commit
stringlengths
40
40
old_file
stringlengths
4
118
new_file
stringlengths
4
118
old_contents
stringlengths
10
2.94k
new_contents
stringlengths
21
3.18k
subject
stringlengths
16
444
message
stringlengths
17
2.63k
lang
stringclasses
1 value
license
stringclasses
13 values
repos
stringlengths
5
43k
ndiff
stringlengths
52
3.32k
instruction
stringlengths
16
444
content
stringlengths
133
4.32k
fuzzy_diff
stringlengths
16
3.18k
7104882ffcd35c24d8df5b9aa909e9bc9619cba7
eli5/__init__.py
eli5/__init__.py
from __future__ import absolute_import __version__ = '0.4.1' from .formatters import format_as_html, format_html_styles, format_as_text from .explain import explain_weights, explain_prediction from .sklearn import explain_weights_sklearn, explain_prediction_sklearn try: from .ipython import show_weights, show_prediction except ImportError: pass # IPython is not installed try: from .lightning import ( explain_prediction_lightning, explain_weights_lightning ) except ImportError as e: # lightning is not available pass try: from .sklearn_crfsuite import ( explain_weights_sklearn_crfsuite ) except ImportError as e: # sklearn-crfsuite is not available pass try: from .xgboost import explain_weights_xgboost except ImportError: # xgboost is not available pass
from __future__ import absolute_import __version__ = '0.4.1' from .formatters import format_as_html, format_html_styles, format_as_text from .explain import explain_weights, explain_prediction from .sklearn import explain_weights_sklearn, explain_prediction_sklearn try: from .ipython import show_weights, show_prediction except ImportError: pass # IPython is not installed try: from .lightning import ( explain_prediction_lightning, explain_weights_lightning ) except ImportError as e: # lightning is not available pass try: from .sklearn_crfsuite import ( explain_weights_sklearn_crfsuite ) except ImportError as e: # sklearn-crfsuite is not available pass try: from .xgboost import explain_weights_xgboost except ImportError: # xgboost is not available pass except Exception as e: if e.__class__.__name__ == 'XGBoostLibraryNotFound': # improperly installed xgboost pass else: raise
Handle improperly installed xgboost. Fixes GH-162.
Handle improperly installed xgboost. Fixes GH-162.
Python
mit
TeamHG-Memex/eli5,TeamHG-Memex/eli5,TeamHG-Memex/eli5
from __future__ import absolute_import __version__ = '0.4.1' from .formatters import format_as_html, format_html_styles, format_as_text from .explain import explain_weights, explain_prediction from .sklearn import explain_weights_sklearn, explain_prediction_sklearn try: from .ipython import show_weights, show_prediction except ImportError: pass # IPython is not installed try: from .lightning import ( explain_prediction_lightning, explain_weights_lightning ) except ImportError as e: # lightning is not available pass try: from .sklearn_crfsuite import ( explain_weights_sklearn_crfsuite ) except ImportError as e: # sklearn-crfsuite is not available pass try: from .xgboost import explain_weights_xgboost except ImportError: # xgboost is not available pass + except Exception as e: + if e.__class__.__name__ == 'XGBoostLibraryNotFound': + # improperly installed xgboost + pass + else: + raise
Handle improperly installed xgboost. Fixes GH-162.
## Code Before: from __future__ import absolute_import __version__ = '0.4.1' from .formatters import format_as_html, format_html_styles, format_as_text from .explain import explain_weights, explain_prediction from .sklearn import explain_weights_sklearn, explain_prediction_sklearn try: from .ipython import show_weights, show_prediction except ImportError: pass # IPython is not installed try: from .lightning import ( explain_prediction_lightning, explain_weights_lightning ) except ImportError as e: # lightning is not available pass try: from .sklearn_crfsuite import ( explain_weights_sklearn_crfsuite ) except ImportError as e: # sklearn-crfsuite is not available pass try: from .xgboost import explain_weights_xgboost except ImportError: # xgboost is not available pass ## Instruction: Handle improperly installed xgboost. Fixes GH-162. ## Code After: from __future__ import absolute_import __version__ = '0.4.1' from .formatters import format_as_html, format_html_styles, format_as_text from .explain import explain_weights, explain_prediction from .sklearn import explain_weights_sklearn, explain_prediction_sklearn try: from .ipython import show_weights, show_prediction except ImportError: pass # IPython is not installed try: from .lightning import ( explain_prediction_lightning, explain_weights_lightning ) except ImportError as e: # lightning is not available pass try: from .sklearn_crfsuite import ( explain_weights_sklearn_crfsuite ) except ImportError as e: # sklearn-crfsuite is not available pass try: from .xgboost import explain_weights_xgboost except ImportError: # xgboost is not available pass except Exception as e: if e.__class__.__name__ == 'XGBoostLibraryNotFound': # improperly installed xgboost pass else: raise
# ... existing code ... pass except Exception as e: if e.__class__.__name__ == 'XGBoostLibraryNotFound': # improperly installed xgboost pass else: raise # ... rest of the code ...
1eacbac722ca949518e1a8e9d6a0a957e193ba9e
tests/functional/staging_and_prod/test_admin.py
tests/functional/staging_and_prod/test_admin.py
from retry.api import retry_call from config import config from tests.pages import UploadCsvPage from tests.postman import ( send_notification_via_csv, get_notification_by_id_via_api, ) from tests.test_utils import assert_notification_body, recordtime, NotificationStatuses @recordtime def test_admin(driver, client, login_user): upload_csv_page = UploadCsvPage(driver) csv_sms_notification_id = send_notification_via_csv(upload_csv_page, 'sms') csv_sms_notification = retry_call( get_notification_by_id_via_api, fargs=[client, csv_sms_notification_id, NotificationStatuses.SENT], tries=config['notification_retry_times'], delay=config['notification_retry_interval'] ) assert_notification_body(csv_sms_notification_id, csv_sms_notification) csv_email_notification_id = send_notification_via_csv(upload_csv_page, 'email') csv_email_notification = retry_call( get_notification_by_id_via_api, fargs=[client, csv_email_notification_id, NotificationStatuses.SENT], tries=config['notification_retry_times'], delay=config['notification_retry_interval'] ) assert_notification_body(csv_email_notification_id, csv_email_notification) upload_csv_page.sign_out()
import pytest from retry.api import retry_call from config import config from tests.pages import UploadCsvPage from tests.postman import ( send_notification_via_csv, get_notification_by_id_via_api, ) from tests.test_utils import assert_notification_body, recordtime, NotificationStatuses @pytest.mark.skip(reason="intermittent pager duty alerts due to queue backlog") def test_admin(driver, client, login_user): upload_csv_page = UploadCsvPage(driver) csv_sms_notification_id = send_notification_via_csv(upload_csv_page, 'sms') csv_sms_notification = retry_call( get_notification_by_id_via_api, fargs=[client, csv_sms_notification_id, NotificationStatuses.SENT], tries=config['notification_retry_times'], delay=config['notification_retry_interval'] ) assert_notification_body(csv_sms_notification_id, csv_sms_notification) csv_email_notification_id = send_notification_via_csv(upload_csv_page, 'email') csv_email_notification = retry_call( get_notification_by_id_via_api, fargs=[client, csv_email_notification_id, NotificationStatuses.SENT], tries=config['notification_retry_times'], delay=config['notification_retry_interval'] ) assert_notification_body(csv_email_notification_id, csv_email_notification) upload_csv_page.sign_out()
Disable CSV upload tests temporarily
Disable CSV upload tests temporarily When the database tasks queue builds up we get false pager duty alerts due to the time it takes for the test csv to get through to the front of the queue.
Python
mit
alphagov/notifications-functional-tests,alphagov/notifications-functional-tests
+ import pytest + from retry.api import retry_call from config import config from tests.pages import UploadCsvPage from tests.postman import ( send_notification_via_csv, get_notification_by_id_via_api, ) from tests.test_utils import assert_notification_body, recordtime, NotificationStatuses - @recordtime + @pytest.mark.skip(reason="intermittent pager duty alerts due to queue backlog") def test_admin(driver, client, login_user): upload_csv_page = UploadCsvPage(driver) csv_sms_notification_id = send_notification_via_csv(upload_csv_page, 'sms') csv_sms_notification = retry_call( get_notification_by_id_via_api, fargs=[client, csv_sms_notification_id, NotificationStatuses.SENT], tries=config['notification_retry_times'], delay=config['notification_retry_interval'] ) assert_notification_body(csv_sms_notification_id, csv_sms_notification) csv_email_notification_id = send_notification_via_csv(upload_csv_page, 'email') csv_email_notification = retry_call( get_notification_by_id_via_api, fargs=[client, csv_email_notification_id, NotificationStatuses.SENT], tries=config['notification_retry_times'], delay=config['notification_retry_interval'] ) assert_notification_body(csv_email_notification_id, csv_email_notification) upload_csv_page.sign_out()
Disable CSV upload tests temporarily
## Code Before: from retry.api import retry_call from config import config from tests.pages import UploadCsvPage from tests.postman import ( send_notification_via_csv, get_notification_by_id_via_api, ) from tests.test_utils import assert_notification_body, recordtime, NotificationStatuses @recordtime def test_admin(driver, client, login_user): upload_csv_page = UploadCsvPage(driver) csv_sms_notification_id = send_notification_via_csv(upload_csv_page, 'sms') csv_sms_notification = retry_call( get_notification_by_id_via_api, fargs=[client, csv_sms_notification_id, NotificationStatuses.SENT], tries=config['notification_retry_times'], delay=config['notification_retry_interval'] ) assert_notification_body(csv_sms_notification_id, csv_sms_notification) csv_email_notification_id = send_notification_via_csv(upload_csv_page, 'email') csv_email_notification = retry_call( get_notification_by_id_via_api, fargs=[client, csv_email_notification_id, NotificationStatuses.SENT], tries=config['notification_retry_times'], delay=config['notification_retry_interval'] ) assert_notification_body(csv_email_notification_id, csv_email_notification) upload_csv_page.sign_out() ## Instruction: Disable CSV upload tests temporarily ## Code After: import pytest from retry.api import retry_call from config import config from tests.pages import UploadCsvPage from tests.postman import ( send_notification_via_csv, get_notification_by_id_via_api, ) from tests.test_utils import assert_notification_body, recordtime, NotificationStatuses @pytest.mark.skip(reason="intermittent pager duty alerts due to queue backlog") def test_admin(driver, client, login_user): upload_csv_page = UploadCsvPage(driver) csv_sms_notification_id = send_notification_via_csv(upload_csv_page, 'sms') csv_sms_notification = retry_call( get_notification_by_id_via_api, fargs=[client, csv_sms_notification_id, NotificationStatuses.SENT], tries=config['notification_retry_times'], delay=config['notification_retry_interval'] ) assert_notification_body(csv_sms_notification_id, csv_sms_notification) csv_email_notification_id = send_notification_via_csv(upload_csv_page, 'email') csv_email_notification = retry_call( get_notification_by_id_via_api, fargs=[client, csv_email_notification_id, NotificationStatuses.SENT], tries=config['notification_retry_times'], delay=config['notification_retry_interval'] ) assert_notification_body(csv_email_notification_id, csv_email_notification) upload_csv_page.sign_out()
... import pytest from retry.api import retry_call ... @pytest.mark.skip(reason="intermittent pager duty alerts due to queue backlog") def test_admin(driver, client, login_user): ...
775170d69862aaff63231b669639a872596ed2cd
test_interpreter.py
test_interpreter.py
import unittest import brainfuck test_cases = [("++++++++[>++++[>++>+++>+++>+<<<<-]>+>+>->>+[<]<-]>>.>---.+++++++..+++.>>.<-.<.+++.------.--------.>>+.>++.", "Hello World!\n")] class InterpreterTestCase(unittest.TestCase): def setUp(self): self.interpreter = brainfuck.BrainfuckInterpreter() def runTest(self): for case in test_cases: self.assertEqual(case[1], self.interpreter.eval(case[0]))
import unittest import brainfuck hello_case = ("++++++++[>++++[>++>+++>+++>+<<<<-]>+>+>->>+[<]<-]>>.>---.+++++++..+++.>>.<-.<.+++.------.--------.>>+.>++.", "Hello World!\n") class InterpreterTestCase(unittest.TestCase): def setUp(self): self.interpreter = brainfuck.BrainfuckInterpreter() def test_hello_world(self): self.assertEqual(hello_case[1], self.interpreter.eval(hello_case[0])) def test_missing_parenthesis(self): self.assertRaises(SyntaxError, self.interpreter.eval, '[++]+]')
Add unittest for missing parenthesis
Add unittest for missing parenthesis
Python
bsd-3-clause
handrake/brainfuck
import unittest import brainfuck - test_cases = [("++++++++[>++++[>++>+++>+++>+<<<<-]>+>+>->>+[<]<-]>>.>---.+++++++..+++.>>.<-.<.+++.------.--------.>>+.>++.", "Hello World!\n")] + hello_case = ("++++++++[>++++[>++>+++>+++>+<<<<-]>+>+>->>+[<]<-]>>.>---.+++++++..+++.>>.<-.<.+++.------.--------.>>+.>++.", "Hello World!\n") class InterpreterTestCase(unittest.TestCase): def setUp(self): self.interpreter = brainfuck.BrainfuckInterpreter() + def test_hello_world(self): - def runTest(self): - for case in test_cases: - self.assertEqual(case[1], self.interpreter.eval(case[0])) + self.assertEqual(hello_case[1], self.interpreter.eval(hello_case[0])) + def test_missing_parenthesis(self): + self.assertRaises(SyntaxError, self.interpreter.eval, '[++]+]')
Add unittest for missing parenthesis
## Code Before: import unittest import brainfuck test_cases = [("++++++++[>++++[>++>+++>+++>+<<<<-]>+>+>->>+[<]<-]>>.>---.+++++++..+++.>>.<-.<.+++.------.--------.>>+.>++.", "Hello World!\n")] class InterpreterTestCase(unittest.TestCase): def setUp(self): self.interpreter = brainfuck.BrainfuckInterpreter() def runTest(self): for case in test_cases: self.assertEqual(case[1], self.interpreter.eval(case[0])) ## Instruction: Add unittest for missing parenthesis ## Code After: import unittest import brainfuck hello_case = ("++++++++[>++++[>++>+++>+++>+<<<<-]>+>+>->>+[<]<-]>>.>---.+++++++..+++.>>.<-.<.+++.------.--------.>>+.>++.", "Hello World!\n") class InterpreterTestCase(unittest.TestCase): def setUp(self): self.interpreter = brainfuck.BrainfuckInterpreter() def test_hello_world(self): self.assertEqual(hello_case[1], self.interpreter.eval(hello_case[0])) def test_missing_parenthesis(self): self.assertRaises(SyntaxError, self.interpreter.eval, '[++]+]')
... hello_case = ("++++++++[>++++[>++>+++>+++>+<<<<-]>+>+>->>+[<]<-]>>.>---.+++++++..+++.>>.<-.<.+++.------.--------.>>+.>++.", "Hello World!\n") ... self.interpreter = brainfuck.BrainfuckInterpreter() def test_hello_world(self): self.assertEqual(hello_case[1], self.interpreter.eval(hello_case[0])) def test_missing_parenthesis(self): self.assertRaises(SyntaxError, self.interpreter.eval, '[++]+]') ...
cd9cb42c16d443a26c7054e27c3ebc254142fbfb
python/ligscore/__init__.py
python/ligscore/__init__.py
import saliweb.backend class Job(saliweb.backend.Job): runnercls = saliweb.backend.SGERunner def run(self): libs = {'PoseScore': 'protein_ligand_pose_score.lib', 'RankScore': 'protein_ligand_rank_score.lib'} pdb, mol2, lib = open('input.txt').readline().strip().split(' ') lib = libs[lib] script = """ module load imp lib=`python -c "import IMP.atom; print IMP.atom.get_data_path('%s')"` ligand_score %s %s "$lib" > score.list 2> score.log """ % (lib, mol2, pdb) r = self.runnercls(script) r.set_sge_options('-l arch=linux-x64') return r def get_web_service(config_file): db = saliweb.backend.Database(Job) config = saliweb.backend.Config(config_file) return saliweb.backend.WebService(config, db)
import saliweb.backend class Job(saliweb.backend.Job): runnercls = saliweb.backend.WyntonSGERunner def run(self): libs = {'PoseScore': 'protein_ligand_pose_score.lib', 'RankScore': 'protein_ligand_rank_score.lib'} pdb, mol2, lib = open('input.txt').readline().strip().split(' ') lib = libs[lib] script = """ module load Sali module load imp lib=`python -c "import IMP.atom; print IMP.atom.get_data_path('%s')"` ligand_score %s %s "$lib" > score.list 2> score.log """ % (lib, mol2, pdb) r = self.runnercls(script) r.set_sge_options('-l arch=lx-amd64') return r def get_web_service(config_file): db = saliweb.backend.Database(Job) config = saliweb.backend.Config(config_file) return saliweb.backend.WebService(config, db)
Switch to new Wynton cluster
Switch to new Wynton cluster
Python
lgpl-2.1
salilab/ligscore,salilab/ligscore
import saliweb.backend class Job(saliweb.backend.Job): - runnercls = saliweb.backend.SGERunner + runnercls = saliweb.backend.WyntonSGERunner def run(self): libs = {'PoseScore': 'protein_ligand_pose_score.lib', 'RankScore': 'protein_ligand_rank_score.lib'} pdb, mol2, lib = open('input.txt').readline().strip().split(' ') lib = libs[lib] script = """ + module load Sali module load imp lib=`python -c "import IMP.atom; print IMP.atom.get_data_path('%s')"` ligand_score %s %s "$lib" > score.list 2> score.log """ % (lib, mol2, pdb) r = self.runnercls(script) - r.set_sge_options('-l arch=linux-x64') + r.set_sge_options('-l arch=lx-amd64') return r def get_web_service(config_file): db = saliweb.backend.Database(Job) config = saliweb.backend.Config(config_file) return saliweb.backend.WebService(config, db)
Switch to new Wynton cluster
## Code Before: import saliweb.backend class Job(saliweb.backend.Job): runnercls = saliweb.backend.SGERunner def run(self): libs = {'PoseScore': 'protein_ligand_pose_score.lib', 'RankScore': 'protein_ligand_rank_score.lib'} pdb, mol2, lib = open('input.txt').readline().strip().split(' ') lib = libs[lib] script = """ module load imp lib=`python -c "import IMP.atom; print IMP.atom.get_data_path('%s')"` ligand_score %s %s "$lib" > score.list 2> score.log """ % (lib, mol2, pdb) r = self.runnercls(script) r.set_sge_options('-l arch=linux-x64') return r def get_web_service(config_file): db = saliweb.backend.Database(Job) config = saliweb.backend.Config(config_file) return saliweb.backend.WebService(config, db) ## Instruction: Switch to new Wynton cluster ## Code After: import saliweb.backend class Job(saliweb.backend.Job): runnercls = saliweb.backend.WyntonSGERunner def run(self): libs = {'PoseScore': 'protein_ligand_pose_score.lib', 'RankScore': 'protein_ligand_rank_score.lib'} pdb, mol2, lib = open('input.txt').readline().strip().split(' ') lib = libs[lib] script = """ module load Sali module load imp lib=`python -c "import IMP.atom; print IMP.atom.get_data_path('%s')"` ligand_score %s %s "$lib" > score.list 2> score.log """ % (lib, mol2, pdb) r = self.runnercls(script) r.set_sge_options('-l arch=lx-amd64') return r def get_web_service(config_file): db = saliweb.backend.Database(Job) config = saliweb.backend.Config(config_file) return saliweb.backend.WebService(config, db)
... class Job(saliweb.backend.Job): runnercls = saliweb.backend.WyntonSGERunner ... script = """ module load Sali module load imp ... r = self.runnercls(script) r.set_sge_options('-l arch=lx-amd64') return r ...
c1ed5eb96b04ca0af2ad8f26023d8cbaa4a75eda
rx/concurrency/threadpoolscheduler.py
rx/concurrency/threadpoolscheduler.py
import logging from concurrent.futures import ThreadPoolExecutor from rx.core import Scheduler, Disposable from rx.disposables import SingleAssignmentDisposable, CompositeDisposable from .timeoutscheduler import TimeoutScheduler log = logging.getLogger("Rx") class ThreadPoolScheduler(TimeoutScheduler): """A scheduler that schedules work via the thread pool and threading timers.""" def __init__(self, max_workers=None): self.executor = ThreadPoolExecutor(max_workers=max_workers) def schedule(self, action, state=None): """Schedules an action to be executed.""" disposable = SingleAssignmentDisposable() def run(): disposable.disposable = self.invoke_action(action, state) future = self.executor.submit(run) def dispose(): future.cancel() return CompositeDisposable(disposable, Disposable.create(dispose)) Scheduler.thread_pool = thread_pool_scheduler = ThreadPoolScheduler()
from concurrent.futures import ThreadPoolExecutor from rx.core import Scheduler from .newthreadscheduler import NewThreadScheduler class ThreadPoolScheduler(NewThreadScheduler): """A scheduler that schedules work via the thread pool.""" class ThreadPoolThread: """Wraps a concurrent future as a thread.""" def __init__(self, executor, run): self.run = run self.future = None self.executor = executor def start(self): self.future = self.executor.submit(self.run) def cancel(self): self.future.cancel() def __init__(self, max_workers=None): super(ThreadPoolScheduler, self).__init__(self.thread_factory) self.executor = ThreadPoolExecutor(max_workers=max_workers) def thread_factory(self, target, *args): return self.ThreadPoolThread(self.executor, target) Scheduler.thread_pool = thread_pool_scheduler = ThreadPoolScheduler()
Make thread pool scheduler behave as a pooled new thread scheduler
Make thread pool scheduler behave as a pooled new thread scheduler
Python
mit
ReactiveX/RxPY,ReactiveX/RxPY
- import logging from concurrent.futures import ThreadPoolExecutor - from rx.core import Scheduler, Disposable + from rx.core import Scheduler - from rx.disposables import SingleAssignmentDisposable, CompositeDisposable - from .timeoutscheduler import TimeoutScheduler + from .newthreadscheduler import NewThreadScheduler - - log = logging.getLogger("Rx") - class ThreadPoolScheduler(TimeoutScheduler): + class ThreadPoolScheduler(NewThreadScheduler): - """A scheduler that schedules work via the thread pool and threading + """A scheduler that schedules work via the thread pool.""" - timers.""" + + class ThreadPoolThread: + """Wraps a concurrent future as a thread.""" + + def __init__(self, executor, run): + self.run = run + self.future = None + self.executor = executor + + def start(self): + self.future = self.executor.submit(self.run) + + def cancel(self): + self.future.cancel() def __init__(self, max_workers=None): + super(ThreadPoolScheduler, self).__init__(self.thread_factory) self.executor = ThreadPoolExecutor(max_workers=max_workers) + def thread_factory(self, target, *args): + return self.ThreadPoolThread(self.executor, target) - def schedule(self, action, state=None): - """Schedules an action to be executed.""" - - disposable = SingleAssignmentDisposable() - - def run(): - disposable.disposable = self.invoke_action(action, state) - future = self.executor.submit(run) - - def dispose(): - future.cancel() - return CompositeDisposable(disposable, Disposable.create(dispose)) Scheduler.thread_pool = thread_pool_scheduler = ThreadPoolScheduler()
Make thread pool scheduler behave as a pooled new thread scheduler
## Code Before: import logging from concurrent.futures import ThreadPoolExecutor from rx.core import Scheduler, Disposable from rx.disposables import SingleAssignmentDisposable, CompositeDisposable from .timeoutscheduler import TimeoutScheduler log = logging.getLogger("Rx") class ThreadPoolScheduler(TimeoutScheduler): """A scheduler that schedules work via the thread pool and threading timers.""" def __init__(self, max_workers=None): self.executor = ThreadPoolExecutor(max_workers=max_workers) def schedule(self, action, state=None): """Schedules an action to be executed.""" disposable = SingleAssignmentDisposable() def run(): disposable.disposable = self.invoke_action(action, state) future = self.executor.submit(run) def dispose(): future.cancel() return CompositeDisposable(disposable, Disposable.create(dispose)) Scheduler.thread_pool = thread_pool_scheduler = ThreadPoolScheduler() ## Instruction: Make thread pool scheduler behave as a pooled new thread scheduler ## Code After: from concurrent.futures import ThreadPoolExecutor from rx.core import Scheduler from .newthreadscheduler import NewThreadScheduler class ThreadPoolScheduler(NewThreadScheduler): """A scheduler that schedules work via the thread pool.""" class ThreadPoolThread: """Wraps a concurrent future as a thread.""" def __init__(self, executor, run): self.run = run self.future = None self.executor = executor def start(self): self.future = self.executor.submit(self.run) def cancel(self): self.future.cancel() def __init__(self, max_workers=None): super(ThreadPoolScheduler, self).__init__(self.thread_factory) self.executor = ThreadPoolExecutor(max_workers=max_workers) def thread_factory(self, target, *args): return self.ThreadPoolThread(self.executor, target) Scheduler.thread_pool = thread_pool_scheduler = ThreadPoolScheduler()
# ... existing code ... from concurrent.futures import ThreadPoolExecutor # ... modified code ... from rx.core import Scheduler from .newthreadscheduler import NewThreadScheduler ... class ThreadPoolScheduler(NewThreadScheduler): """A scheduler that schedules work via the thread pool.""" class ThreadPoolThread: """Wraps a concurrent future as a thread.""" def __init__(self, executor, run): self.run = run self.future = None self.executor = executor def start(self): self.future = self.executor.submit(self.run) def cancel(self): self.future.cancel() ... def __init__(self, max_workers=None): super(ThreadPoolScheduler, self).__init__(self.thread_factory) self.executor = ThreadPoolExecutor(max_workers=max_workers) ... def thread_factory(self, target, *args): return self.ThreadPoolThread(self.executor, target) # ... rest of the code ...
faba2bc98f08cddea51d2e0093aa5c2981c8bf15
gdrived.py
gdrived.py
import sys import time import daemon class GDriveDaemon(daemon.Daemon): def run(self): while True: time.sleep(1)
import sys import time import daemon UPDATE_INTERVAL = 30 # Sync update interval in seconds. class GDriveDaemon(daemon.Daemon, object): def __init__(self): "Class constructor." # Use pidfile in Gdrive config directory. pidfile = None # Use loglevel from GDrive config. loglevel = None # Use logfile in GDrive config directory. stdout = None super(GDriveDaemon, self).__init__(pidfile, loglevel, stdout) def run(self): "Run the daemon." while True: time.sleep(UPDATE_INTERVAL)
Add update interval constant. Add detail to constructor.
Add update interval constant. Add detail to constructor.
Python
apache-2.0
babycaseny/gdrive-linux,jimlawton/gdrive-linux-googlecode,jimlawton/gdrive-linux,jmfield2/gdrive-linux
import sys import time import daemon + UPDATE_INTERVAL = 30 # Sync update interval in seconds. + - class GDriveDaemon(daemon.Daemon): + class GDriveDaemon(daemon.Daemon, object): + + def __init__(self): + "Class constructor." + + # Use pidfile in Gdrive config directory. + pidfile = None + + # Use loglevel from GDrive config. + loglevel = None + + # Use logfile in GDrive config directory. + stdout = None + + super(GDriveDaemon, self).__init__(pidfile, loglevel, stdout) + def run(self): + "Run the daemon." + while True: + - time.sleep(1) + time.sleep(UPDATE_INTERVAL)
Add update interval constant. Add detail to constructor.
## Code Before: import sys import time import daemon class GDriveDaemon(daemon.Daemon): def run(self): while True: time.sleep(1) ## Instruction: Add update interval constant. Add detail to constructor. ## Code After: import sys import time import daemon UPDATE_INTERVAL = 30 # Sync update interval in seconds. class GDriveDaemon(daemon.Daemon, object): def __init__(self): "Class constructor." # Use pidfile in Gdrive config directory. pidfile = None # Use loglevel from GDrive config. loglevel = None # Use logfile in GDrive config directory. stdout = None super(GDriveDaemon, self).__init__(pidfile, loglevel, stdout) def run(self): "Run the daemon." while True: time.sleep(UPDATE_INTERVAL)
... UPDATE_INTERVAL = 30 # Sync update interval in seconds. class GDriveDaemon(daemon.Daemon, object): def __init__(self): "Class constructor." # Use pidfile in Gdrive config directory. pidfile = None # Use loglevel from GDrive config. loglevel = None # Use logfile in GDrive config directory. stdout = None super(GDriveDaemon, self).__init__(pidfile, loglevel, stdout) def run(self): "Run the daemon." while True: time.sleep(UPDATE_INTERVAL) ...
e86901ac2b074d42d2e388353bbe60fcdd8f0240
wagtail/contrib/postgres_search/apps.py
wagtail/contrib/postgres_search/apps.py
from django.apps import AppConfig from django.core.checks import Error, Tags, register from .utils import get_postgresql_connections, set_weights class PostgresSearchConfig(AppConfig): name = 'wagtail.contrib.postgres_search' def ready(self): @register(Tags.compatibility, Tags.database) def check_if_postgresql(app_configs, **kwargs): if get_postgresql_connections(): return [] return [Error('You must use a PostgreSQL database ' 'to use PostgreSQL search.', id='wagtail.contrib.postgres_search.E001')] set_weights() from .models import IndexEntry IndexEntry.add_generic_relations()
from django.apps import AppConfig from django.core.checks import Error, Tags, register from .utils import get_postgresql_connections, set_weights class PostgresSearchConfig(AppConfig): name = 'wagtail.contrib.postgres_search' default_auto_field = 'django.db.models.AutoField' def ready(self): @register(Tags.compatibility, Tags.database) def check_if_postgresql(app_configs, **kwargs): if get_postgresql_connections(): return [] return [Error('You must use a PostgreSQL database ' 'to use PostgreSQL search.', id='wagtail.contrib.postgres_search.E001')] set_weights() from .models import IndexEntry IndexEntry.add_generic_relations()
Set default_auto_field in wagtail.contrib.postgres_search AppConfig
Set default_auto_field in wagtail.contrib.postgres_search AppConfig Add default_auto_field = 'django.db.models.AutoField' Co-authored-by: Nick Moreton <[email protected]>
Python
bsd-3-clause
jnns/wagtail,zerolab/wagtail,gasman/wagtail,gasman/wagtail,gasman/wagtail,rsalmaso/wagtail,rsalmaso/wagtail,thenewguy/wagtail,thenewguy/wagtail,jnns/wagtail,jnns/wagtail,rsalmaso/wagtail,wagtail/wagtail,mixxorz/wagtail,torchbox/wagtail,jnns/wagtail,gasman/wagtail,thenewguy/wagtail,thenewguy/wagtail,wagtail/wagtail,mixxorz/wagtail,zerolab/wagtail,torchbox/wagtail,wagtail/wagtail,mixxorz/wagtail,mixxorz/wagtail,zerolab/wagtail,torchbox/wagtail,mixxorz/wagtail,rsalmaso/wagtail,wagtail/wagtail,wagtail/wagtail,gasman/wagtail,zerolab/wagtail,thenewguy/wagtail,zerolab/wagtail,torchbox/wagtail,rsalmaso/wagtail
from django.apps import AppConfig from django.core.checks import Error, Tags, register from .utils import get_postgresql_connections, set_weights class PostgresSearchConfig(AppConfig): name = 'wagtail.contrib.postgres_search' + default_auto_field = 'django.db.models.AutoField' def ready(self): @register(Tags.compatibility, Tags.database) def check_if_postgresql(app_configs, **kwargs): if get_postgresql_connections(): return [] return [Error('You must use a PostgreSQL database ' 'to use PostgreSQL search.', id='wagtail.contrib.postgres_search.E001')] set_weights() from .models import IndexEntry IndexEntry.add_generic_relations()
Set default_auto_field in wagtail.contrib.postgres_search AppConfig
## Code Before: from django.apps import AppConfig from django.core.checks import Error, Tags, register from .utils import get_postgresql_connections, set_weights class PostgresSearchConfig(AppConfig): name = 'wagtail.contrib.postgres_search' def ready(self): @register(Tags.compatibility, Tags.database) def check_if_postgresql(app_configs, **kwargs): if get_postgresql_connections(): return [] return [Error('You must use a PostgreSQL database ' 'to use PostgreSQL search.', id='wagtail.contrib.postgres_search.E001')] set_weights() from .models import IndexEntry IndexEntry.add_generic_relations() ## Instruction: Set default_auto_field in wagtail.contrib.postgres_search AppConfig ## Code After: from django.apps import AppConfig from django.core.checks import Error, Tags, register from .utils import get_postgresql_connections, set_weights class PostgresSearchConfig(AppConfig): name = 'wagtail.contrib.postgres_search' default_auto_field = 'django.db.models.AutoField' def ready(self): @register(Tags.compatibility, Tags.database) def check_if_postgresql(app_configs, **kwargs): if get_postgresql_connections(): return [] return [Error('You must use a PostgreSQL database ' 'to use PostgreSQL search.', id='wagtail.contrib.postgres_search.E001')] set_weights() from .models import IndexEntry IndexEntry.add_generic_relations()
... name = 'wagtail.contrib.postgres_search' default_auto_field = 'django.db.models.AutoField' ...
5decd7e68c6454e455bc1debe232ea37f7260c58
mixins.py
mixins.py
class DepthSerializerMixin(object): """Custom method 'get_serializer_class', set attribute 'depth' based on query parameter in the url""" def get_serializer_class(self): serializer_class = self.serializer_class query_params = self.request.QUERY_PARAMS depth = query_params.get('__depth', None) serializer_class.Meta.depth = int(depth) if(depth != None and depth.isdigit()) else 0 return serializer_class
class DepthSerializerMixin(object): """Custom method 'get_serializer_class', set attribute 'depth' based on query parameter in the url""" def get_serializer_class(self, *args, **kwargs): serializer_class = super(DepthSerializerMixin, self).get_serializer_class(*args, **kwargs) query_params = self.request.QUERY_PARAMS depth = query_params.get('__depth', None) serializer_class.Meta.depth = int(depth) if(depth != None and depth.isdigit()) else 0 return serializer_class
Call method 'get_serializer_class' of the Class parent
Call method 'get_serializer_class' of the Class parent
Python
mit
krescruz/depth-serializer-mixin
class DepthSerializerMixin(object): """Custom method 'get_serializer_class', set attribute 'depth' based on query parameter in the url""" - def get_serializer_class(self): + def get_serializer_class(self, *args, **kwargs): - serializer_class = self.serializer_class + serializer_class = super(DepthSerializerMixin, self).get_serializer_class(*args, **kwargs) query_params = self.request.QUERY_PARAMS depth = query_params.get('__depth', None) serializer_class.Meta.depth = int(depth) if(depth != None and depth.isdigit()) else 0 return serializer_class
Call method 'get_serializer_class' of the Class parent
## Code Before: class DepthSerializerMixin(object): """Custom method 'get_serializer_class', set attribute 'depth' based on query parameter in the url""" def get_serializer_class(self): serializer_class = self.serializer_class query_params = self.request.QUERY_PARAMS depth = query_params.get('__depth', None) serializer_class.Meta.depth = int(depth) if(depth != None and depth.isdigit()) else 0 return serializer_class ## Instruction: Call method 'get_serializer_class' of the Class parent ## Code After: class DepthSerializerMixin(object): """Custom method 'get_serializer_class', set attribute 'depth' based on query parameter in the url""" def get_serializer_class(self, *args, **kwargs): serializer_class = super(DepthSerializerMixin, self).get_serializer_class(*args, **kwargs) query_params = self.request.QUERY_PARAMS depth = query_params.get('__depth', None) serializer_class.Meta.depth = int(depth) if(depth != None and depth.isdigit()) else 0 return serializer_class
... def get_serializer_class(self, *args, **kwargs): serializer_class = super(DepthSerializerMixin, self).get_serializer_class(*args, **kwargs) query_params = self.request.QUERY_PARAMS ...
593bab981f36f7af52ae55914c18e368e8c1a94f
examples/app-on-ws-init.py
examples/app-on-ws-init.py
from argparse import ArgumentParser import i3ipc i3 = i3ipc.Connection() parser = ArgumentParser(description='Open an application on a given workspace when it is initialized') parser.add_argument('--workspace', metavar='NAME', help='The name of the workspace') parser.add_argument('--command', metavar='CMD', help='The command to run on the newly initted workspace') args = parser.parse_args() def on_workspace(i3, e): if e.current.props.name == args.workspace and not len(e.current.leaves()): i3.command('exec {}'.format(args.command)) i3.on('workspace::focus', on_workspace) i3.main()
from argparse import ArgumentParser import i3ipc i3 = i3ipc.Connection() parser = ArgumentParser(description="""Open the given application each time the given workspace is created. For instance, running 'app-on-ws-init.py 6 i3-sensible-terminal' should open your terminal as soon as you create the workspace 6. """) parser.add_argument('workspace', metavar='WS_NAME', help='The name of the workspace') parser.add_argument('command', metavar='CMD', help='The command to run on the newly initted workspace') args = parser.parse_args() def on_workspace(i3, e): if e.current.props.name == args.workspace and not len(e.current.leaves()): i3.command('exec {}'.format(args.command)) i3.on('workspace::focus', on_workspace) i3.main()
Make the 2 mandatory parameters mandatory. Make the help message a bit clearer and provides an example.
Make the 2 mandatory parameters mandatory. Make the help message a bit clearer and provides an example.
Python
bsd-3-clause
xenomachina/i3ipc-python,nicoe/i3ipc-python,acrisci/i3ipc-python,chrsclmn/i3ipc-python
from argparse import ArgumentParser import i3ipc i3 = i3ipc.Connection() - parser = ArgumentParser(description='Open an application on a given workspace when it is initialized') + parser = ArgumentParser(description="""Open the given application each time the + given workspace is created. For instance, running 'app-on-ws-init.py 6 + i3-sensible-terminal' should open your terminal as soon as you create the + workspace 6. + """) - parser.add_argument('--workspace', metavar='NAME', help='The name of the workspace') + parser.add_argument('workspace', metavar='WS_NAME', help='The name of the workspace') - parser.add_argument('--command', metavar='CMD', help='The command to run on the newly initted workspace') + parser.add_argument('command', metavar='CMD', help='The command to run on the newly initted workspace') args = parser.parse_args() def on_workspace(i3, e): if e.current.props.name == args.workspace and not len(e.current.leaves()): i3.command('exec {}'.format(args.command)) i3.on('workspace::focus', on_workspace) i3.main()
Make the 2 mandatory parameters mandatory. Make the help message a bit clearer and provides an example.
## Code Before: from argparse import ArgumentParser import i3ipc i3 = i3ipc.Connection() parser = ArgumentParser(description='Open an application on a given workspace when it is initialized') parser.add_argument('--workspace', metavar='NAME', help='The name of the workspace') parser.add_argument('--command', metavar='CMD', help='The command to run on the newly initted workspace') args = parser.parse_args() def on_workspace(i3, e): if e.current.props.name == args.workspace and not len(e.current.leaves()): i3.command('exec {}'.format(args.command)) i3.on('workspace::focus', on_workspace) i3.main() ## Instruction: Make the 2 mandatory parameters mandatory. Make the help message a bit clearer and provides an example. ## Code After: from argparse import ArgumentParser import i3ipc i3 = i3ipc.Connection() parser = ArgumentParser(description="""Open the given application each time the given workspace is created. For instance, running 'app-on-ws-init.py 6 i3-sensible-terminal' should open your terminal as soon as you create the workspace 6. """) parser.add_argument('workspace', metavar='WS_NAME', help='The name of the workspace') parser.add_argument('command', metavar='CMD', help='The command to run on the newly initted workspace') args = parser.parse_args() def on_workspace(i3, e): if e.current.props.name == args.workspace and not len(e.current.leaves()): i3.command('exec {}'.format(args.command)) i3.on('workspace::focus', on_workspace) i3.main()
... parser = ArgumentParser(description="""Open the given application each time the given workspace is created. For instance, running 'app-on-ws-init.py 6 i3-sensible-terminal' should open your terminal as soon as you create the workspace 6. """) parser.add_argument('workspace', metavar='WS_NAME', help='The name of the workspace') parser.add_argument('command', metavar='CMD', help='The command to run on the newly initted workspace') ...
0b8b32a044e92f4e996af734d44a2d93d1492684
project_code/bulk_fitting.py
project_code/bulk_fitting.py
''' Bulk spectral line fitting with SDSS galaxy spectra ''' import os from astropy.io import fits from pandas import concat # Bring in the package funcs from specfit import do_specfit from download_spectra import download_spectra def bulk_fit(obs_file, output_file, keep_spectra=False): ''' Downloads files based off of the entries in the given file, performs spectral line fitting and saves the results to a FITS table. ''' # Open the file data_file = fits.open(obs_file) spectra_data = data_file[1].data del data_file num_spectra = spectra_data['Z'].shape[0] for i in range(num_spectra): spec_info = spectra_data[i] # Download the spectrum spec_name = \ download_spectra(spec_info['PLATEID'], spec_info['FIBREID'], spec_info['MJD'], spec_info['SURVEY']) spec_df = do_specfit(spec_name, verbose=False) if i == 0: df = spec_df else: df = concat([df, spec_df]) if not keep_spectra: os.system('rm ' + spec_name) df.write(output_file) return
''' Bulk spectral line fitting with SDSS galaxy spectra ''' import os from astropy.io import fits from pandas import DataFrame # Bring in the package funcs from specfit import do_specfit from download_spectra import download_spectra def bulk_fit(obs_file, output_file, keep_spectra=False): ''' Downloads files based off of the entries in the given file, performs spectral line fitting and saves the results to a FITS table. ''' # Open the file data_file = fits.open(obs_file) spectra_data = data_file[1].data del data_file num_spectra = spectra_data['Z'].shape[0] for i in range(num_spectra): spec_info = spectra_data[i] # Download the spectrum spec_name = \ download_spectra(spec_info['PLATE'], spec_info['FIBERID'], spec_info['MJD'], spec_info['SURVEY']) spec_df = do_specfit(spec_name, verbose=False) if i == 0: df = DataFrame(spec_df, columns=[spec_name[:-5]]) else: df[spec_name[:-5]] = spec_df if not keep_spectra: os.system('rm ' + spec_name) df.to_csv(output_file) return
Correct names, concat dataframes properly
Correct names, concat dataframes properly
Python
mit
e-koch/Phys-595
''' Bulk spectral line fitting with SDSS galaxy spectra ''' import os from astropy.io import fits - from pandas import concat + from pandas import DataFrame # Bring in the package funcs from specfit import do_specfit from download_spectra import download_spectra def bulk_fit(obs_file, output_file, keep_spectra=False): ''' Downloads files based off of the entries in the given file, performs spectral line fitting and saves the results to a FITS table. ''' # Open the file data_file = fits.open(obs_file) spectra_data = data_file[1].data del data_file num_spectra = spectra_data['Z'].shape[0] for i in range(num_spectra): spec_info = spectra_data[i] # Download the spectrum spec_name = \ - download_spectra(spec_info['PLATEID'], spec_info['FIBREID'], + download_spectra(spec_info['PLATE'], spec_info['FIBERID'], spec_info['MJD'], spec_info['SURVEY']) spec_df = do_specfit(spec_name, verbose=False) if i == 0: - df = spec_df + df = DataFrame(spec_df, columns=[spec_name[:-5]]) else: - df = concat([df, spec_df]) + df[spec_name[:-5]] = spec_df if not keep_spectra: os.system('rm ' + spec_name) - df.write(output_file) + df.to_csv(output_file) return
Correct names, concat dataframes properly
## Code Before: ''' Bulk spectral line fitting with SDSS galaxy spectra ''' import os from astropy.io import fits from pandas import concat # Bring in the package funcs from specfit import do_specfit from download_spectra import download_spectra def bulk_fit(obs_file, output_file, keep_spectra=False): ''' Downloads files based off of the entries in the given file, performs spectral line fitting and saves the results to a FITS table. ''' # Open the file data_file = fits.open(obs_file) spectra_data = data_file[1].data del data_file num_spectra = spectra_data['Z'].shape[0] for i in range(num_spectra): spec_info = spectra_data[i] # Download the spectrum spec_name = \ download_spectra(spec_info['PLATEID'], spec_info['FIBREID'], spec_info['MJD'], spec_info['SURVEY']) spec_df = do_specfit(spec_name, verbose=False) if i == 0: df = spec_df else: df = concat([df, spec_df]) if not keep_spectra: os.system('rm ' + spec_name) df.write(output_file) return ## Instruction: Correct names, concat dataframes properly ## Code After: ''' Bulk spectral line fitting with SDSS galaxy spectra ''' import os from astropy.io import fits from pandas import DataFrame # Bring in the package funcs from specfit import do_specfit from download_spectra import download_spectra def bulk_fit(obs_file, output_file, keep_spectra=False): ''' Downloads files based off of the entries in the given file, performs spectral line fitting and saves the results to a FITS table. ''' # Open the file data_file = fits.open(obs_file) spectra_data = data_file[1].data del data_file num_spectra = spectra_data['Z'].shape[0] for i in range(num_spectra): spec_info = spectra_data[i] # Download the spectrum spec_name = \ download_spectra(spec_info['PLATE'], spec_info['FIBERID'], spec_info['MJD'], spec_info['SURVEY']) spec_df = do_specfit(spec_name, verbose=False) if i == 0: df = DataFrame(spec_df, columns=[spec_name[:-5]]) else: df[spec_name[:-5]] = spec_df if not keep_spectra: os.system('rm ' + spec_name) df.to_csv(output_file) return
# ... existing code ... from astropy.io import fits from pandas import DataFrame # ... modified code ... spec_name = \ download_spectra(spec_info['PLATE'], spec_info['FIBERID'], spec_info['MJD'], spec_info['SURVEY']) ... if i == 0: df = DataFrame(spec_df, columns=[spec_name[:-5]]) else: df[spec_name[:-5]] = spec_df ... df.to_csv(output_file) # ... rest of the code ...
f622e11536c4ebf8f82985329d06efc58c2fe60e
blog/tests/test_views.py
blog/tests/test_views.py
from django.test import TestCase class BlogViewsTestCase(TestCase): def setUp(self):
from django import test from django.core.urlresolvers import reverse from blog.models import Post, Category class BlogViewsTestCase(test.TestCase): def setUp(self): # Add parent category and post category parent = Category(name='Writing', parent=None) parent.save() category = Category(name='Thoughts', parent=parent) category.save() # Create a draft _post = Post(title='Random thoughts of the author', body='Thoughts turned to words.', category=category) _post.save() self.draft = _post # Publish a post post = Post(title='New thoughts from without', body='A post fit to be published!', category=category) post.save() post.publish() self.post = post self.client = test.Client() def test_index(self): response = self.client.get('/') self.assertEqual(response.status_code, 200) posts = response.context['posts'] self.assertNotIn(self.draft, posts) self.assertIn(self.post, posts) def test_post_view(self): post_url = reverse('blog:post', kwargs=dict(pid=self.post.id, slug=self.post.slug)) response = self.client.get(post_url) self.assertEqual(response.status_code, 200) post = response.context['post'] posts = response.context['posts'] self.assertEqual(post, self.post) self.assertEqual(posts.count(), 0) def test_draft_view(self): draft_url = reverse('blog:post', kwargs=dict(pid=self.draft.id, slug=self.draft.slug)) response = self.client.get(draft_url) self.assertEqual(response.status_code, 404)
Add tests for blog index view and post view
Add tests for blog index view and post view
Python
mit
ajoyoommen/weblog,ajoyoommen/weblog
- from django.test import TestCase + from django import test + from django.core.urlresolvers import reverse + + from blog.models import Post, Category - class BlogViewsTestCase(TestCase): + class BlogViewsTestCase(test.TestCase): def setUp(self): + # Add parent category and post category + parent = Category(name='Writing', parent=None) + parent.save() + category = Category(name='Thoughts', parent=parent) + category.save() + # Create a draft + _post = Post(title='Random thoughts of the author', + body='Thoughts turned to words.', category=category) + _post.save() + self.draft = _post + # Publish a post + post = Post(title='New thoughts from without', + body='A post fit to be published!', category=category) + post.save() + post.publish() + self.post = post + self.client = test.Client() + + def test_index(self): + response = self.client.get('/') + self.assertEqual(response.status_code, 200) + posts = response.context['posts'] + self.assertNotIn(self.draft, posts) + self.assertIn(self.post, posts) + + def test_post_view(self): + post_url = reverse('blog:post', kwargs=dict(pid=self.post.id, + slug=self.post.slug)) + response = self.client.get(post_url) + self.assertEqual(response.status_code, 200) + post = response.context['post'] + posts = response.context['posts'] + self.assertEqual(post, self.post) + self.assertEqual(posts.count(), 0) + + def test_draft_view(self): + draft_url = reverse('blog:post', kwargs=dict(pid=self.draft.id, + slug=self.draft.slug)) + response = self.client.get(draft_url) + self.assertEqual(response.status_code, 404) +
Add tests for blog index view and post view
## Code Before: from django.test import TestCase class BlogViewsTestCase(TestCase): def setUp(self): ## Instruction: Add tests for blog index view and post view ## Code After: from django import test from django.core.urlresolvers import reverse from blog.models import Post, Category class BlogViewsTestCase(test.TestCase): def setUp(self): # Add parent category and post category parent = Category(name='Writing', parent=None) parent.save() category = Category(name='Thoughts', parent=parent) category.save() # Create a draft _post = Post(title='Random thoughts of the author', body='Thoughts turned to words.', category=category) _post.save() self.draft = _post # Publish a post post = Post(title='New thoughts from without', body='A post fit to be published!', category=category) post.save() post.publish() self.post = post self.client = test.Client() def test_index(self): response = self.client.get('/') self.assertEqual(response.status_code, 200) posts = response.context['posts'] self.assertNotIn(self.draft, posts) self.assertIn(self.post, posts) def test_post_view(self): post_url = reverse('blog:post', kwargs=dict(pid=self.post.id, slug=self.post.slug)) response = self.client.get(post_url) self.assertEqual(response.status_code, 200) post = response.context['post'] posts = response.context['posts'] self.assertEqual(post, self.post) self.assertEqual(posts.count(), 0) def test_draft_view(self): draft_url = reverse('blog:post', kwargs=dict(pid=self.draft.id, slug=self.draft.slug)) response = self.client.get(draft_url) self.assertEqual(response.status_code, 404)
// ... existing code ... from django import test from django.core.urlresolvers import reverse from blog.models import Post, Category // ... modified code ... class BlogViewsTestCase(test.TestCase): def setUp(self): # Add parent category and post category parent = Category(name='Writing', parent=None) parent.save() category = Category(name='Thoughts', parent=parent) category.save() # Create a draft _post = Post(title='Random thoughts of the author', body='Thoughts turned to words.', category=category) _post.save() self.draft = _post # Publish a post post = Post(title='New thoughts from without', body='A post fit to be published!', category=category) post.save() post.publish() self.post = post self.client = test.Client() def test_index(self): response = self.client.get('/') self.assertEqual(response.status_code, 200) posts = response.context['posts'] self.assertNotIn(self.draft, posts) self.assertIn(self.post, posts) def test_post_view(self): post_url = reverse('blog:post', kwargs=dict(pid=self.post.id, slug=self.post.slug)) response = self.client.get(post_url) self.assertEqual(response.status_code, 200) post = response.context['post'] posts = response.context['posts'] self.assertEqual(post, self.post) self.assertEqual(posts.count(), 0) def test_draft_view(self): draft_url = reverse('blog:post', kwargs=dict(pid=self.draft.id, slug=self.draft.slug)) response = self.client.get(draft_url) self.assertEqual(response.status_code, 404) // ... rest of the code ...
edf151feea948ebf4a9f00a0248ab1f363cacfac
scaffolder/commands/install.py
scaffolder/commands/install.py
from optparse import make_option from optparse import OptionParser from scaffolder import get_minion_path from scaffolder.core.template import TemplateManager from scaffolder.core.commands import BaseCommand class InstallCommand(BaseCommand): option_list = BaseCommand.option_list + ( make_option( "-t", "--target", dest="target_dir", default=get_minion_path('weaver'), help='Project Templates directory.', metavar="TEMPLATES_DIR" ), ) def __init__(self, name, help='', aliases=(), stdout=None, stderr=None): help = 'install: Installs a Project Template.' parser = OptionParser( version=self.get_version(), option_list=self.get_option_list(), usage='\n %prog {0} ACTION [OPTIONS]'.format(name) ) aliases = ('tmp',) BaseCommand.__init__(self, name, parser=parser, help=help, aliases=aliases) def run(self, *args, **options): src = args[0] tgt = options.get('target_dir') manager = TemplateManager() manager.install(src=src, dest=tgt)
from optparse import make_option from optparse import OptionParser from scaffolder import get_minion_path from scaffolder.core.template import TemplateManager from scaffolder.core.commands import BaseCommand class InstallCommand(BaseCommand): option_list = BaseCommand.option_list + ( make_option( "-t", "--target", dest="target_dir", default=get_minion_path('weaver'), help='Project Templates directory.', metavar="TEMPLATES_DIR" ), ) help = 'Installs a Project Template.' def run(self, *args, **options): src = args[0] tgt = options.get('target_dir') manager = TemplateManager() manager.install(src=src, dest=tgt)
Remove __init__ method, not needed.
InstallCommand: Remove __init__ method, not needed.
Python
mit
goliatone/minions
from optparse import make_option from optparse import OptionParser from scaffolder import get_minion_path from scaffolder.core.template import TemplateManager from scaffolder.core.commands import BaseCommand class InstallCommand(BaseCommand): option_list = BaseCommand.option_list + ( make_option( "-t", "--target", dest="target_dir", default=get_minion_path('weaver'), help='Project Templates directory.', metavar="TEMPLATES_DIR" ), ) - def __init__(self, name, help='', aliases=(), stdout=None, stderr=None): - help = 'install: Installs a Project Template.' + help = 'Installs a Project Template.' + - parser = OptionParser( - version=self.get_version(), - option_list=self.get_option_list(), - usage='\n %prog {0} ACTION [OPTIONS]'.format(name) - ) - aliases = ('tmp',) - BaseCommand.__init__(self, name, parser=parser, help=help, aliases=aliases) def run(self, *args, **options): src = args[0] tgt = options.get('target_dir') manager = TemplateManager() manager.install(src=src, dest=tgt)
Remove __init__ method, not needed.
## Code Before: from optparse import make_option from optparse import OptionParser from scaffolder import get_minion_path from scaffolder.core.template import TemplateManager from scaffolder.core.commands import BaseCommand class InstallCommand(BaseCommand): option_list = BaseCommand.option_list + ( make_option( "-t", "--target", dest="target_dir", default=get_minion_path('weaver'), help='Project Templates directory.', metavar="TEMPLATES_DIR" ), ) def __init__(self, name, help='', aliases=(), stdout=None, stderr=None): help = 'install: Installs a Project Template.' parser = OptionParser( version=self.get_version(), option_list=self.get_option_list(), usage='\n %prog {0} ACTION [OPTIONS]'.format(name) ) aliases = ('tmp',) BaseCommand.__init__(self, name, parser=parser, help=help, aliases=aliases) def run(self, *args, **options): src = args[0] tgt = options.get('target_dir') manager = TemplateManager() manager.install(src=src, dest=tgt) ## Instruction: Remove __init__ method, not needed. ## Code After: from optparse import make_option from optparse import OptionParser from scaffolder import get_minion_path from scaffolder.core.template import TemplateManager from scaffolder.core.commands import BaseCommand class InstallCommand(BaseCommand): option_list = BaseCommand.option_list + ( make_option( "-t", "--target", dest="target_dir", default=get_minion_path('weaver'), help='Project Templates directory.', metavar="TEMPLATES_DIR" ), ) help = 'Installs a Project Template.' def run(self, *args, **options): src = args[0] tgt = options.get('target_dir') manager = TemplateManager() manager.install(src=src, dest=tgt)
// ... existing code ... help = 'Installs a Project Template.' // ... rest of the code ...
796f9ff27f579557237c48196eb50d40269c6840
glitch/__main__.py
glitch/__main__.py
from . import config from . import apikeys import argparse import logging parser = argparse.ArgumentParser(description="Invoke the Infinite Glitch server(s)") parser.add_argument("server", help="Server to invoke", choices=["main", "renderer"], nargs="?", default="main") parser.add_argument("-l", "--log", help="Logging level", type=lambda x: x.upper(), choices=logging._nameToLevel, # NAUGHTY default="INFO") parser.add_argument("--dev", help="Dev mode (no logins)", action='store_true') arguments = parser.parse_args() log = logging.getLogger(__name__) logging.basicConfig(level=getattr(logging, arguments.log), format='%(asctime)s:%(levelname)s:%(name)s:%(message)s') if arguments.server == "renderer": from . import renderer renderer.run() # doesn't return else: from . import server server.run(disable_logins=arguments.dev) # doesn't return
from . import config from . import apikeys import argparse # Hack: Allow "python -m glitch database" to be the same as "glitch.database" import sys if len(sys.argv) > 1 and sys.argv[1] == "database": from . import database import clize sys.exit(clize.run(*database.commands, args=sys.argv[1:])) import logging parser = argparse.ArgumentParser(description="Invoke the Infinite Glitch server(s)") parser.add_argument("server", help="Server to invoke", choices=["main", "renderer"], nargs="?", default="main") parser.add_argument("-l", "--log", help="Logging level", type=lambda x: x.upper(), choices=logging._nameToLevel, # NAUGHTY default="INFO") parser.add_argument("--dev", help="Dev mode (no logins)", action='store_true') arguments = parser.parse_args() log = logging.getLogger(__name__) logging.basicConfig(level=getattr(logging, arguments.log), format='%(asctime)s:%(levelname)s:%(name)s:%(message)s') if arguments.server == "renderer": from . import renderer renderer.run() # doesn't return else: from . import server server.run(disable_logins=arguments.dev) # doesn't return
Allow 'python -m glitch database' as well as with a dot
Allow 'python -m glitch database' as well as with a dot
Python
artistic-2.0
MikeiLL/appension,MikeiLL/appension,Rosuav/appension,Rosuav/appension,MikeiLL/appension,MikeiLL/appension,Rosuav/appension,Rosuav/appension
from . import config from . import apikeys import argparse + + # Hack: Allow "python -m glitch database" to be the same as "glitch.database" + import sys + if len(sys.argv) > 1 and sys.argv[1] == "database": + from . import database + import clize + sys.exit(clize.run(*database.commands, args=sys.argv[1:])) import logging parser = argparse.ArgumentParser(description="Invoke the Infinite Glitch server(s)") parser.add_argument("server", help="Server to invoke", choices=["main", "renderer"], nargs="?", default="main") parser.add_argument("-l", "--log", help="Logging level", type=lambda x: x.upper(), choices=logging._nameToLevel, # NAUGHTY default="INFO") parser.add_argument("--dev", help="Dev mode (no logins)", action='store_true') arguments = parser.parse_args() log = logging.getLogger(__name__) logging.basicConfig(level=getattr(logging, arguments.log), format='%(asctime)s:%(levelname)s:%(name)s:%(message)s') if arguments.server == "renderer": from . import renderer renderer.run() # doesn't return else: from . import server server.run(disable_logins=arguments.dev) # doesn't return
Allow 'python -m glitch database' as well as with a dot
## Code Before: from . import config from . import apikeys import argparse import logging parser = argparse.ArgumentParser(description="Invoke the Infinite Glitch server(s)") parser.add_argument("server", help="Server to invoke", choices=["main", "renderer"], nargs="?", default="main") parser.add_argument("-l", "--log", help="Logging level", type=lambda x: x.upper(), choices=logging._nameToLevel, # NAUGHTY default="INFO") parser.add_argument("--dev", help="Dev mode (no logins)", action='store_true') arguments = parser.parse_args() log = logging.getLogger(__name__) logging.basicConfig(level=getattr(logging, arguments.log), format='%(asctime)s:%(levelname)s:%(name)s:%(message)s') if arguments.server == "renderer": from . import renderer renderer.run() # doesn't return else: from . import server server.run(disable_logins=arguments.dev) # doesn't return ## Instruction: Allow 'python -m glitch database' as well as with a dot ## Code After: from . import config from . import apikeys import argparse # Hack: Allow "python -m glitch database" to be the same as "glitch.database" import sys if len(sys.argv) > 1 and sys.argv[1] == "database": from . import database import clize sys.exit(clize.run(*database.commands, args=sys.argv[1:])) import logging parser = argparse.ArgumentParser(description="Invoke the Infinite Glitch server(s)") parser.add_argument("server", help="Server to invoke", choices=["main", "renderer"], nargs="?", default="main") parser.add_argument("-l", "--log", help="Logging level", type=lambda x: x.upper(), choices=logging._nameToLevel, # NAUGHTY default="INFO") parser.add_argument("--dev", help="Dev mode (no logins)", action='store_true') arguments = parser.parse_args() log = logging.getLogger(__name__) logging.basicConfig(level=getattr(logging, arguments.log), format='%(asctime)s:%(levelname)s:%(name)s:%(message)s') if arguments.server == "renderer": from . import renderer renderer.run() # doesn't return else: from . import server server.run(disable_logins=arguments.dev) # doesn't return
... import argparse # Hack: Allow "python -m glitch database" to be the same as "glitch.database" import sys if len(sys.argv) > 1 and sys.argv[1] == "database": from . import database import clize sys.exit(clize.run(*database.commands, args=sys.argv[1:])) ...
4b93e5aa8c0ce90189fb852e75ee213d3be0d01a
flicks/base/urls.py
flicks/base/urls.py
from django.conf.urls.defaults import patterns, url from flicks.base import views urlpatterns = patterns('', url(r'^/?$', views.home, name='flicks.base.home'), url(r'^strings/?$', views.strings, name='flicks.base.strings'), )
from django.conf.urls.defaults import patterns, url from flicks.base import views urlpatterns = patterns('', url(r'^/?$', views.home, name='flicks.base.home'), url(r'^faq/?$', views.faq, name='flicks.base.faq'), url(r'^strings/?$', views.strings, name='flicks.base.strings'), )
Add back in FAQ url that was removed accidentally.
Add back in FAQ url that was removed accidentally.
Python
bsd-3-clause
mozilla/firefox-flicks,mozilla/firefox-flicks,mozilla/firefox-flicks,mozilla/firefox-flicks
from django.conf.urls.defaults import patterns, url from flicks.base import views urlpatterns = patterns('', url(r'^/?$', views.home, name='flicks.base.home'), + url(r'^faq/?$', views.faq, name='flicks.base.faq'), url(r'^strings/?$', views.strings, name='flicks.base.strings'), )
Add back in FAQ url that was removed accidentally.
## Code Before: from django.conf.urls.defaults import patterns, url from flicks.base import views urlpatterns = patterns('', url(r'^/?$', views.home, name='flicks.base.home'), url(r'^strings/?$', views.strings, name='flicks.base.strings'), ) ## Instruction: Add back in FAQ url that was removed accidentally. ## Code After: from django.conf.urls.defaults import patterns, url from flicks.base import views urlpatterns = patterns('', url(r'^/?$', views.home, name='flicks.base.home'), url(r'^faq/?$', views.faq, name='flicks.base.faq'), url(r'^strings/?$', views.strings, name='flicks.base.strings'), )
# ... existing code ... url(r'^/?$', views.home, name='flicks.base.home'), url(r'^faq/?$', views.faq, name='flicks.base.faq'), url(r'^strings/?$', views.strings, name='flicks.base.strings'), # ... rest of the code ...
9df3f3a2d0660b8e8166aa944bf45f261a51d987
ies_base/serializers.py
ies_base/serializers.py
from rest_framework import serializers class Tag(object): def __init__(self, name, related_tags, equivalent_names): self.name = name self.related_tags = related_tags self.equivalent_names = equivalent_names class TagSerializer(serializers.Serializer): name = serializers.CharField() related_tags = serializers.ListField(child=serializers.CharField(), allow_null=True) equivalent_names = serializers.ListField(child=serializers.CharField(), allow_null=True) def create(self, validated_data): return Tag(**validated_data) class Followable(object): def __init__(self, name, type, game, object_id, thumbnail_url="", **kwargs): self.name = name self.type = type self.game = game self.object_id = object_id self.thumbnail_url = thumbnail_url for field, value in kwargs.items(): setattr(self, field, value) class FollowableSerializer(serializers.Serializer): name = serializers.CharField() type = serializers.IntegerField() game = serializers.IntegerField() object_id = serializers.IntegerField() default_color = serializers.CharField(allow_null=True, allow_blank=True) thumbnail_url = serializers.CharField(allow_blank=True, allow_null=True) def create(self, validated_data): return Followable(**validated_data)
from rest_framework import serializers class Tag(object): def __init__(self, name, related_tags, equivalent_names): self.name = name self.related_tags = related_tags self.equivalent_names = equivalent_names class TagSerializer(serializers.Serializer): name = serializers.CharField() related_tags = serializers.ListField(child=serializers.CharField(), allow_null=True) equivalent_names = serializers.ListField(child=serializers.CharField(), allow_null=True) def create(self, validated_data): return Tag(**validated_data) class Followable(object): def __init__(self, name, type, game, object_id, thumbnail_url="", **kwargs): self.name = name self.type = type self.game = game self.object_id = object_id self.thumbnail_url = thumbnail_url for field, value in kwargs.items(): setattr(self, field, value) class FollowableSerializer(serializers.Serializer): name = serializers.CharField() type = serializers.IntegerField() game = serializers.IntegerField() object_id = serializers.IntegerField() default_color = serializers.CharField(allow_null=True, allow_blank=True, required=False) thumbnail_url = serializers.CharField(allow_blank=True, allow_null=True) def create(self, validated_data): return Followable(**validated_data)
Make default color not required
Make default color not required
Python
mit
InstanteSports/ies-django-base
from rest_framework import serializers class Tag(object): def __init__(self, name, related_tags, equivalent_names): self.name = name self.related_tags = related_tags self.equivalent_names = equivalent_names class TagSerializer(serializers.Serializer): name = serializers.CharField() related_tags = serializers.ListField(child=serializers.CharField(), allow_null=True) equivalent_names = serializers.ListField(child=serializers.CharField(), allow_null=True) def create(self, validated_data): return Tag(**validated_data) class Followable(object): def __init__(self, name, type, game, object_id, thumbnail_url="", **kwargs): self.name = name self.type = type self.game = game self.object_id = object_id self.thumbnail_url = thumbnail_url for field, value in kwargs.items(): setattr(self, field, value) class FollowableSerializer(serializers.Serializer): name = serializers.CharField() type = serializers.IntegerField() game = serializers.IntegerField() object_id = serializers.IntegerField() - default_color = serializers.CharField(allow_null=True, allow_blank=True) + default_color = serializers.CharField(allow_null=True, allow_blank=True, required=False) thumbnail_url = serializers.CharField(allow_blank=True, allow_null=True) def create(self, validated_data): return Followable(**validated_data)
Make default color not required
## Code Before: from rest_framework import serializers class Tag(object): def __init__(self, name, related_tags, equivalent_names): self.name = name self.related_tags = related_tags self.equivalent_names = equivalent_names class TagSerializer(serializers.Serializer): name = serializers.CharField() related_tags = serializers.ListField(child=serializers.CharField(), allow_null=True) equivalent_names = serializers.ListField(child=serializers.CharField(), allow_null=True) def create(self, validated_data): return Tag(**validated_data) class Followable(object): def __init__(self, name, type, game, object_id, thumbnail_url="", **kwargs): self.name = name self.type = type self.game = game self.object_id = object_id self.thumbnail_url = thumbnail_url for field, value in kwargs.items(): setattr(self, field, value) class FollowableSerializer(serializers.Serializer): name = serializers.CharField() type = serializers.IntegerField() game = serializers.IntegerField() object_id = serializers.IntegerField() default_color = serializers.CharField(allow_null=True, allow_blank=True) thumbnail_url = serializers.CharField(allow_blank=True, allow_null=True) def create(self, validated_data): return Followable(**validated_data) ## Instruction: Make default color not required ## Code After: from rest_framework import serializers class Tag(object): def __init__(self, name, related_tags, equivalent_names): self.name = name self.related_tags = related_tags self.equivalent_names = equivalent_names class TagSerializer(serializers.Serializer): name = serializers.CharField() related_tags = serializers.ListField(child=serializers.CharField(), allow_null=True) equivalent_names = serializers.ListField(child=serializers.CharField(), allow_null=True) def create(self, validated_data): return Tag(**validated_data) class Followable(object): def __init__(self, name, type, game, object_id, thumbnail_url="", **kwargs): self.name = name self.type = type self.game = game self.object_id = object_id self.thumbnail_url = thumbnail_url for field, value in kwargs.items(): setattr(self, field, value) class FollowableSerializer(serializers.Serializer): name = serializers.CharField() type = serializers.IntegerField() game = serializers.IntegerField() object_id = serializers.IntegerField() default_color = serializers.CharField(allow_null=True, allow_blank=True, required=False) thumbnail_url = serializers.CharField(allow_blank=True, allow_null=True) def create(self, validated_data): return Followable(**validated_data)
// ... existing code ... object_id = serializers.IntegerField() default_color = serializers.CharField(allow_null=True, allow_blank=True, required=False) thumbnail_url = serializers.CharField(allow_blank=True, allow_null=True) // ... rest of the code ...
7519bebe1d9d87930275858a537dcc0a0a64f007
tools/strip_filenames.py
tools/strip_filenames.py
import os directory = os.listdir() illegal_characters = "%?_'*+$!\"" tolowercase=True for a in range(len(directory)): newname="" for c in directory[a]: if c in illegal_characters: continue if c.isalnum() or c == '.': newname=newname+c.lower() print("convert {} to {}".format(directory[a],newname)) os.rename(directory[a], newname)
import sys import os from docopt import docopt # docopt(doc, argv=None, help=True, version=None, options_first=False)) def main(): opt = docopt(__doc__, sys.argv[1:]) directory = opt.get("filename", os.listdir()) legal_characters = "" list_N010 = list(range(size)) list_alpha = [ chr(x+97) for x in range(26) ] list_ALPHA = [ chr(x+65) for x in range(26) ] legal_characters += "".join(list_N010) legal_characters += "".join(list_alpha) if not opt.get("--lowercase", False): legal_characters += "".join(list_N010) for a in range(len(directory)): newname="" for c in directory[a]: if c not in legal_characters: continue newname += c print("convert {} to {}".format(directory[a],newname)) os.rename(directory[a], newname) if __name__ == "__main__": main()
Use legal characters for stripping filenames
Use legal characters for stripping filenames
Python
mit
dgengtek/scripts,dgengtek/scripts
+ + import sys import os + from docopt import docopt - directory = os.listdir() - illegal_characters = "%?_'*+$!\"" - tolowercase=True - for a in range(len(directory)): - newname="" - for c in directory[a]: - if c in illegal_characters: - continue - if c.isalnum() or c == '.': - newname=newname+c.lower() - print("convert {} to {}".format(directory[a],newname)) - os.rename(directory[a], newname) + # docopt(doc, argv=None, help=True, version=None, options_first=False)) + + def main(): + opt = docopt(__doc__, sys.argv[1:]) + directory = opt.get("filename", os.listdir()) + legal_characters = "" + list_N010 = list(range(size)) + list_alpha = [ chr(x+97) for x in range(26) ] + list_ALPHA = [ chr(x+65) for x in range(26) ] + + legal_characters += "".join(list_N010) + legal_characters += "".join(list_alpha) + if not opt.get("--lowercase", False): + legal_characters += "".join(list_N010) + for a in range(len(directory)): + newname="" + for c in directory[a]: + if c not in legal_characters: + continue + newname += c + print("convert {} to {}".format(directory[a],newname)) + os.rename(directory[a], newname) + + if __name__ == "__main__": + main() +
Use legal characters for stripping filenames
## Code Before: import os directory = os.listdir() illegal_characters = "%?_'*+$!\"" tolowercase=True for a in range(len(directory)): newname="" for c in directory[a]: if c in illegal_characters: continue if c.isalnum() or c == '.': newname=newname+c.lower() print("convert {} to {}".format(directory[a],newname)) os.rename(directory[a], newname) ## Instruction: Use legal characters for stripping filenames ## Code After: import sys import os from docopt import docopt # docopt(doc, argv=None, help=True, version=None, options_first=False)) def main(): opt = docopt(__doc__, sys.argv[1:]) directory = opt.get("filename", os.listdir()) legal_characters = "" list_N010 = list(range(size)) list_alpha = [ chr(x+97) for x in range(26) ] list_ALPHA = [ chr(x+65) for x in range(26) ] legal_characters += "".join(list_N010) legal_characters += "".join(list_alpha) if not opt.get("--lowercase", False): legal_characters += "".join(list_N010) for a in range(len(directory)): newname="" for c in directory[a]: if c not in legal_characters: continue newname += c print("convert {} to {}".format(directory[a],newname)) os.rename(directory[a], newname) if __name__ == "__main__": main()
// ... existing code ... import sys import os from docopt import docopt # docopt(doc, argv=None, help=True, version=None, options_first=False)) def main(): opt = docopt(__doc__, sys.argv[1:]) directory = opt.get("filename", os.listdir()) legal_characters = "" list_N010 = list(range(size)) list_alpha = [ chr(x+97) for x in range(26) ] list_ALPHA = [ chr(x+65) for x in range(26) ] legal_characters += "".join(list_N010) legal_characters += "".join(list_alpha) if not opt.get("--lowercase", False): legal_characters += "".join(list_N010) for a in range(len(directory)): newname="" for c in directory[a]: if c not in legal_characters: continue newname += c print("convert {} to {}".format(directory[a],newname)) os.rename(directory[a], newname) if __name__ == "__main__": main() // ... rest of the code ...
4f8aed6ed3491e62911619eaa9aa4b86b30065e4
leonardo/module/leonardo_auth/widget/userlogin/models.py
leonardo/module/leonardo_auth/widget/userlogin/models.py
from django.db import models from django.utils.translation import ugettext_lazy as _ from leonardo.module.web.models import Widget LOGIN_TYPE_CHOICES = ( (1, _("Admin")), (2, _("Public")), ) class UserLoginWidget(Widget): type = models.PositiveIntegerField(verbose_name=_( "type"), choices=LOGIN_TYPE_CHOICES, default=2) def get_context_data(self, request): context = super(UserLoginWidget, self).get_context_data(request) if 'next' in request.GET: context['next'] = request.GET['next'] return context class Meta: abstract = True verbose_name = _("user login") verbose_name_plural = _("user logins")
from django.db import models from django.utils.translation import ugettext_lazy as _ from leonardo.module.web.models import Widget LOGIN_TYPE_CHOICES = ( (1, _("Admin")), (2, _("Public")), ) class UserLoginWidget(Widget): type = models.PositiveIntegerField(verbose_name=_( "type"), choices=LOGIN_TYPE_CHOICES, default=2) def get_context_data(self, request): context = super(UserLoginWidget, self).get_context_data(request) if 'next' in request.GET: context['next'] = request.GET['next'] else: context['next'] = request.path return context class Meta: abstract = True verbose_name = _("user login") verbose_name_plural = _("user logins")
Fix missing next in context.
Fix missing next in context.
Python
bsd-3-clause
django-leonardo/django-leonardo,django-leonardo/django-leonardo,django-leonardo/django-leonardo,django-leonardo/django-leonardo
from django.db import models from django.utils.translation import ugettext_lazy as _ from leonardo.module.web.models import Widget LOGIN_TYPE_CHOICES = ( (1, _("Admin")), (2, _("Public")), ) class UserLoginWidget(Widget): type = models.PositiveIntegerField(verbose_name=_( "type"), choices=LOGIN_TYPE_CHOICES, default=2) def get_context_data(self, request): context = super(UserLoginWidget, self).get_context_data(request) if 'next' in request.GET: context['next'] = request.GET['next'] + else: + context['next'] = request.path return context class Meta: abstract = True verbose_name = _("user login") verbose_name_plural = _("user logins")
Fix missing next in context.
## Code Before: from django.db import models from django.utils.translation import ugettext_lazy as _ from leonardo.module.web.models import Widget LOGIN_TYPE_CHOICES = ( (1, _("Admin")), (2, _("Public")), ) class UserLoginWidget(Widget): type = models.PositiveIntegerField(verbose_name=_( "type"), choices=LOGIN_TYPE_CHOICES, default=2) def get_context_data(self, request): context = super(UserLoginWidget, self).get_context_data(request) if 'next' in request.GET: context['next'] = request.GET['next'] return context class Meta: abstract = True verbose_name = _("user login") verbose_name_plural = _("user logins") ## Instruction: Fix missing next in context. ## Code After: from django.db import models from django.utils.translation import ugettext_lazy as _ from leonardo.module.web.models import Widget LOGIN_TYPE_CHOICES = ( (1, _("Admin")), (2, _("Public")), ) class UserLoginWidget(Widget): type = models.PositiveIntegerField(verbose_name=_( "type"), choices=LOGIN_TYPE_CHOICES, default=2) def get_context_data(self, request): context = super(UserLoginWidget, self).get_context_data(request) if 'next' in request.GET: context['next'] = request.GET['next'] else: context['next'] = request.path return context class Meta: abstract = True verbose_name = _("user login") verbose_name_plural = _("user logins")
... context['next'] = request.GET['next'] else: context['next'] = request.path ...
b6ab579fa65f816704142716fbd68645ac5f2ff8
zenaida/contrib/feedback/models.py
zenaida/contrib/feedback/models.py
from django.conf import settings from django.db import models class FeedbackItem(models.Model): timestamp = models.DateTimeField(auto_now_add=True) user = models.ForeignKey(settings.AUTH_USER_MODEL) resolved = models.BooleanField(default=False) content = models.TextField() screenshot = models.FileField(blank=True, null=True, upload_to="feedback/screenshots") # Request Data view = models.CharField(max_length=255) request_path = models.CharField(max_length=255) # The longest methods should be 7 chars, but we'll allow custom methods up # to 20 chars just in case. request_method = models.CharField(max_length=20, blank=True, null=True) # How long is the longest encoding name? request_encoding = models.CharField(max_length=20, blank=True, null=True) request_meta = models.TextField(blank=True, null=True) request_get = models.TextField(blank=True, null=True) request_post = models.TextField(blank=True, null=True) request_files = models.TextField(blank=True, null=True) def __unicode__(self): return "{username} at {path}".format( username=self.user.get_full_name(), path = self.request_path )
from django.conf import settings from django.db import models class FeedbackItem(models.Model): timestamp = models.DateTimeField(auto_now_add=True) user = models.ForeignKey(settings.AUTH_USER_MODEL) resolved = models.BooleanField(default=False) content = models.TextField() screenshot = models.FileField(blank=True, null=True, upload_to="feedback/screenshots") # Request Data view = models.CharField(max_length=255) request_path = models.CharField(max_length=255) # The longest methods should be 7 chars, but we'll allow custom methods up # to 20 chars just in case. request_method = models.CharField(max_length=20, blank=True, null=True) # How long is the longest encoding name? request_encoding = models.CharField(max_length=20, blank=True, null=True) request_meta = models.TextField(blank=True, null=True) request_get = models.TextField(blank=True, null=True) request_post = models.TextField(blank=True, null=True) request_files = models.TextField(blank=True, null=True) def __unicode__(self): return "{username} at {path}".format( username=self.user.get_full_name(), path = self.request_path ) class Meta: ordering = ["-timestamp"]
Order feedback items by their timestamp.
Order feedback items by their timestamp.
Python
bsd-3-clause
littleweaver/django-zenaida,littleweaver/django-zenaida,littleweaver/django-zenaida,littleweaver/django-zenaida
from django.conf import settings from django.db import models class FeedbackItem(models.Model): timestamp = models.DateTimeField(auto_now_add=True) user = models.ForeignKey(settings.AUTH_USER_MODEL) resolved = models.BooleanField(default=False) content = models.TextField() screenshot = models.FileField(blank=True, null=True, upload_to="feedback/screenshots") # Request Data view = models.CharField(max_length=255) request_path = models.CharField(max_length=255) # The longest methods should be 7 chars, but we'll allow custom methods up # to 20 chars just in case. request_method = models.CharField(max_length=20, blank=True, null=True) # How long is the longest encoding name? request_encoding = models.CharField(max_length=20, blank=True, null=True) request_meta = models.TextField(blank=True, null=True) request_get = models.TextField(blank=True, null=True) request_post = models.TextField(blank=True, null=True) request_files = models.TextField(blank=True, null=True) def __unicode__(self): return "{username} at {path}".format( username=self.user.get_full_name(), path = self.request_path ) + class Meta: + ordering = ["-timestamp"] +
Order feedback items by their timestamp.
## Code Before: from django.conf import settings from django.db import models class FeedbackItem(models.Model): timestamp = models.DateTimeField(auto_now_add=True) user = models.ForeignKey(settings.AUTH_USER_MODEL) resolved = models.BooleanField(default=False) content = models.TextField() screenshot = models.FileField(blank=True, null=True, upload_to="feedback/screenshots") # Request Data view = models.CharField(max_length=255) request_path = models.CharField(max_length=255) # The longest methods should be 7 chars, but we'll allow custom methods up # to 20 chars just in case. request_method = models.CharField(max_length=20, blank=True, null=True) # How long is the longest encoding name? request_encoding = models.CharField(max_length=20, blank=True, null=True) request_meta = models.TextField(blank=True, null=True) request_get = models.TextField(blank=True, null=True) request_post = models.TextField(blank=True, null=True) request_files = models.TextField(blank=True, null=True) def __unicode__(self): return "{username} at {path}".format( username=self.user.get_full_name(), path = self.request_path ) ## Instruction: Order feedback items by their timestamp. ## Code After: from django.conf import settings from django.db import models class FeedbackItem(models.Model): timestamp = models.DateTimeField(auto_now_add=True) user = models.ForeignKey(settings.AUTH_USER_MODEL) resolved = models.BooleanField(default=False) content = models.TextField() screenshot = models.FileField(blank=True, null=True, upload_to="feedback/screenshots") # Request Data view = models.CharField(max_length=255) request_path = models.CharField(max_length=255) # The longest methods should be 7 chars, but we'll allow custom methods up # to 20 chars just in case. request_method = models.CharField(max_length=20, blank=True, null=True) # How long is the longest encoding name? request_encoding = models.CharField(max_length=20, blank=True, null=True) request_meta = models.TextField(blank=True, null=True) request_get = models.TextField(blank=True, null=True) request_post = models.TextField(blank=True, null=True) request_files = models.TextField(blank=True, null=True) def __unicode__(self): return "{username} at {path}".format( username=self.user.get_full_name(), path = self.request_path ) class Meta: ordering = ["-timestamp"]
... ) class Meta: ordering = ["-timestamp"] ...
29cc59bc478c4c6bc936141d19a3386468ff8f07
tests/test_general_attributes.py
tests/test_general_attributes.py
from jawa.attribute import get_attribute_classes def test_mandatory_attributes(): for parser_class in get_attribute_classes().values(): assert hasattr(parser_class, 'ADDED_IN'), ( 'Attribute parser missing mandatory ADDED_IN property' ) assert hasattr(parser_class, 'MINIMUM_CLASS_VERSION'), ( 'Attribute parser missing mandatory MINIMUM_CLASS_VERSION ' 'property' )
from jawa.attribute import get_attribute_classes def test_mandatory_attributes(): required_properities = ['ADDED_IN', 'MINIMUM_CLASS_VERSION'] for name, class_ in get_attribute_classes().items(): for p in required_properities: assert hasattr(class_, p), ( '{name} parser missing mandatory {p} property'.format( name=name, p=p ) ) def test_attribute_naming(): for name, class_ in get_attribute_classes().items(): if hasattr(class_, 'ATTRIBUTE_NAME'): continue assert class_.__name__.endswith('Attribute'), ( '{name} parser does not follow naming convention and does' ' not explicity set it.'.format(name=name) )
Add a simple test for Attribuet class naming conventions.
Add a simple test for Attribuet class naming conventions.
Python
mit
TkTech/Jawa,TkTech/Jawa
from jawa.attribute import get_attribute_classes def test_mandatory_attributes(): + required_properities = ['ADDED_IN', 'MINIMUM_CLASS_VERSION'] - for parser_class in get_attribute_classes().values(): + for name, class_ in get_attribute_classes().items(): - assert hasattr(parser_class, 'ADDED_IN'), ( - 'Attribute parser missing mandatory ADDED_IN property' + for p in required_properities: + assert hasattr(class_, p), ( + '{name} parser missing mandatory {p} property'.format( + name=name, + p=p + ) - ) + ) - assert hasattr(parser_class, 'MINIMUM_CLASS_VERSION'), ( - 'Attribute parser missing mandatory MINIMUM_CLASS_VERSION ' - 'property' + + + def test_attribute_naming(): + for name, class_ in get_attribute_classes().items(): + if hasattr(class_, 'ATTRIBUTE_NAME'): + continue + + assert class_.__name__.endswith('Attribute'), ( + '{name} parser does not follow naming convention and does' + ' not explicity set it.'.format(name=name) )
Add a simple test for Attribuet class naming conventions.
## Code Before: from jawa.attribute import get_attribute_classes def test_mandatory_attributes(): for parser_class in get_attribute_classes().values(): assert hasattr(parser_class, 'ADDED_IN'), ( 'Attribute parser missing mandatory ADDED_IN property' ) assert hasattr(parser_class, 'MINIMUM_CLASS_VERSION'), ( 'Attribute parser missing mandatory MINIMUM_CLASS_VERSION ' 'property' ) ## Instruction: Add a simple test for Attribuet class naming conventions. ## Code After: from jawa.attribute import get_attribute_classes def test_mandatory_attributes(): required_properities = ['ADDED_IN', 'MINIMUM_CLASS_VERSION'] for name, class_ in get_attribute_classes().items(): for p in required_properities: assert hasattr(class_, p), ( '{name} parser missing mandatory {p} property'.format( name=name, p=p ) ) def test_attribute_naming(): for name, class_ in get_attribute_classes().items(): if hasattr(class_, 'ATTRIBUTE_NAME'): continue assert class_.__name__.endswith('Attribute'), ( '{name} parser does not follow naming convention and does' ' not explicity set it.'.format(name=name) )
# ... existing code ... def test_mandatory_attributes(): required_properities = ['ADDED_IN', 'MINIMUM_CLASS_VERSION'] for name, class_ in get_attribute_classes().items(): for p in required_properities: assert hasattr(class_, p), ( '{name} parser missing mandatory {p} property'.format( name=name, p=p ) ) def test_attribute_naming(): for name, class_ in get_attribute_classes().items(): if hasattr(class_, 'ATTRIBUTE_NAME'): continue assert class_.__name__.endswith('Attribute'), ( '{name} parser does not follow naming convention and does' ' not explicity set it.'.format(name=name) ) # ... rest of the code ...
0a3eb4b966dff69cbe582c60bf4444facb4b683d
tcconfig/_tc_command_helper.py
tcconfig/_tc_command_helper.py
from __future__ import absolute_import, unicode_literals import subprocrunner as spr from ._common import find_bin_path from ._const import TcSubCommand from ._error import NetworkInterfaceNotFoundError def get_tc_base_command(tc_subcommand): if tc_subcommand not in TcSubCommand: raise ValueError("the argument must be a TcSubCommand value") return "{:s} {:s}".format(find_bin_path("tc"), tc_subcommand.value) def run_tc_show(subcommand, device, tc_command_output): from ._network import verify_network_interface verify_network_interface(device, tc_command_output) runner = spr.SubprocessRunner( "{:s} show dev {:s}".format(get_tc_base_command(subcommand), device) ) if runner.run() != 0 and runner.stderr.find("Cannot find device") != -1: # reach here if the device does not exist at the system and netiface # not installed. raise NetworkInterfaceNotFoundError(target=device) return runner.stdout
from __future__ import absolute_import, unicode_literals import subprocrunner as spr from ._common import find_bin_path from ._const import TcSubCommand from ._error import NetworkInterfaceNotFoundError def get_tc_base_command(tc_subcommand): if not isinstance(tc_subcommand, TcSubCommand): raise ValueError("the argument must be a TcSubCommand value") return "{:s} {:s}".format(find_bin_path("tc"), tc_subcommand.value) def run_tc_show(subcommand, device, tc_command_output): from ._network import verify_network_interface verify_network_interface(device, tc_command_output) runner = spr.SubprocessRunner( "{:s} show dev {:s}".format(get_tc_base_command(subcommand), device) ) if runner.run() != 0 and runner.stderr.find("Cannot find device") != -1: # reach here if the device does not exist at the system and netiface # not installed. raise NetworkInterfaceNotFoundError(target=device) return runner.stdout
Change to avoid a DeprecationWarning
Change to avoid a DeprecationWarning
Python
mit
thombashi/tcconfig,thombashi/tcconfig
from __future__ import absolute_import, unicode_literals import subprocrunner as spr from ._common import find_bin_path from ._const import TcSubCommand from ._error import NetworkInterfaceNotFoundError def get_tc_base_command(tc_subcommand): - if tc_subcommand not in TcSubCommand: + if not isinstance(tc_subcommand, TcSubCommand): raise ValueError("the argument must be a TcSubCommand value") return "{:s} {:s}".format(find_bin_path("tc"), tc_subcommand.value) def run_tc_show(subcommand, device, tc_command_output): from ._network import verify_network_interface verify_network_interface(device, tc_command_output) runner = spr.SubprocessRunner( "{:s} show dev {:s}".format(get_tc_base_command(subcommand), device) ) if runner.run() != 0 and runner.stderr.find("Cannot find device") != -1: # reach here if the device does not exist at the system and netiface # not installed. raise NetworkInterfaceNotFoundError(target=device) return runner.stdout
Change to avoid a DeprecationWarning
## Code Before: from __future__ import absolute_import, unicode_literals import subprocrunner as spr from ._common import find_bin_path from ._const import TcSubCommand from ._error import NetworkInterfaceNotFoundError def get_tc_base_command(tc_subcommand): if tc_subcommand not in TcSubCommand: raise ValueError("the argument must be a TcSubCommand value") return "{:s} {:s}".format(find_bin_path("tc"), tc_subcommand.value) def run_tc_show(subcommand, device, tc_command_output): from ._network import verify_network_interface verify_network_interface(device, tc_command_output) runner = spr.SubprocessRunner( "{:s} show dev {:s}".format(get_tc_base_command(subcommand), device) ) if runner.run() != 0 and runner.stderr.find("Cannot find device") != -1: # reach here if the device does not exist at the system and netiface # not installed. raise NetworkInterfaceNotFoundError(target=device) return runner.stdout ## Instruction: Change to avoid a DeprecationWarning ## Code After: from __future__ import absolute_import, unicode_literals import subprocrunner as spr from ._common import find_bin_path from ._const import TcSubCommand from ._error import NetworkInterfaceNotFoundError def get_tc_base_command(tc_subcommand): if not isinstance(tc_subcommand, TcSubCommand): raise ValueError("the argument must be a TcSubCommand value") return "{:s} {:s}".format(find_bin_path("tc"), tc_subcommand.value) def run_tc_show(subcommand, device, tc_command_output): from ._network import verify_network_interface verify_network_interface(device, tc_command_output) runner = spr.SubprocessRunner( "{:s} show dev {:s}".format(get_tc_base_command(subcommand), device) ) if runner.run() != 0 and runner.stderr.find("Cannot find device") != -1: # reach here if the device does not exist at the system and netiface # not installed. raise NetworkInterfaceNotFoundError(target=device) return runner.stdout
# ... existing code ... def get_tc_base_command(tc_subcommand): if not isinstance(tc_subcommand, TcSubCommand): raise ValueError("the argument must be a TcSubCommand value") # ... rest of the code ...
817b597f3a45a8b16de84d480458a66499604f5a
owned_models/models.py
owned_models/models.py
from django.conf import settings from django.db import models class UserOwnedModelManager(models.Manager): def filter_for_user(self, user, *args, **kwargs): return super(UserOwnedModelManager, self).get_queryset().filter(user = user, *args, **kwargs) def get_for_user(self, user, *args, **kwargs): if 'user' in kwargs: kwargs.pop('user') return super(UserOwnedModelManager, self).get_queryset().get(user = user, *args, **kwargs) class UserOwnedModel(models.Model): user = models.ForeignKey(settings.AUTH_USER_MODEL, editable = False) objects = UserOwnedModelManager() class Meta: abstract = True
from django.conf import settings from django.db import models class UserOwnedModelManager(models.Manager): def filter_for_user(self, user, *args, **kwargs): return super(UserOwnedModelManager, self).get_queryset().filter(user = user, *args, **kwargs) def get_for_user(self, user, *args, **kwargs): if 'user' in kwargs: kwargs.pop('user') return super(UserOwnedModelManager, self).get_queryset().get(user = user, *args, **kwargs) def get_or_create_for_user(self, user, **kwargs): return super(UserOwnedModelManager, self).get_or_create(user = user, **kwargs) class UserOwnedModel(models.Model): user = models.ForeignKey(settings.AUTH_USER_MODEL, editable = False) objects = UserOwnedModelManager() class Meta: abstract = True
Add `get_or_create_for_user` method to default Manager.
Add `get_or_create_for_user` method to default Manager.
Python
mit
discolabs/django-owned-models
from django.conf import settings from django.db import models class UserOwnedModelManager(models.Manager): def filter_for_user(self, user, *args, **kwargs): return super(UserOwnedModelManager, self).get_queryset().filter(user = user, *args, **kwargs) def get_for_user(self, user, *args, **kwargs): if 'user' in kwargs: kwargs.pop('user') return super(UserOwnedModelManager, self).get_queryset().get(user = user, *args, **kwargs) + def get_or_create_for_user(self, user, **kwargs): + return super(UserOwnedModelManager, self).get_or_create(user = user, **kwargs) class UserOwnedModel(models.Model): user = models.ForeignKey(settings.AUTH_USER_MODEL, editable = False) objects = UserOwnedModelManager() class Meta: abstract = True
Add `get_or_create_for_user` method to default Manager.
## Code Before: from django.conf import settings from django.db import models class UserOwnedModelManager(models.Manager): def filter_for_user(self, user, *args, **kwargs): return super(UserOwnedModelManager, self).get_queryset().filter(user = user, *args, **kwargs) def get_for_user(self, user, *args, **kwargs): if 'user' in kwargs: kwargs.pop('user') return super(UserOwnedModelManager, self).get_queryset().get(user = user, *args, **kwargs) class UserOwnedModel(models.Model): user = models.ForeignKey(settings.AUTH_USER_MODEL, editable = False) objects = UserOwnedModelManager() class Meta: abstract = True ## Instruction: Add `get_or_create_for_user` method to default Manager. ## Code After: from django.conf import settings from django.db import models class UserOwnedModelManager(models.Manager): def filter_for_user(self, user, *args, **kwargs): return super(UserOwnedModelManager, self).get_queryset().filter(user = user, *args, **kwargs) def get_for_user(self, user, *args, **kwargs): if 'user' in kwargs: kwargs.pop('user') return super(UserOwnedModelManager, self).get_queryset().get(user = user, *args, **kwargs) def get_or_create_for_user(self, user, **kwargs): return super(UserOwnedModelManager, self).get_or_create(user = user, **kwargs) class UserOwnedModel(models.Model): user = models.ForeignKey(settings.AUTH_USER_MODEL, editable = False) objects = UserOwnedModelManager() class Meta: abstract = True
// ... existing code ... def get_or_create_for_user(self, user, **kwargs): return super(UserOwnedModelManager, self).get_or_create(user = user, **kwargs) // ... rest of the code ...
15c8215415d36da4fac9c7333e62239f7b81c12d
test/support/mock_definitions.py
test/support/mock_definitions.py
class MockDefinitions(object): def __init__(self, session_key=None): self.session_key = session_key if session_key is not None else '123456789' @property def metadata(self): host = os.getenv('SPLUNK_API_HOST', 'localhost') port = os.getenv('SPLUNK_API_PORT', 8089), return {'server_uri': 'https://{host}:{port}/', 'session_key': self.session_key, 'name': 'amp4e_events_test_input'}
import os # Generates validation/input definitions as if they were created by splunk for tests class MockDefinitions(object): def __init__(self, session_key=None): self.session_key = session_key if session_key is not None else '123456789' @property def metadata(self): host = os.getenv('SPLUNK_API_HOST', '127.0.0.1') return {'server_uri': 'https://{host}:8089/'.format(host=host), 'session_key': self.session_key, 'name': 'amp4e_events_test_input'}
Change mock to be env dependant
Change mock to be env dependant
Python
bsd-2-clause
Cisco-AMP/amp4e_splunk_events_input,Cisco-AMP/amp4e_splunk_events_input,Cisco-AMP/amp4e_splunk_events_input,Cisco-AMP/amp4e_splunk_events_input
+ import os + # Generates validation/input definitions as if they were created by splunk for tests class MockDefinitions(object): def __init__(self, session_key=None): self.session_key = session_key if session_key is not None else '123456789' @property def metadata(self): - host = os.getenv('SPLUNK_API_HOST', 'localhost') + host = os.getenv('SPLUNK_API_HOST', '127.0.0.1') - port = os.getenv('SPLUNK_API_PORT', 8089), - return {'server_uri': 'https://{host}:{port}/', 'session_key': self.session_key, + return {'server_uri': 'https://{host}:8089/'.format(host=host), 'session_key': self.session_key, 'name': 'amp4e_events_test_input'}
Change mock to be env dependant
## Code Before: class MockDefinitions(object): def __init__(self, session_key=None): self.session_key = session_key if session_key is not None else '123456789' @property def metadata(self): host = os.getenv('SPLUNK_API_HOST', 'localhost') port = os.getenv('SPLUNK_API_PORT', 8089), return {'server_uri': 'https://{host}:{port}/', 'session_key': self.session_key, 'name': 'amp4e_events_test_input'} ## Instruction: Change mock to be env dependant ## Code After: import os # Generates validation/input definitions as if they were created by splunk for tests class MockDefinitions(object): def __init__(self, session_key=None): self.session_key = session_key if session_key is not None else '123456789' @property def metadata(self): host = os.getenv('SPLUNK_API_HOST', '127.0.0.1') return {'server_uri': 'https://{host}:8089/'.format(host=host), 'session_key': self.session_key, 'name': 'amp4e_events_test_input'}
# ... existing code ... import os # Generates validation/input definitions as if they were created by splunk for tests class MockDefinitions(object): # ... modified code ... def metadata(self): host = os.getenv('SPLUNK_API_HOST', '127.0.0.1') return {'server_uri': 'https://{host}:8089/'.format(host=host), 'session_key': self.session_key, 'name': 'amp4e_events_test_input'} # ... rest of the code ...
1e182ec0fd7cf550c809f2e6792629caeb8d5553
sauce/lib/helpers.py
sauce/lib/helpers.py
from datetime import datetime from tg import url as tgurl #from webhelpers import date, feedgenerator, html, number, misc, text import webhelpers as w from webhelpers.html.tags import link_to from webhelpers.text import truncate from webhelpers.date import distance_of_time_in_words import re #log = logging.getLogger(__name__) cut = lambda text, max=200: truncate(text, max, whole_word=True) strftimedelta = lambda delta, granularity='minute': distance_of_time_in_words(datetime.now(), datetime.now()+delta, granularity) def link(label, url='', **attrs): return link_to(label, tgurl(url), **attrs) def striphtml(text): return re.sub('<[^<]+?>', ' ', text).strip()
from datetime import datetime from tg import url as tgurl #from webhelpers import date, feedgenerator, html, number, misc, text import webhelpers as w from webhelpers.html.tags import link_to from webhelpers.text import truncate from webhelpers.date import distance_of_time_in_words from genshi.core import striptags import re #log = logging.getLogger(__name__) cut = lambda text, max=200: truncate(text, max, whole_word=True) strftimedelta = lambda delta, granularity='minute': distance_of_time_in_words(datetime.now(), datetime.now()+delta, granularity) striphtml = striptags def link(label, url='', **attrs): return link_to(label, tgurl(url), **attrs)
Use striptags from genshi for striphtml, since we have to have genshi anyway
Use striptags from genshi for striphtml, since we have to have genshi anyway
Python
agpl-3.0
moschlar/SAUCE,moschlar/SAUCE,moschlar/SAUCE,moschlar/SAUCE
from datetime import datetime from tg import url as tgurl #from webhelpers import date, feedgenerator, html, number, misc, text import webhelpers as w from webhelpers.html.tags import link_to from webhelpers.text import truncate from webhelpers.date import distance_of_time_in_words + from genshi.core import striptags + import re #log = logging.getLogger(__name__) cut = lambda text, max=200: truncate(text, max, whole_word=True) strftimedelta = lambda delta, granularity='minute': distance_of_time_in_words(datetime.now(), datetime.now()+delta, granularity) + striphtml = striptags def link(label, url='', **attrs): return link_to(label, tgurl(url), **attrs) - def striphtml(text): - return re.sub('<[^<]+?>', ' ', text).strip()
Use striptags from genshi for striphtml, since we have to have genshi anyway
## Code Before: from datetime import datetime from tg import url as tgurl #from webhelpers import date, feedgenerator, html, number, misc, text import webhelpers as w from webhelpers.html.tags import link_to from webhelpers.text import truncate from webhelpers.date import distance_of_time_in_words import re #log = logging.getLogger(__name__) cut = lambda text, max=200: truncate(text, max, whole_word=True) strftimedelta = lambda delta, granularity='minute': distance_of_time_in_words(datetime.now(), datetime.now()+delta, granularity) def link(label, url='', **attrs): return link_to(label, tgurl(url), **attrs) def striphtml(text): return re.sub('<[^<]+?>', ' ', text).strip() ## Instruction: Use striptags from genshi for striphtml, since we have to have genshi anyway ## Code After: from datetime import datetime from tg import url as tgurl #from webhelpers import date, feedgenerator, html, number, misc, text import webhelpers as w from webhelpers.html.tags import link_to from webhelpers.text import truncate from webhelpers.date import distance_of_time_in_words from genshi.core import striptags import re #log = logging.getLogger(__name__) cut = lambda text, max=200: truncate(text, max, whole_word=True) strftimedelta = lambda delta, granularity='minute': distance_of_time_in_words(datetime.now(), datetime.now()+delta, granularity) striphtml = striptags def link(label, url='', **attrs): return link_to(label, tgurl(url), **attrs)
// ... existing code ... from genshi.core import striptags import re // ... modified code ... strftimedelta = lambda delta, granularity='minute': distance_of_time_in_words(datetime.now(), datetime.now()+delta, granularity) striphtml = striptags ... // ... rest of the code ...
cd00388bdc4c1963ac8ff81f9b7132ba32272fc8
adwords_client/__init__.py
adwords_client/__init__.py
__version__ = '17.07' # Date based versioning # See: https://packaging.python.org/tutorials/distributing-packages/#date-based-versioning
__version__ = '17.07' # Date based versioning # See: https://packaging.python.org/tutorials/distributing-packages/#date-based-versioning import sys print(__doc__, file=sys.stderr, flush=True)
Print license on each import
Print license on each import
Python
apache-2.0
getninjas/adwords-client
+ __version__ = '17.07' # Date based versioning # See: https://packaging.python.org/tutorials/distributing-packages/#date-based-versioning + import sys + print(__doc__, file=sys.stderr, flush=True) +
Print license on each import
## Code Before: __version__ = '17.07' # Date based versioning # See: https://packaging.python.org/tutorials/distributing-packages/#date-based-versioning ## Instruction: Print license on each import ## Code After: __version__ = '17.07' # Date based versioning # See: https://packaging.python.org/tutorials/distributing-packages/#date-based-versioning import sys print(__doc__, file=sys.stderr, flush=True)
// ... existing code ... __version__ = '17.07' # Date based versioning // ... modified code ... # See: https://packaging.python.org/tutorials/distributing-packages/#date-based-versioning import sys print(__doc__, file=sys.stderr, flush=True) // ... rest of the code ...
14d6955118893c532c1d9f8f6037d1da1b18dbbb
analysis/plot-skeleton.py
analysis/plot-skeleton.py
import climate import database import plots @climate.annotate( root='plot data rooted at this path', pattern=('plot data from files matching this pattern', 'option'), ) def main(root, pattern='*/*block02/*trial00*.csv.gz'): for trial in database.Experiment(root).trials_matching(pattern): with plots.space() as ax: plots.skeleton(ax, trial, 100) if __name__ == '__main__': climate.call(main)
import climate import pandas as pd import database import plots @climate.annotate( root='plot data rooted at this path', pattern=('plot data from files matching this pattern', 'option'), ) def main(root, pattern='*/*block03/*trial00*.csv.gz'): for trial in database.Experiment(root).trials_matching(pattern): with plots.space() as ax: for i in range(3): plots.skeleton(ax, trial, 1000 + 300 * i, lw=2, color='#fd3220', alpha=0.3) #trial.rotate_heading(pd.Series([-6.28 / 10] * len(trial.df))) trial.make_body_relative() for i in range(3): plots.skeleton(ax, trial, 1000 + 300 * i, offset=(0.5 * i, 0.5 * i), lw=2, color='#111111', alpha=0.3) if __name__ == '__main__': climate.call(main)
Add multiple skeletons for the moment.
Add multiple skeletons for the moment.
Python
mit
lmjohns3/cube-experiment,lmjohns3/cube-experiment,lmjohns3/cube-experiment
import climate + import pandas as pd import database import plots @climate.annotate( root='plot data rooted at this path', pattern=('plot data from files matching this pattern', 'option'), ) - def main(root, pattern='*/*block02/*trial00*.csv.gz'): + def main(root, pattern='*/*block03/*trial00*.csv.gz'): for trial in database.Experiment(root).trials_matching(pattern): with plots.space() as ax: - plots.skeleton(ax, trial, 100) + for i in range(3): + plots.skeleton(ax, trial, 1000 + 300 * i, lw=2, color='#fd3220', alpha=0.3) + #trial.rotate_heading(pd.Series([-6.28 / 10] * len(trial.df))) + trial.make_body_relative() + for i in range(3): + plots.skeleton(ax, trial, 1000 + 300 * i, offset=(0.5 * i, 0.5 * i), lw=2, color='#111111', alpha=0.3) if __name__ == '__main__': climate.call(main)
Add multiple skeletons for the moment.
## Code Before: import climate import database import plots @climate.annotate( root='plot data rooted at this path', pattern=('plot data from files matching this pattern', 'option'), ) def main(root, pattern='*/*block02/*trial00*.csv.gz'): for trial in database.Experiment(root).trials_matching(pattern): with plots.space() as ax: plots.skeleton(ax, trial, 100) if __name__ == '__main__': climate.call(main) ## Instruction: Add multiple skeletons for the moment. ## Code After: import climate import pandas as pd import database import plots @climate.annotate( root='plot data rooted at this path', pattern=('plot data from files matching this pattern', 'option'), ) def main(root, pattern='*/*block03/*trial00*.csv.gz'): for trial in database.Experiment(root).trials_matching(pattern): with plots.space() as ax: for i in range(3): plots.skeleton(ax, trial, 1000 + 300 * i, lw=2, color='#fd3220', alpha=0.3) #trial.rotate_heading(pd.Series([-6.28 / 10] * len(trial.df))) trial.make_body_relative() for i in range(3): plots.skeleton(ax, trial, 1000 + 300 * i, offset=(0.5 * i, 0.5 * i), lw=2, color='#111111', alpha=0.3) if __name__ == '__main__': climate.call(main)
// ... existing code ... import climate import pandas as pd // ... modified code ... ) def main(root, pattern='*/*block03/*trial00*.csv.gz'): for trial in database.Experiment(root).trials_matching(pattern): ... with plots.space() as ax: for i in range(3): plots.skeleton(ax, trial, 1000 + 300 * i, lw=2, color='#fd3220', alpha=0.3) #trial.rotate_heading(pd.Series([-6.28 / 10] * len(trial.df))) trial.make_body_relative() for i in range(3): plots.skeleton(ax, trial, 1000 + 300 * i, offset=(0.5 * i, 0.5 * i), lw=2, color='#111111', alpha=0.3) // ... rest of the code ...
f2fc7f1015fc24fdbb69069ac74a21437e94657b
xmantissa/plugins/sineoff.py
xmantissa/plugins/sineoff.py
from axiom import iaxiom, userbase from xmantissa import website, offering, provisioning from sine import sipserver, sinetheme sineproxy = provisioning.BenefactorFactory( name = u'sineproxy', description = u'Sine SIP Proxy', benefactorClass = sipserver.SineBenefactor) plugin = offering.Offering( name = u"Sine", description = u""" The Sine SIP proxy and registrar. """, siteRequirements = ( (userbase.IRealm, userbase.LoginSystem), (None, website.WebSite), (None, sipserver.SIPServer)), appPowerups = (sipserver.SinePublicPage, ), benefactorFactories = (sineproxy,), loginInterfaces=(), themes = (sinetheme.XHTMLDirectoryTheme('base'),) )
from axiom import iaxiom, userbase from xmantissa import website, offering, provisioning from sine import sipserver, sinetheme sineproxy = provisioning.BenefactorFactory( name = u'sineproxy', description = u'Sine SIP Proxy', benefactorClass = sipserver.SineBenefactor) plugin = offering.Offering( name = u"Sine", description = u""" The Sine SIP proxy and registrar. """, siteRequirements = ( (userbase.IRealm, userbase.LoginSystem), (None, website.WebSite), (None, sipserver.SIPServer)), appPowerups = (sipserver.SinePublicPage, ), benefactorFactories = (sineproxy,), themes = (sinetheme.XHTMLDirectoryTheme('base'),) )
Revert 5505 - introduced numerous regressions into the test suite
Revert 5505 - introduced numerous regressions into the test suite
Python
mit
habnabit/divmod-sine,twisted/sine
from axiom import iaxiom, userbase from xmantissa import website, offering, provisioning from sine import sipserver, sinetheme sineproxy = provisioning.BenefactorFactory( name = u'sineproxy', description = u'Sine SIP Proxy', benefactorClass = sipserver.SineBenefactor) plugin = offering.Offering( name = u"Sine", description = u""" The Sine SIP proxy and registrar. """, siteRequirements = ( (userbase.IRealm, userbase.LoginSystem), (None, website.WebSite), (None, sipserver.SIPServer)), appPowerups = (sipserver.SinePublicPage, ), benefactorFactories = (sineproxy,), - loginInterfaces=(), + themes = (sinetheme.XHTMLDirectoryTheme('base'),) )
Revert 5505 - introduced numerous regressions into the test suite
## Code Before: from axiom import iaxiom, userbase from xmantissa import website, offering, provisioning from sine import sipserver, sinetheme sineproxy = provisioning.BenefactorFactory( name = u'sineproxy', description = u'Sine SIP Proxy', benefactorClass = sipserver.SineBenefactor) plugin = offering.Offering( name = u"Sine", description = u""" The Sine SIP proxy and registrar. """, siteRequirements = ( (userbase.IRealm, userbase.LoginSystem), (None, website.WebSite), (None, sipserver.SIPServer)), appPowerups = (sipserver.SinePublicPage, ), benefactorFactories = (sineproxy,), loginInterfaces=(), themes = (sinetheme.XHTMLDirectoryTheme('base'),) ) ## Instruction: Revert 5505 - introduced numerous regressions into the test suite ## Code After: from axiom import iaxiom, userbase from xmantissa import website, offering, provisioning from sine import sipserver, sinetheme sineproxy = provisioning.BenefactorFactory( name = u'sineproxy', description = u'Sine SIP Proxy', benefactorClass = sipserver.SineBenefactor) plugin = offering.Offering( name = u"Sine", description = u""" The Sine SIP proxy and registrar. """, siteRequirements = ( (userbase.IRealm, userbase.LoginSystem), (None, website.WebSite), (None, sipserver.SIPServer)), appPowerups = (sipserver.SinePublicPage, ), benefactorFactories = (sineproxy,), themes = (sinetheme.XHTMLDirectoryTheme('base'),) )
... benefactorFactories = (sineproxy,), themes = (sinetheme.XHTMLDirectoryTheme('base'),) ...
ca8fa466638c0ef405a82dfc3cfecfdb400faaa7
sublime_jedi/helper.py
sublime_jedi/helper.py
import sublime import sublime_plugin from .utils import ask_daemon class HelpMessageCommand(sublime_plugin.TextCommand): def run(self, edit, docstring): self.view.close() self.view.insert(edit, self.view.size(), docstring) class SublimeJediDocstring(sublime_plugin.TextCommand): """ Show doctring in output panel """ def run(self, edit): ask_daemon(self.view, self.show_docstring, 'docstring') def show_docstring(self, view, docstring): window = sublime.active_window() if docstring: output = window.get_output_panel('help_panel') output.set_read_only(False) output.run_command('help_message', {'docstring': docstring}) output.set_read_only(True) window.run_command("show_panel", {"panel": "output.help_panel"}) else: window.run_command("hide_panel", {"panel": "output.help_panel"}) sublime.status_message('Jedi: No results!') class SublimeJediSignature(sublime_plugin.TextCommand): """ Show signature in statusbar """ def run(self, edit): ask_daemon(self.view, self.show_signature, 'signature') def show_signature(self, view, signature): if signature: sublime.status_message('Jedi: {0}'.format(signature))
import sublime import sublime_plugin from .utils import ask_daemon, PythonCommandMixin class HelpMessageCommand(sublime_plugin.TextCommand): def run(self, edit, docstring): self.view.close() self.view.insert(edit, self.view.size(), docstring) class SublimeJediDocstring(PythonCommandMixin, sublime_plugin.TextCommand): """ Show doctring in output panel """ def run(self, edit): ask_daemon(self.view, self.show_docstring, 'docstring') def show_docstring(self, view, docstring): window = sublime.active_window() if docstring: output = window.get_output_panel('help_panel') output.set_read_only(False) output.run_command('help_message', {'docstring': docstring}) output.set_read_only(True) window.run_command("show_panel", {"panel": "output.help_panel"}) else: window.run_command("hide_panel", {"panel": "output.help_panel"}) sublime.status_message('Jedi: No results!') class SublimeJediSignature(PythonCommandMixin, sublime_plugin.TextCommand): """ Show signature in statusbar """ def run(self, edit): ask_daemon(self.view, self.show_signature, 'signature') def show_signature(self, view, signature): if signature: sublime.status_message('Jedi: {0}'.format(signature))
Hide documentation commands in non-python scope
Hide documentation commands in non-python scope
Python
mit
srusskih/SublimeJEDI,edelvalle/SublimeJEDI
import sublime import sublime_plugin - from .utils import ask_daemon + from .utils import ask_daemon, PythonCommandMixin class HelpMessageCommand(sublime_plugin.TextCommand): def run(self, edit, docstring): self.view.close() self.view.insert(edit, self.view.size(), docstring) - class SublimeJediDocstring(sublime_plugin.TextCommand): + class SublimeJediDocstring(PythonCommandMixin, sublime_plugin.TextCommand): """ Show doctring in output panel """ def run(self, edit): ask_daemon(self.view, self.show_docstring, 'docstring') def show_docstring(self, view, docstring): window = sublime.active_window() if docstring: output = window.get_output_panel('help_panel') output.set_read_only(False) output.run_command('help_message', {'docstring': docstring}) output.set_read_only(True) window.run_command("show_panel", {"panel": "output.help_panel"}) else: window.run_command("hide_panel", {"panel": "output.help_panel"}) sublime.status_message('Jedi: No results!') - class SublimeJediSignature(sublime_plugin.TextCommand): + class SublimeJediSignature(PythonCommandMixin, sublime_plugin.TextCommand): """ Show signature in statusbar """ def run(self, edit): ask_daemon(self.view, self.show_signature, 'signature') def show_signature(self, view, signature): if signature: sublime.status_message('Jedi: {0}'.format(signature))
Hide documentation commands in non-python scope
## Code Before: import sublime import sublime_plugin from .utils import ask_daemon class HelpMessageCommand(sublime_plugin.TextCommand): def run(self, edit, docstring): self.view.close() self.view.insert(edit, self.view.size(), docstring) class SublimeJediDocstring(sublime_plugin.TextCommand): """ Show doctring in output panel """ def run(self, edit): ask_daemon(self.view, self.show_docstring, 'docstring') def show_docstring(self, view, docstring): window = sublime.active_window() if docstring: output = window.get_output_panel('help_panel') output.set_read_only(False) output.run_command('help_message', {'docstring': docstring}) output.set_read_only(True) window.run_command("show_panel", {"panel": "output.help_panel"}) else: window.run_command("hide_panel", {"panel": "output.help_panel"}) sublime.status_message('Jedi: No results!') class SublimeJediSignature(sublime_plugin.TextCommand): """ Show signature in statusbar """ def run(self, edit): ask_daemon(self.view, self.show_signature, 'signature') def show_signature(self, view, signature): if signature: sublime.status_message('Jedi: {0}'.format(signature)) ## Instruction: Hide documentation commands in non-python scope ## Code After: import sublime import sublime_plugin from .utils import ask_daemon, PythonCommandMixin class HelpMessageCommand(sublime_plugin.TextCommand): def run(self, edit, docstring): self.view.close() self.view.insert(edit, self.view.size(), docstring) class SublimeJediDocstring(PythonCommandMixin, sublime_plugin.TextCommand): """ Show doctring in output panel """ def run(self, edit): ask_daemon(self.view, self.show_docstring, 'docstring') def show_docstring(self, view, docstring): window = sublime.active_window() if docstring: output = window.get_output_panel('help_panel') output.set_read_only(False) output.run_command('help_message', {'docstring': docstring}) output.set_read_only(True) window.run_command("show_panel", {"panel": "output.help_panel"}) else: window.run_command("hide_panel", {"panel": "output.help_panel"}) sublime.status_message('Jedi: No results!') class SublimeJediSignature(PythonCommandMixin, sublime_plugin.TextCommand): """ Show signature in statusbar """ def run(self, edit): ask_daemon(self.view, self.show_signature, 'signature') def show_signature(self, view, signature): if signature: sublime.status_message('Jedi: {0}'.format(signature))
... from .utils import ask_daemon, PythonCommandMixin ... class SublimeJediDocstring(PythonCommandMixin, sublime_plugin.TextCommand): """ ... class SublimeJediSignature(PythonCommandMixin, sublime_plugin.TextCommand): """ ...
276f22927890051f66976468585d8351c0ccf5b9
sum-of-multiples/sum_of_multiples.py
sum-of-multiples/sum_of_multiples.py
def sum_of_multiples(limit, factors): return sum(filter(lambda n: n < limit, {f*i for i in range(1, limit) for f in factors}))
def sum_of_multiples(limit, factors): return sum({n for f in factors for n in range(f, limit, f)})
Use more optimal method of getting multiples
Use more optimal method of getting multiples
Python
agpl-3.0
CubicComet/exercism-python-solutions
def sum_of_multiples(limit, factors): + return sum({n for f in factors for n in range(f, limit, f)}) - return sum(filter(lambda n: n < limit, - {f*i for i in range(1, limit) for f in factors}))
Use more optimal method of getting multiples
## Code Before: def sum_of_multiples(limit, factors): return sum(filter(lambda n: n < limit, {f*i for i in range(1, limit) for f in factors})) ## Instruction: Use more optimal method of getting multiples ## Code After: def sum_of_multiples(limit, factors): return sum({n for f in factors for n in range(f, limit, f)})
... def sum_of_multiples(limit, factors): return sum({n for f in factors for n in range(f, limit, f)}) ...
19354bd82a89383d795cdada8d6af78e8f12eed8
src/server/test_client.py
src/server/test_client.py
import socket import sys from RemoteFunctionCaller import * from SocketNetworker import SocketNetworker HOST = 'localhost' # The remote host PORT = 8553 # The same port as used by the server s = None for res in socket.getaddrinfo(HOST, PORT, socket.AF_UNSPEC, socket.SOCK_STREAM): af, socktype, proto, canonname, sa = res try: s = socket.socket(af, socktype, proto) except OSError as msg: s = None continue try: s.connect(sa) except OSError as msg: s.close() s = None continue break if s is None: print('could not open socket') sys.exit(1) nw = SocketNetworker(s) caller = RemoteFunctionCaller(nw) try: caller.setData("test", "success") print(caller.getData("test", default="failish")) except TimeoutError: print("Timed out.") nw.close()
import socket import sys from RemoteFunctionCaller import * from SocketNetworker import SocketNetworker HOST = 'localhost' # The remote host PORT = 8553 # The same port as used by the server s = None for res in socket.getaddrinfo(HOST, PORT, socket.AF_UNSPEC, socket.SOCK_STREAM): af, socktype, proto, canonname, sa = res try: s = socket.socket(af, socktype, proto) except OSError as msg: s = None continue try: s.connect(sa) except OSError as msg: s.close() s = None continue break if s is None: print('could not open socket') sys.exit(1) nw = SocketNetworker(s) caller = RemoteFunctionCaller(nw) try: print(caller.SharedClientDataStore__set("test", "success")) print(caller.SharedClientDtaStore__get("test", default="failish")) except TimeoutError: print("Timed out.") nw.close()
Update call method in test client
Update call method in test client
Python
mit
cnlohr/bridgesim,cnlohr/bridgesim,cnlohr/bridgesim,cnlohr/bridgesim
import socket import sys from RemoteFunctionCaller import * from SocketNetworker import SocketNetworker HOST = 'localhost' # The remote host PORT = 8553 # The same port as used by the server s = None for res in socket.getaddrinfo(HOST, PORT, socket.AF_UNSPEC, socket.SOCK_STREAM): af, socktype, proto, canonname, sa = res try: s = socket.socket(af, socktype, proto) except OSError as msg: s = None continue try: s.connect(sa) except OSError as msg: s.close() s = None continue break if s is None: print('could not open socket') sys.exit(1) nw = SocketNetworker(s) caller = RemoteFunctionCaller(nw) try: - caller.setData("test", "success") + print(caller.SharedClientDataStore__set("test", "success")) - print(caller.getData("test", default="failish")) + print(caller.SharedClientDtaStore__get("test", default="failish")) except TimeoutError: print("Timed out.") nw.close()
Update call method in test client
## Code Before: import socket import sys from RemoteFunctionCaller import * from SocketNetworker import SocketNetworker HOST = 'localhost' # The remote host PORT = 8553 # The same port as used by the server s = None for res in socket.getaddrinfo(HOST, PORT, socket.AF_UNSPEC, socket.SOCK_STREAM): af, socktype, proto, canonname, sa = res try: s = socket.socket(af, socktype, proto) except OSError as msg: s = None continue try: s.connect(sa) except OSError as msg: s.close() s = None continue break if s is None: print('could not open socket') sys.exit(1) nw = SocketNetworker(s) caller = RemoteFunctionCaller(nw) try: caller.setData("test", "success") print(caller.getData("test", default="failish")) except TimeoutError: print("Timed out.") nw.close() ## Instruction: Update call method in test client ## Code After: import socket import sys from RemoteFunctionCaller import * from SocketNetworker import SocketNetworker HOST = 'localhost' # The remote host PORT = 8553 # The same port as used by the server s = None for res in socket.getaddrinfo(HOST, PORT, socket.AF_UNSPEC, socket.SOCK_STREAM): af, socktype, proto, canonname, sa = res try: s = socket.socket(af, socktype, proto) except OSError as msg: s = None continue try: s.connect(sa) except OSError as msg: s.close() s = None continue break if s is None: print('could not open socket') sys.exit(1) nw = SocketNetworker(s) caller = RemoteFunctionCaller(nw) try: print(caller.SharedClientDataStore__set("test", "success")) print(caller.SharedClientDtaStore__get("test", default="failish")) except TimeoutError: print("Timed out.") nw.close()
# ... existing code ... try: print(caller.SharedClientDataStore__set("test", "success")) print(caller.SharedClientDtaStore__get("test", default="failish")) except TimeoutError: # ... rest of the code ...
945d64464857581052e18d79e62a6fde8bdecb9b
fabfile.py
fabfile.py
import sys from fabric.api import local, task @task def start_db(): if sys.platform.startswith('darwin'): # Mac OSX local('postgres -D /usr/local/var/postgres -s')
import sys from pathlib import Path from fabric.api import local, task, lcd, env from fabric.contrib.console import confirm from fabric.utils import abort src_p = Path(env.real_fabfile).parent / 'src' @task def start_db(): if sys.platform.startswith('darwin'): # Mac OSX local('postgres -D /usr/local/var/postgres -s') @task def backup(): cmd_dumpdata = 'python manage.py dumpdata ' with lcd(src_p): local( cmd_dumpdata + 'users.EmailUser data_sources.DataSource | ' 'tee ../db_dump/user_sources.json' ) local( cmd_dumpdata + 'experiments | ' 'tee ../db_dump/experiments.json' ) local( cmd_dumpdata + 'analyses.GenomeReference | ' 'tee ../db_dump/genome_reference.json' ) @task def reborn(): with lcd(src_p.as_posix()): db_dump_dir = Path(env.cwd, '../db_dump') if not ( db_dump_dir.joinpath('user_sources.json').exists() and db_dump_dir.joinpath('genome_reference.json').exists() and db_dump_dir.joinpath('experiments.json').exists() ): abort('Backup the import database content first!') confirm('Destory and re-create the current database?', False) local('dropdb biocloud') local('createdb biocloud') local('python manage.py migrate') local('python manage.py loaddata ../db_dump/user_sources.json') local('python manage.py loaddata ../db_dump/genome_reference.json') local('python manage.py loaddata ../db_dump/experiments.json')
Add fab command to backup and destroy database
Add fab command to backup and destroy database
Python
mit
ccwang002/biocloud-server-kai,ccwang002/biocloud-server-kai,ccwang002/biocloud-server-kai
import sys + from pathlib import Path - from fabric.api import local, task + from fabric.api import local, task, lcd, env + from fabric.contrib.console import confirm + from fabric.utils import abort + + src_p = Path(env.real_fabfile).parent / 'src' + @task def start_db(): if sys.platform.startswith('darwin'): # Mac OSX local('postgres -D /usr/local/var/postgres -s') + + @task + def backup(): + cmd_dumpdata = 'python manage.py dumpdata ' + with lcd(src_p): + local( + cmd_dumpdata + 'users.EmailUser data_sources.DataSource | ' + 'tee ../db_dump/user_sources.json' + ) + local( + cmd_dumpdata + 'experiments | ' + 'tee ../db_dump/experiments.json' + ) + local( + cmd_dumpdata + 'analyses.GenomeReference | ' + 'tee ../db_dump/genome_reference.json' + ) + + + @task + def reborn(): + with lcd(src_p.as_posix()): + db_dump_dir = Path(env.cwd, '../db_dump') + if not ( + db_dump_dir.joinpath('user_sources.json').exists() and + db_dump_dir.joinpath('genome_reference.json').exists() and + db_dump_dir.joinpath('experiments.json').exists() + ): + abort('Backup the import database content first!') + confirm('Destory and re-create the current database?', False) + + local('dropdb biocloud') + local('createdb biocloud') + local('python manage.py migrate') + local('python manage.py loaddata ../db_dump/user_sources.json') + local('python manage.py loaddata ../db_dump/genome_reference.json') + local('python manage.py loaddata ../db_dump/experiments.json') +
Add fab command to backup and destroy database
## Code Before: import sys from fabric.api import local, task @task def start_db(): if sys.platform.startswith('darwin'): # Mac OSX local('postgres -D /usr/local/var/postgres -s') ## Instruction: Add fab command to backup and destroy database ## Code After: import sys from pathlib import Path from fabric.api import local, task, lcd, env from fabric.contrib.console import confirm from fabric.utils import abort src_p = Path(env.real_fabfile).parent / 'src' @task def start_db(): if sys.platform.startswith('darwin'): # Mac OSX local('postgres -D /usr/local/var/postgres -s') @task def backup(): cmd_dumpdata = 'python manage.py dumpdata ' with lcd(src_p): local( cmd_dumpdata + 'users.EmailUser data_sources.DataSource | ' 'tee ../db_dump/user_sources.json' ) local( cmd_dumpdata + 'experiments | ' 'tee ../db_dump/experiments.json' ) local( cmd_dumpdata + 'analyses.GenomeReference | ' 'tee ../db_dump/genome_reference.json' ) @task def reborn(): with lcd(src_p.as_posix()): db_dump_dir = Path(env.cwd, '../db_dump') if not ( db_dump_dir.joinpath('user_sources.json').exists() and db_dump_dir.joinpath('genome_reference.json').exists() and db_dump_dir.joinpath('experiments.json').exists() ): abort('Backup the import database content first!') confirm('Destory and re-create the current database?', False) local('dropdb biocloud') local('createdb biocloud') local('python manage.py migrate') local('python manage.py loaddata ../db_dump/user_sources.json') local('python manage.py loaddata ../db_dump/genome_reference.json') local('python manage.py loaddata ../db_dump/experiments.json')
// ... existing code ... import sys from pathlib import Path from fabric.api import local, task, lcd, env from fabric.contrib.console import confirm from fabric.utils import abort src_p = Path(env.real_fabfile).parent / 'src' // ... modified code ... local('postgres -D /usr/local/var/postgres -s') @task def backup(): cmd_dumpdata = 'python manage.py dumpdata ' with lcd(src_p): local( cmd_dumpdata + 'users.EmailUser data_sources.DataSource | ' 'tee ../db_dump/user_sources.json' ) local( cmd_dumpdata + 'experiments | ' 'tee ../db_dump/experiments.json' ) local( cmd_dumpdata + 'analyses.GenomeReference | ' 'tee ../db_dump/genome_reference.json' ) @task def reborn(): with lcd(src_p.as_posix()): db_dump_dir = Path(env.cwd, '../db_dump') if not ( db_dump_dir.joinpath('user_sources.json').exists() and db_dump_dir.joinpath('genome_reference.json').exists() and db_dump_dir.joinpath('experiments.json').exists() ): abort('Backup the import database content first!') confirm('Destory and re-create the current database?', False) local('dropdb biocloud') local('createdb biocloud') local('python manage.py migrate') local('python manage.py loaddata ../db_dump/user_sources.json') local('python manage.py loaddata ../db_dump/genome_reference.json') local('python manage.py loaddata ../db_dump/experiments.json') // ... rest of the code ...
56e6ab84025f071c04701d3dc736b68e82361139
apitestcase/testcase.py
apitestcase/testcase.py
import types import unittest import requests class TestCase(object): """ Add assetion methods for HTTP Requests to TestCase """ hosts = [] def assertGet(self, endpoint="", status_code=200, body=""): """ Asserts GET requests on a given endpoint """ for host in self.hosts: response = requests.get(host+endpoint) self.assertEqual(response.status_code, status_code) if isinstance(body, types.StringType): self.assertIn(body, response.content) elif isinstance(body, list): for content in body: self.assertIn(content, response.content)
import unittest import requests class TestCase(object): """ Add assetion methods for HTTP Requests to TestCase """ hosts = [] def assertGet(self, endpoint="", status_code=200, body=""): """ Asserts GET requests on a given endpoint """ for host in self.hosts: response = requests.get(host+endpoint) self.assertEqual(response.status_code, status_code) if isinstance(body, str): self.assertIn(body, response.content) elif isinstance(body, list): for content in body: self.assertIn(content, response.content)
Change assertGet body check from StringType to str
Change assertGet body check from StringType to str
Python
mit
bramwelt/apitestcase
- import types import unittest import requests class TestCase(object): """ Add assetion methods for HTTP Requests to TestCase """ hosts = [] def assertGet(self, endpoint="", status_code=200, body=""): """ Asserts GET requests on a given endpoint """ for host in self.hosts: response = requests.get(host+endpoint) self.assertEqual(response.status_code, status_code) - if isinstance(body, types.StringType): + if isinstance(body, str): self.assertIn(body, response.content) elif isinstance(body, list): for content in body: self.assertIn(content, response.content)
Change assertGet body check from StringType to str
## Code Before: import types import unittest import requests class TestCase(object): """ Add assetion methods for HTTP Requests to TestCase """ hosts = [] def assertGet(self, endpoint="", status_code=200, body=""): """ Asserts GET requests on a given endpoint """ for host in self.hosts: response = requests.get(host+endpoint) self.assertEqual(response.status_code, status_code) if isinstance(body, types.StringType): self.assertIn(body, response.content) elif isinstance(body, list): for content in body: self.assertIn(content, response.content) ## Instruction: Change assertGet body check from StringType to str ## Code After: import unittest import requests class TestCase(object): """ Add assetion methods for HTTP Requests to TestCase """ hosts = [] def assertGet(self, endpoint="", status_code=200, body=""): """ Asserts GET requests on a given endpoint """ for host in self.hosts: response = requests.get(host+endpoint) self.assertEqual(response.status_code, status_code) if isinstance(body, str): self.assertIn(body, response.content) elif isinstance(body, list): for content in body: self.assertIn(content, response.content)
... import unittest ... self.assertEqual(response.status_code, status_code) if isinstance(body, str): self.assertIn(body, response.content) ...
153c832f083e8ec801ecb8dbddd2f8e79b735eed
utilities.py
utilities.py
import pkg_resources pkg_resources.require('aphla') import aphla as ap def get_pv_names(mode): ''' Given a certain ring mode as a string, return all available pvs ''' ap.machines.load(mode) result = set() elements = ap.getElements('*') for element in elements: pvs = element.pv() if(len(pvs) > 0): pv_name = pvs[0].split(':')[0] result.add(pv_name) return result def get_pvs_from_file(filepath): ''' Return a list of pvs from a given file ''' with open(filepath) as f: contents = f.read().splitlines() return contents
import pkg_resources pkg_resources.require('aphla') import aphla as ap def get_pv_names(mode): ''' Given a certain ring mode as a string, return all available pvs ''' ap.machines.load(mode) result = set() elements = ap.getElements('*') for element in elements: pvs = element.pv() if(len(pvs) > 0): pv_name = pvs[0].split(':')[0] result.add(pv_name) return result def get_pvs_from_file(filepath): ''' Return a list of pvs from a given file ''' with open(filepath) as f: contents = f.read().splitlines() return contents def write_pvs_to_file(filename, data): ''' Write given pvs to file ''' f = open(filename, 'w') for element in data: f.write(element, '\n') f.close()
Add utility function to write pvs to file
Add utility function to write pvs to file
Python
apache-2.0
razvanvasile/Work-Mini-Projects,razvanvasile/Work-Mini-Projects,razvanvasile/Work-Mini-Projects
import pkg_resources pkg_resources.require('aphla') import aphla as ap + def get_pv_names(mode): ''' Given a certain ring mode as a string, return all available pvs ''' ap.machines.load(mode) result = set() elements = ap.getElements('*') for element in elements: pvs = element.pv() if(len(pvs) > 0): pv_name = pvs[0].split(':')[0] result.add(pv_name) return result def get_pvs_from_file(filepath): ''' Return a list of pvs from a given file ''' with open(filepath) as f: contents = f.read().splitlines() return contents + def write_pvs_to_file(filename, data): + ''' Write given pvs to file ''' + f = open(filename, 'w') + for element in data: + f.write(element, '\n') + f.close() +
Add utility function to write pvs to file
## Code Before: import pkg_resources pkg_resources.require('aphla') import aphla as ap def get_pv_names(mode): ''' Given a certain ring mode as a string, return all available pvs ''' ap.machines.load(mode) result = set() elements = ap.getElements('*') for element in elements: pvs = element.pv() if(len(pvs) > 0): pv_name = pvs[0].split(':')[0] result.add(pv_name) return result def get_pvs_from_file(filepath): ''' Return a list of pvs from a given file ''' with open(filepath) as f: contents = f.read().splitlines() return contents ## Instruction: Add utility function to write pvs to file ## Code After: import pkg_resources pkg_resources.require('aphla') import aphla as ap def get_pv_names(mode): ''' Given a certain ring mode as a string, return all available pvs ''' ap.machines.load(mode) result = set() elements = ap.getElements('*') for element in elements: pvs = element.pv() if(len(pvs) > 0): pv_name = pvs[0].split(':')[0] result.add(pv_name) return result def get_pvs_from_file(filepath): ''' Return a list of pvs from a given file ''' with open(filepath) as f: contents = f.read().splitlines() return contents def write_pvs_to_file(filename, data): ''' Write given pvs to file ''' f = open(filename, 'w') for element in data: f.write(element, '\n') f.close()
... import aphla as ap ... def write_pvs_to_file(filename, data): ''' Write given pvs to file ''' f = open(filename, 'w') for element in data: f.write(element, '\n') f.close() ...
82b45c3ec1344bed87ac7d572d82f43a4320492c
craigomatic/wsgi.py
craigomatic/wsgi.py
import os from os.path import abspath, dirname from sys import path from django.core.wsgi import get_wsgi_application SITE_ROOT = dirname(dirname(abspath(__file__))) path.append(SITE_ROOT) os.environ.setdefault("DJANGO_SETTINGS_MODULE", "craigomatic.settings") application = get_wsgi_application()
import os from os.path import abspath, dirname from sys import path from django.core.wsgi import get_wsgi_application from whitenoise.django import DjangoWhiteNoise SITE_ROOT = dirname(dirname(abspath(__file__))) path.append(SITE_ROOT) os.environ.setdefault("DJANGO_SETTINGS_MODULE", "craigomatic.settings") application = get_wsgi_application() application = DjangoWhiteNoise(application)
Integrate whitenoise with the Django application
Integrate whitenoise with the Django application This allows Django to serve static files in production.
Python
mit
rgreinho/craigomatic,rgreinho/craigomatic,rgreinho/craigomatic,rgreinho/craigomatic
import os from os.path import abspath, dirname from sys import path from django.core.wsgi import get_wsgi_application + from whitenoise.django import DjangoWhiteNoise + SITE_ROOT = dirname(dirname(abspath(__file__))) path.append(SITE_ROOT) os.environ.setdefault("DJANGO_SETTINGS_MODULE", "craigomatic.settings") application = get_wsgi_application() + application = DjangoWhiteNoise(application)
Integrate whitenoise with the Django application
## Code Before: import os from os.path import abspath, dirname from sys import path from django.core.wsgi import get_wsgi_application SITE_ROOT = dirname(dirname(abspath(__file__))) path.append(SITE_ROOT) os.environ.setdefault("DJANGO_SETTINGS_MODULE", "craigomatic.settings") application = get_wsgi_application() ## Instruction: Integrate whitenoise with the Django application ## Code After: import os from os.path import abspath, dirname from sys import path from django.core.wsgi import get_wsgi_application from whitenoise.django import DjangoWhiteNoise SITE_ROOT = dirname(dirname(abspath(__file__))) path.append(SITE_ROOT) os.environ.setdefault("DJANGO_SETTINGS_MODULE", "craigomatic.settings") application = get_wsgi_application() application = DjangoWhiteNoise(application)
... from whitenoise.django import DjangoWhiteNoise SITE_ROOT = dirname(dirname(abspath(__file__))) ... application = get_wsgi_application() application = DjangoWhiteNoise(application) ...
f0bd7658b961daceaace56e4ada415c5c9410d54
UM/Operations/ScaleToBoundsOperation.py
UM/Operations/ScaleToBoundsOperation.py
from UM.Operations.Operation import Operation from UM.Math.Vector import Vector ## Operation subclass that will scale a node to fit within the bounds provided. class ScaleToBoundsOperation(Operation): def __init__(self, node, bounds): super().__init__() self._node = node self._old_scale = node.getScale() bbox = self._node.getBoundingBox() largest_dimension = max(bbox.width, bbox.height, bbox.depth) scale_factor = 1.0 if largest_dimension == bbox.width: scale_factor = self._old_scale.x * (bounds.width / bbox.width) elif largest_dimension == bbox.height: scale_factor = self._old_scale.y * (bounds.height / bbox.height) else: scale_factor = self._old_scale.z * (bounds.depth / bbox.depth) self._new_scale = Vector(scale_factor, scale_factor, scale_factor) def undo(self): self._node.setScale(self._old_scale) def redo(self): self._node.setPosition(Vector(0, 0, 0)) self._node.setScale(self._new_scale)
from UM.Operations.Operation import Operation from UM.Math.Vector import Vector ## Operation subclass that will scale a node to fit within the bounds provided. class ScaleToBoundsOperation(Operation): def __init__(self, node, bounds): super().__init__() self._node = node self._old_scale = node.getScale() bbox = self._node.getBoundingBox() largest_dimension = max(bbox.width, bbox.height, bbox.depth) scale_factor = 1.0 if largest_dimension == bbox.depth: scale_factor = self._old_scale.z * (bounds.depth / bbox.depth) elif largest_dimension == bbox.width: scale_factor = self._old_scale.x * (bounds.width / bbox.width) elif largest_dimension == bbox.height: scale_factor = self._old_scale.y * (bounds.height / bbox.height) self._new_scale = Vector(scale_factor, scale_factor, scale_factor) def undo(self): self._node.setScale(self._old_scale) def redo(self): self._node.setPosition(Vector(0, 0, 0)) self._node.setScale(self._new_scale)
Check depth before width since that is more likely to be the smaller dimension
Check depth before width since that is more likely to be the smaller dimension Contributes to Asana issue 37107676459484
Python
agpl-3.0
onitake/Uranium,onitake/Uranium
from UM.Operations.Operation import Operation from UM.Math.Vector import Vector ## Operation subclass that will scale a node to fit within the bounds provided. class ScaleToBoundsOperation(Operation): def __init__(self, node, bounds): super().__init__() self._node = node self._old_scale = node.getScale() bbox = self._node.getBoundingBox() largest_dimension = max(bbox.width, bbox.height, bbox.depth) scale_factor = 1.0 + if largest_dimension == bbox.depth: + scale_factor = self._old_scale.z * (bounds.depth / bbox.depth) - if largest_dimension == bbox.width: + elif largest_dimension == bbox.width: scale_factor = self._old_scale.x * (bounds.width / bbox.width) elif largest_dimension == bbox.height: scale_factor = self._old_scale.y * (bounds.height / bbox.height) - else: - scale_factor = self._old_scale.z * (bounds.depth / bbox.depth) self._new_scale = Vector(scale_factor, scale_factor, scale_factor) def undo(self): self._node.setScale(self._old_scale) def redo(self): self._node.setPosition(Vector(0, 0, 0)) self._node.setScale(self._new_scale)
Check depth before width since that is more likely to be the smaller dimension
## Code Before: from UM.Operations.Operation import Operation from UM.Math.Vector import Vector ## Operation subclass that will scale a node to fit within the bounds provided. class ScaleToBoundsOperation(Operation): def __init__(self, node, bounds): super().__init__() self._node = node self._old_scale = node.getScale() bbox = self._node.getBoundingBox() largest_dimension = max(bbox.width, bbox.height, bbox.depth) scale_factor = 1.0 if largest_dimension == bbox.width: scale_factor = self._old_scale.x * (bounds.width / bbox.width) elif largest_dimension == bbox.height: scale_factor = self._old_scale.y * (bounds.height / bbox.height) else: scale_factor = self._old_scale.z * (bounds.depth / bbox.depth) self._new_scale = Vector(scale_factor, scale_factor, scale_factor) def undo(self): self._node.setScale(self._old_scale) def redo(self): self._node.setPosition(Vector(0, 0, 0)) self._node.setScale(self._new_scale) ## Instruction: Check depth before width since that is more likely to be the smaller dimension ## Code After: from UM.Operations.Operation import Operation from UM.Math.Vector import Vector ## Operation subclass that will scale a node to fit within the bounds provided. class ScaleToBoundsOperation(Operation): def __init__(self, node, bounds): super().__init__() self._node = node self._old_scale = node.getScale() bbox = self._node.getBoundingBox() largest_dimension = max(bbox.width, bbox.height, bbox.depth) scale_factor = 1.0 if largest_dimension == bbox.depth: scale_factor = self._old_scale.z * (bounds.depth / bbox.depth) elif largest_dimension == bbox.width: scale_factor = self._old_scale.x * (bounds.width / bbox.width) elif largest_dimension == bbox.height: scale_factor = self._old_scale.y * (bounds.height / bbox.height) self._new_scale = Vector(scale_factor, scale_factor, scale_factor) def undo(self): self._node.setScale(self._old_scale) def redo(self): self._node.setPosition(Vector(0, 0, 0)) self._node.setScale(self._new_scale)
// ... existing code ... scale_factor = 1.0 if largest_dimension == bbox.depth: scale_factor = self._old_scale.z * (bounds.depth / bbox.depth) elif largest_dimension == bbox.width: scale_factor = self._old_scale.x * (bounds.width / bbox.width) // ... modified code ... scale_factor = self._old_scale.y * (bounds.height / bbox.height) // ... rest of the code ...
57d3f5a78385b07fb4d7f91ac97edb6e9dc850aa
waterbutler/providers/osfstorage/metadata.py
waterbutler/providers/osfstorage/metadata.py
from waterbutler.core import metadata class BaseOsfStorageMetadata: @property def provider(self): return 'osfstorage' class OsfStorageFileMetadata(BaseOsfStorageMetadata, metadata.BaseFileMetadata): @property def name(self): return self.raw['name'] @property def path(self): if self.raw['path'][0] != '/': return '/' + self.raw['path'] return self.raw['path'] @property def modified(self): return self.raw.get('modified') @property def size(self): return self.raw.get('size') @property def content_type(self): return None @property def extra(self): return { 'downloads': self.raw['downloads'] } class OsfStorageFolderMetadata(BaseOsfStorageMetadata, metadata.BaseFolderMetadata): @property def name(self): return self.raw['name'] @property def path(self): return self.raw['path']
from waterbutler.core import metadata class BaseOsfStorageMetadata: @property def provider(self): return 'osfstorage' class OsfStorageFileMetadata(BaseOsfStorageMetadata, metadata.BaseFileMetadata): @property def name(self): return self.raw['name'] @property def path(self): if self.raw['path'][0].startswith('/'): return self.raw['path'] return '/' + self.raw['path'] @property def modified(self): return self.raw.get('modified') @property def size(self): return self.raw.get('size') @property def content_type(self): return None @property def extra(self): return { 'downloads': self.raw['downloads'] } class OsfStorageFolderMetadata(BaseOsfStorageMetadata, metadata.BaseFolderMetadata): @property def name(self): return self.raw['name'] @property def path(self): return self.raw['path']
Use startswith in favor of indexing
Use startswith in favor of indexing
Python
apache-2.0
CenterForOpenScience/waterbutler,cosenal/waterbutler,icereval/waterbutler,chrisseto/waterbutler,RCOSDP/waterbutler,rafaeldelucena/waterbutler,Johnetordoff/waterbutler,TomBaxter/waterbutler,felliott/waterbutler,rdhyee/waterbutler,kwierman/waterbutler,Ghalko/waterbutler,hmoco/waterbutler
from waterbutler.core import metadata class BaseOsfStorageMetadata: @property def provider(self): return 'osfstorage' class OsfStorageFileMetadata(BaseOsfStorageMetadata, metadata.BaseFileMetadata): @property def name(self): return self.raw['name'] @property def path(self): - if self.raw['path'][0] != '/': + if self.raw['path'][0].startswith('/'): + return self.raw['path'] - return '/' + self.raw['path'] + return '/' + self.raw['path'] - return self.raw['path'] @property def modified(self): return self.raw.get('modified') @property def size(self): return self.raw.get('size') @property def content_type(self): return None @property def extra(self): return { 'downloads': self.raw['downloads'] } class OsfStorageFolderMetadata(BaseOsfStorageMetadata, metadata.BaseFolderMetadata): @property def name(self): return self.raw['name'] @property def path(self): return self.raw['path']
Use startswith in favor of indexing
## Code Before: from waterbutler.core import metadata class BaseOsfStorageMetadata: @property def provider(self): return 'osfstorage' class OsfStorageFileMetadata(BaseOsfStorageMetadata, metadata.BaseFileMetadata): @property def name(self): return self.raw['name'] @property def path(self): if self.raw['path'][0] != '/': return '/' + self.raw['path'] return self.raw['path'] @property def modified(self): return self.raw.get('modified') @property def size(self): return self.raw.get('size') @property def content_type(self): return None @property def extra(self): return { 'downloads': self.raw['downloads'] } class OsfStorageFolderMetadata(BaseOsfStorageMetadata, metadata.BaseFolderMetadata): @property def name(self): return self.raw['name'] @property def path(self): return self.raw['path'] ## Instruction: Use startswith in favor of indexing ## Code After: from waterbutler.core import metadata class BaseOsfStorageMetadata: @property def provider(self): return 'osfstorage' class OsfStorageFileMetadata(BaseOsfStorageMetadata, metadata.BaseFileMetadata): @property def name(self): return self.raw['name'] @property def path(self): if self.raw['path'][0].startswith('/'): return self.raw['path'] return '/' + self.raw['path'] @property def modified(self): return self.raw.get('modified') @property def size(self): return self.raw.get('size') @property def content_type(self): return None @property def extra(self): return { 'downloads': self.raw['downloads'] } class OsfStorageFolderMetadata(BaseOsfStorageMetadata, metadata.BaseFolderMetadata): @property def name(self): return self.raw['name'] @property def path(self): return self.raw['path']
# ... existing code ... def path(self): if self.raw['path'][0].startswith('/'): return self.raw['path'] return '/' + self.raw['path'] # ... rest of the code ...
a6061ef140e371101c3d04c5b85562586293eee8
scrappyr/scraps/tests/test_models.py
scrappyr/scraps/tests/test_models.py
from test_plus.test import TestCase from ..models import Scrap class TestScrap(TestCase): def test__str__(self): scrap = Scrap(raw_title='hello') assert str(scrap) == 'hello' def test_html_title(self): scrap = Scrap(raw_title='hello') assert scrap.html_title == 'hello' def test_html_title_bold(self): scrap = Scrap(raw_title='**hello**') assert scrap.html_title == '<strong>hello</strong>'
from test_plus.test import TestCase from ..models import Scrap class TestScrap(TestCase): def test__str__(self): scrap = Scrap(raw_title='hello') assert str(scrap) == 'hello' def test_html_title(self): scrap = Scrap(raw_title='hello') assert scrap.html_title == 'hello' def test_html_title_bold(self): scrap = Scrap(raw_title='**hello**') assert scrap.html_title == '<strong>hello</strong>' def test_html_title_with_block_element_gets_escaped(self): scrap = Scrap(raw_title='<div>hello</div>') assert scrap.html_title == '&lt;div&gt;hello&lt;/div&gt;'
Add test of block-elements in scrap title
Add test of block-elements in scrap title
Python
mit
tonysyu/scrappyr-app,tonysyu/scrappyr-app,tonysyu/scrappyr-app,tonysyu/scrappyr-app
from test_plus.test import TestCase from ..models import Scrap class TestScrap(TestCase): def test__str__(self): scrap = Scrap(raw_title='hello') assert str(scrap) == 'hello' def test_html_title(self): scrap = Scrap(raw_title='hello') assert scrap.html_title == 'hello' def test_html_title_bold(self): scrap = Scrap(raw_title='**hello**') assert scrap.html_title == '<strong>hello</strong>' + def test_html_title_with_block_element_gets_escaped(self): + scrap = Scrap(raw_title='<div>hello</div>') + assert scrap.html_title == '&lt;div&gt;hello&lt;/div&gt;' +
Add test of block-elements in scrap title
## Code Before: from test_plus.test import TestCase from ..models import Scrap class TestScrap(TestCase): def test__str__(self): scrap = Scrap(raw_title='hello') assert str(scrap) == 'hello' def test_html_title(self): scrap = Scrap(raw_title='hello') assert scrap.html_title == 'hello' def test_html_title_bold(self): scrap = Scrap(raw_title='**hello**') assert scrap.html_title == '<strong>hello</strong>' ## Instruction: Add test of block-elements in scrap title ## Code After: from test_plus.test import TestCase from ..models import Scrap class TestScrap(TestCase): def test__str__(self): scrap = Scrap(raw_title='hello') assert str(scrap) == 'hello' def test_html_title(self): scrap = Scrap(raw_title='hello') assert scrap.html_title == 'hello' def test_html_title_bold(self): scrap = Scrap(raw_title='**hello**') assert scrap.html_title == '<strong>hello</strong>' def test_html_title_with_block_element_gets_escaped(self): scrap = Scrap(raw_title='<div>hello</div>') assert scrap.html_title == '&lt;div&gt;hello&lt;/div&gt;'
... assert scrap.html_title == '<strong>hello</strong>' def test_html_title_with_block_element_gets_escaped(self): scrap = Scrap(raw_title='<div>hello</div>') assert scrap.html_title == '&lt;div&gt;hello&lt;/div&gt;' ...
2a0b1d070996bfb3d950d4fae70b264ddabc7d2f
sheldon/config.py
sheldon/config.py
import os class Config: def __init__(self, prefix='SHELDON_'): """ Load config from environment variables. :param prefix: string, all needed environment variables starts from it. Default - 'SHELDON_'. So, environment variables will be looking like: 'SHELDON_BOT_NAME', 'SHELDON_TWITTER_KEY' :return: """ # Bot config variables self.variables = {} for variable in os.environ: if variable.startswith(prefix): self.variables[variable] = os.environ[variable] def get(self, variable, default_value): """ :param variable: string, needed variable :param default_value: string, value that returns if variable is not set :return: """ if variable not in self.variables: return default_value return self.variables[variable]
import os class Config: def __init__(self, prefix='SHELDON_'): """ Load config from environment variables. :param prefix: string, all needed environment variables starts from it. Default - 'SHELDON_'. So, environment variables will be looking like: 'SHELDON_BOT_NAME', 'SHELDON_TWITTER_KEY' :return: """ # Bot config variables self.variables = {} for variable in os.environ: if variable.startswith(prefix): self.variables[variable] = os.environ[variable] def get(self, variable, default_value): """ Get variable value from environment :param variable: string, needed variable :param default_value: string, value that returns if variable is not set :return: variable value """ if variable not in self.variables: return default_value return self.variables[variable] def get_installed_plugins(self): """ Return list of installed plugins from installed_plugins.txt :return: list of strings with names of plugins """ plugins_file = open('installed_plugins.txt') return plugins_file.readlines()
Add function for getting installed plugins
Add function for getting installed plugins
Python
mit
lises/sheldon
import os class Config: def __init__(self, prefix='SHELDON_'): """ Load config from environment variables. :param prefix: string, all needed environment variables starts from it. Default - 'SHELDON_'. So, environment variables will be looking like: 'SHELDON_BOT_NAME', 'SHELDON_TWITTER_KEY' :return: """ # Bot config variables self.variables = {} for variable in os.environ: if variable.startswith(prefix): self.variables[variable] = os.environ[variable] def get(self, variable, default_value): """ + Get variable value from environment :param variable: string, needed variable :param default_value: string, value that returns if variable is not set - :return: + :return: variable value """ if variable not in self.variables: return default_value return self.variables[variable] + def get_installed_plugins(self): + """ + Return list of installed plugins from installed_plugins.txt + :return: list of strings with names of plugins + """ + plugins_file = open('installed_plugins.txt') + return plugins_file.readlines() +
Add function for getting installed plugins
## Code Before: import os class Config: def __init__(self, prefix='SHELDON_'): """ Load config from environment variables. :param prefix: string, all needed environment variables starts from it. Default - 'SHELDON_'. So, environment variables will be looking like: 'SHELDON_BOT_NAME', 'SHELDON_TWITTER_KEY' :return: """ # Bot config variables self.variables = {} for variable in os.environ: if variable.startswith(prefix): self.variables[variable] = os.environ[variable] def get(self, variable, default_value): """ :param variable: string, needed variable :param default_value: string, value that returns if variable is not set :return: """ if variable not in self.variables: return default_value return self.variables[variable] ## Instruction: Add function for getting installed plugins ## Code After: import os class Config: def __init__(self, prefix='SHELDON_'): """ Load config from environment variables. :param prefix: string, all needed environment variables starts from it. Default - 'SHELDON_'. So, environment variables will be looking like: 'SHELDON_BOT_NAME', 'SHELDON_TWITTER_KEY' :return: """ # Bot config variables self.variables = {} for variable in os.environ: if variable.startswith(prefix): self.variables[variable] = os.environ[variable] def get(self, variable, default_value): """ Get variable value from environment :param variable: string, needed variable :param default_value: string, value that returns if variable is not set :return: variable value """ if variable not in self.variables: return default_value return self.variables[variable] def get_installed_plugins(self): """ Return list of installed plugins from installed_plugins.txt :return: list of strings with names of plugins """ plugins_file = open('installed_plugins.txt') return plugins_file.readlines()
... """ Get variable value from environment ... variable is not set :return: variable value """ ... def get_installed_plugins(self): """ Return list of installed plugins from installed_plugins.txt :return: list of strings with names of plugins """ plugins_file = open('installed_plugins.txt') return plugins_file.readlines() ...
e51d35545d038b5cb7035cc74f39e4a5c2b0756a
thinglang/execution/classes.py
thinglang/execution/classes.py
from thinglang.lexer.symbols.base import LexicalIdentifier class ThingInstance(object): def __init__(self, cls): self.cls = cls self.methods = { x.name: x for x in self.cls.children } self.members = {} def __contains__(self, item): return item in self.members or item in self.methods def __getitem__(self, item): return self.members.get(item) or self.methods.get(item) def __str__(self): return f'Thing<{self.cls}>(members={self.members}, methods={self.methods})'
from thinglang.lexer.symbols.base import LexicalIdentifier class ThingInstance(object): def __init__(self, cls): self.cls = cls self.methods = { x.name: x for x in self.cls.children } self.members = {} def __contains__(self, item): return item in self.members or item in self.methods def __getitem__(self, item): return self.members.get(item) if item in self.members else self.methods[item] def __str__(self): return f'Thing<{self.cls}>(members={self.members}, methods={self.methods})'
Fix bug in ThingInstace __setitem__
Fix bug in ThingInstace __setitem__
Python
mit
ytanay/thinglang,ytanay/thinglang,ytanay/thinglang,ytanay/thinglang
from thinglang.lexer.symbols.base import LexicalIdentifier class ThingInstance(object): def __init__(self, cls): self.cls = cls self.methods = { x.name: x for x in self.cls.children } self.members = {} def __contains__(self, item): return item in self.members or item in self.methods def __getitem__(self, item): - return self.members.get(item) or self.methods.get(item) + return self.members.get(item) if item in self.members else self.methods[item] def __str__(self): return f'Thing<{self.cls}>(members={self.members}, methods={self.methods})'
Fix bug in ThingInstace __setitem__
## Code Before: from thinglang.lexer.symbols.base import LexicalIdentifier class ThingInstance(object): def __init__(self, cls): self.cls = cls self.methods = { x.name: x for x in self.cls.children } self.members = {} def __contains__(self, item): return item in self.members or item in self.methods def __getitem__(self, item): return self.members.get(item) or self.methods.get(item) def __str__(self): return f'Thing<{self.cls}>(members={self.members}, methods={self.methods})' ## Instruction: Fix bug in ThingInstace __setitem__ ## Code After: from thinglang.lexer.symbols.base import LexicalIdentifier class ThingInstance(object): def __init__(self, cls): self.cls = cls self.methods = { x.name: x for x in self.cls.children } self.members = {} def __contains__(self, item): return item in self.members or item in self.methods def __getitem__(self, item): return self.members.get(item) if item in self.members else self.methods[item] def __str__(self): return f'Thing<{self.cls}>(members={self.members}, methods={self.methods})'
# ... existing code ... def __getitem__(self, item): return self.members.get(item) if item in self.members else self.methods[item] # ... rest of the code ...
a10e21a8fe811e896998ba510255592a966f0782
infra/recipes/build_windows.py
infra/recipes/build_windows.py
from recipe_engine.post_process import Filter PYTHON_VERSION_COMPATIBILITY = "PY3" DEPS = [ "crosvm", "recipe_engine/buildbucket", "recipe_engine/context", "recipe_engine/properties", "recipe_engine/step", ] def RunSteps(api): # Note: The recipe does work on linux as well, if the required dependencies have been installed # on the host via ./tools/install-deps. # This allows the build to be tested via `./recipe.py run build_windows` with api.crosvm.host_build_context(): api.step( "Build crosvm tests", [ "vpython3", "./tools/run_tests", "--verbose", "--build-only", ], ) api.step( "Run crosvm tests", [ "vpython3", "./tools/run_tests", "--verbose", ], ) def GenTests(api): filter_steps = Filter("Build crosvm tests", "Run crosvm tests") yield ( api.test( "build", api.buildbucket.ci_build(project="crosvm/crosvm"), ) + api.post_process(filter_steps) )
from recipe_engine.post_process import Filter PYTHON_VERSION_COMPATIBILITY = "PY3" DEPS = [ "crosvm", "recipe_engine/buildbucket", "recipe_engine/context", "recipe_engine/properties", "recipe_engine/step", ] def RunSteps(api): # Note: The recipe does work on linux as well, if the required dependencies have been installed # on the host via ./tools/install-deps. # This allows the build to be tested via `./recipe.py run build_windows` with api.crosvm.host_build_context(): api.step( "Build crosvm tests", [ "vpython3", "./tools/run_tests", "--verbose", "--build-only", ], ) api.step( "Run crosvm tests", [ "vpython3", "./tools/run_tests", "--verbose", ], ) api.step( "Clippy windows crosvm", [ "vpython3", "./tools/clippy", ], ) def GenTests(api): filter_steps = Filter("Build crosvm tests", "Run crosvm tests") yield ( api.test( "build", api.buildbucket.ci_build(project="crosvm/crosvm"), ) + api.post_process(filter_steps) )
Enable clippy in windows LUCI
crosvm: Enable clippy in windows LUCI For linux based systems, clippy continues to run in health_check BUG=b:257249038 TEST=CQ Change-Id: I39d3d45a0db72c61e79fd2c51b195b82c067a244 Reviewed-on: https://chromium-review.googlesource.com/c/crosvm/crosvm/+/3993934 Reviewed-by: Dennis Kempin <[email protected]> Commit-Queue: Vikram Auradkar <[email protected]>
Python
bsd-3-clause
google/crosvm,google/crosvm,google/crosvm,google/crosvm
from recipe_engine.post_process import Filter PYTHON_VERSION_COMPATIBILITY = "PY3" DEPS = [ "crosvm", "recipe_engine/buildbucket", "recipe_engine/context", "recipe_engine/properties", "recipe_engine/step", ] def RunSteps(api): # Note: The recipe does work on linux as well, if the required dependencies have been installed # on the host via ./tools/install-deps. # This allows the build to be tested via `./recipe.py run build_windows` with api.crosvm.host_build_context(): api.step( "Build crosvm tests", [ "vpython3", "./tools/run_tests", "--verbose", "--build-only", ], ) api.step( "Run crosvm tests", [ "vpython3", "./tools/run_tests", "--verbose", ], ) + api.step( + "Clippy windows crosvm", + [ + "vpython3", + "./tools/clippy", + ], + ) def GenTests(api): filter_steps = Filter("Build crosvm tests", "Run crosvm tests") yield ( api.test( "build", api.buildbucket.ci_build(project="crosvm/crosvm"), ) + api.post_process(filter_steps) )
Enable clippy in windows LUCI
## Code Before: from recipe_engine.post_process import Filter PYTHON_VERSION_COMPATIBILITY = "PY3" DEPS = [ "crosvm", "recipe_engine/buildbucket", "recipe_engine/context", "recipe_engine/properties", "recipe_engine/step", ] def RunSteps(api): # Note: The recipe does work on linux as well, if the required dependencies have been installed # on the host via ./tools/install-deps. # This allows the build to be tested via `./recipe.py run build_windows` with api.crosvm.host_build_context(): api.step( "Build crosvm tests", [ "vpython3", "./tools/run_tests", "--verbose", "--build-only", ], ) api.step( "Run crosvm tests", [ "vpython3", "./tools/run_tests", "--verbose", ], ) def GenTests(api): filter_steps = Filter("Build crosvm tests", "Run crosvm tests") yield ( api.test( "build", api.buildbucket.ci_build(project="crosvm/crosvm"), ) + api.post_process(filter_steps) ) ## Instruction: Enable clippy in windows LUCI ## Code After: from recipe_engine.post_process import Filter PYTHON_VERSION_COMPATIBILITY = "PY3" DEPS = [ "crosvm", "recipe_engine/buildbucket", "recipe_engine/context", "recipe_engine/properties", "recipe_engine/step", ] def RunSteps(api): # Note: The recipe does work on linux as well, if the required dependencies have been installed # on the host via ./tools/install-deps. # This allows the build to be tested via `./recipe.py run build_windows` with api.crosvm.host_build_context(): api.step( "Build crosvm tests", [ "vpython3", "./tools/run_tests", "--verbose", "--build-only", ], ) api.step( "Run crosvm tests", [ "vpython3", "./tools/run_tests", "--verbose", ], ) api.step( "Clippy windows crosvm", [ "vpython3", "./tools/clippy", ], ) def GenTests(api): filter_steps = Filter("Build crosvm tests", "Run crosvm tests") yield ( api.test( "build", api.buildbucket.ci_build(project="crosvm/crosvm"), ) + api.post_process(filter_steps) )
# ... existing code ... ) api.step( "Clippy windows crosvm", [ "vpython3", "./tools/clippy", ], ) # ... rest of the code ...
fc39c6afa49a312413468dfdffcc2de94bb7d78e
tests/test_runner.py
tests/test_runner.py
import unittest from mo.runner import Variable class TestVariable(unittest.TestCase): def test_default(self): v = Variable('name', {'default': 'default'}) self.assertEqual(v.value, 'default') self.assertEqual(str(v), 'default') def test_value(self): v = Variable('name', {'default': 'default'}, 'value') self.assertEqual(v.value, 'value') self.assertEqual(str(v), 'value') def test_str(self): v = Variable('name', {'default': 'abc'}) self.assertEqual(str(v), v.value)
import unittest from mo.runner import Task, Variable class TestVariable(unittest.TestCase): def test_default(self): v = Variable('name', {'default': 'default'}) self.assertEqual(v.value, 'default') self.assertEqual(str(v), 'default') def test_value(self): v = Variable('name', {'default': 'default'}, 'value') self.assertEqual(v.value, 'value') self.assertEqual(str(v), 'value') def test_str(self): v = Variable('name', {'default': 'abc'}) self.assertEqual(str(v), v.value) class TestTask(unittest.TestCase): def test_variables(self): t = Task('name', {'description': '', 'command': '{{ v }}'}, {'v': 'variable'}) self.assertEqual(t.commands[0], 'variable')
Add some more tests for tasks
Add some more tests for tasks
Python
mit
thomasleese/mo
import unittest - from mo.runner import Variable + from mo.runner import Task, Variable class TestVariable(unittest.TestCase): def test_default(self): v = Variable('name', {'default': 'default'}) self.assertEqual(v.value, 'default') self.assertEqual(str(v), 'default') def test_value(self): v = Variable('name', {'default': 'default'}, 'value') self.assertEqual(v.value, 'value') self.assertEqual(str(v), 'value') def test_str(self): v = Variable('name', {'default': 'abc'}) self.assertEqual(str(v), v.value) + + class TestTask(unittest.TestCase): + def test_variables(self): + t = Task('name', {'description': '', 'command': '{{ v }}'}, + {'v': 'variable'}) + self.assertEqual(t.commands[0], 'variable') +
Add some more tests for tasks
## Code Before: import unittest from mo.runner import Variable class TestVariable(unittest.TestCase): def test_default(self): v = Variable('name', {'default': 'default'}) self.assertEqual(v.value, 'default') self.assertEqual(str(v), 'default') def test_value(self): v = Variable('name', {'default': 'default'}, 'value') self.assertEqual(v.value, 'value') self.assertEqual(str(v), 'value') def test_str(self): v = Variable('name', {'default': 'abc'}) self.assertEqual(str(v), v.value) ## Instruction: Add some more tests for tasks ## Code After: import unittest from mo.runner import Task, Variable class TestVariable(unittest.TestCase): def test_default(self): v = Variable('name', {'default': 'default'}) self.assertEqual(v.value, 'default') self.assertEqual(str(v), 'default') def test_value(self): v = Variable('name', {'default': 'default'}, 'value') self.assertEqual(v.value, 'value') self.assertEqual(str(v), 'value') def test_str(self): v = Variable('name', {'default': 'abc'}) self.assertEqual(str(v), v.value) class TestTask(unittest.TestCase): def test_variables(self): t = Task('name', {'description': '', 'command': '{{ v }}'}, {'v': 'variable'}) self.assertEqual(t.commands[0], 'variable')
// ... existing code ... from mo.runner import Task, Variable // ... modified code ... self.assertEqual(str(v), v.value) class TestTask(unittest.TestCase): def test_variables(self): t = Task('name', {'description': '', 'command': '{{ v }}'}, {'v': 'variable'}) self.assertEqual(t.commands[0], 'variable') // ... rest of the code ...
7643635278fc1c92289e8fdd456614ce85a2c2f3
addons/osfstorage/models.py
addons/osfstorage/models.py
import logging from addons.base.models import BaseNodeSettings, BaseStorageAddon logger = logging.getLogger(__name__) class NodeSettings(BaseStorageAddon, BaseNodeSettings): pass
import logging from addons.base.models import BaseNodeSettings, BaseStorageAddon logger = logging.getLogger(__name__) class NodeSettings(BaseStorageAddon, BaseNodeSettings): # Required overrides complete = True has_auth = True
Add required overrides to osfstorage.NodeSettings
Add required overrides to osfstorage.NodeSettings
Python
apache-2.0
felliott/osf.io,CenterForOpenScience/osf.io,laurenrevere/osf.io,cslzchen/osf.io,monikagrabowska/osf.io,caneruguz/osf.io,Nesiehr/osf.io,alexschiller/osf.io,aaxelb/osf.io,caseyrollins/osf.io,Johnetordoff/osf.io,caneruguz/osf.io,mluo613/osf.io,chrisseto/osf.io,monikagrabowska/osf.io,cslzchen/osf.io,monikagrabowska/osf.io,monikagrabowska/osf.io,saradbowman/osf.io,leb2dg/osf.io,cslzchen/osf.io,alexschiller/osf.io,hmoco/osf.io,mfraezz/osf.io,saradbowman/osf.io,CenterForOpenScience/osf.io,mluo613/osf.io,brianjgeiger/osf.io,acshi/osf.io,acshi/osf.io,leb2dg/osf.io,Nesiehr/osf.io,erinspace/osf.io,erinspace/osf.io,chrisseto/osf.io,alexschiller/osf.io,binoculars/osf.io,icereval/osf.io,brianjgeiger/osf.io,alexschiller/osf.io,Nesiehr/osf.io,crcresearch/osf.io,adlius/osf.io,adlius/osf.io,chrisseto/osf.io,sloria/osf.io,sloria/osf.io,chennan47/osf.io,laurenrevere/osf.io,HalcyonChimera/osf.io,laurenrevere/osf.io,chennan47/osf.io,leb2dg/osf.io,icereval/osf.io,felliott/osf.io,mattclark/osf.io,HalcyonChimera/osf.io,acshi/osf.io,erinspace/osf.io,baylee-d/osf.io,TomBaxter/osf.io,cwisecarver/osf.io,TomBaxter/osf.io,mfraezz/osf.io,leb2dg/osf.io,baylee-d/osf.io,acshi/osf.io,aaxelb/osf.io,icereval/osf.io,monikagrabowska/osf.io,Johnetordoff/osf.io,caneruguz/osf.io,pattisdr/osf.io,hmoco/osf.io,caseyrollins/osf.io,chennan47/osf.io,TomBaxter/osf.io,caseyrollins/osf.io,felliott/osf.io,Johnetordoff/osf.io,mluo613/osf.io,binoculars/osf.io,mluo613/osf.io,acshi/osf.io,adlius/osf.io,HalcyonChimera/osf.io,baylee-d/osf.io,Johnetordoff/osf.io,cslzchen/osf.io,aaxelb/osf.io,sloria/osf.io,cwisecarver/osf.io,mluo613/osf.io,adlius/osf.io,alexschiller/osf.io,cwisecarver/osf.io,CenterForOpenScience/osf.io,cwisecarver/osf.io,crcresearch/osf.io,mfraezz/osf.io,caneruguz/osf.io,hmoco/osf.io,brianjgeiger/osf.io,brianjgeiger/osf.io,mfraezz/osf.io,mattclark/osf.io,HalcyonChimera/osf.io,CenterForOpenScience/osf.io,Nesiehr/osf.io,mattclark/osf.io,chrisseto/osf.io,crcresearch/osf.io,pattisdr/osf.io,binoculars/osf.io,pattisdr/osf.io,felliott/osf.io,aaxelb/osf.io,hmoco/osf.io
import logging from addons.base.models import BaseNodeSettings, BaseStorageAddon logger = logging.getLogger(__name__) class NodeSettings(BaseStorageAddon, BaseNodeSettings): - pass + # Required overrides + complete = True + has_auth = True
Add required overrides to osfstorage.NodeSettings
## Code Before: import logging from addons.base.models import BaseNodeSettings, BaseStorageAddon logger = logging.getLogger(__name__) class NodeSettings(BaseStorageAddon, BaseNodeSettings): pass ## Instruction: Add required overrides to osfstorage.NodeSettings ## Code After: import logging from addons.base.models import BaseNodeSettings, BaseStorageAddon logger = logging.getLogger(__name__) class NodeSettings(BaseStorageAddon, BaseNodeSettings): # Required overrides complete = True has_auth = True
# ... existing code ... class NodeSettings(BaseStorageAddon, BaseNodeSettings): # Required overrides complete = True has_auth = True # ... rest of the code ...
28ecf02c3d08eae725512e1563cf74f1831bd02d
gears/engines/base.py
gears/engines/base.py
import subprocess from functools import wraps class EngineProcessFailed(Exception): pass class BaseEngine(object): result_mimetype = None @classmethod def as_engine(cls, **initkwargs): @wraps(cls, updated=()) def engine(asset): instance = engine.engine_class(**initkwargs) return instance.process(asset) engine.engine_class = cls engine.result_mimetype = cls.result_mimetype return engine def process(self, asset): raise NotImplementedError() class ExecEngine(BaseEngine): executable = None params = [] def __init__(self, executable=None): if executable is not None: self.executable = executable def process(self, asset): self.asset = asset p = subprocess.Popen( args=self.get_args(), stdin=subprocess.PIPE, stdout=subprocess.PIPE, stderr=subprocess.PIPE) output, errors = p.communicate(input=asset.processed_source) if p.returncode != 0: raise EngineProcessFailed(errors) asset.processed_source = output def get_args(self): return [self.executable] + self.params
import subprocess from functools import wraps class EngineProcessFailed(Exception): pass class BaseEngine(object): result_mimetype = None @classmethod def as_engine(cls, **initkwargs): @wraps(cls, updated=()) def engine(asset): instance = engine.engine_class(**initkwargs) return instance.process(asset) engine.engine_class = cls engine.result_mimetype = cls.result_mimetype return engine def process(self, asset): raise NotImplementedError() class ExecEngine(BaseEngine): executable = None params = [] def __init__(self, executable=None): if executable is not None: self.executable = executable def process(self, asset): self.asset = asset p = subprocess.Popen( args=self.get_args(), stdin=subprocess.PIPE, stdout=subprocess.PIPE, stderr=subprocess.PIPE) output, errors = p.communicate(input=asset.processed_source.encode('utf-8')) if p.returncode != 0: raise EngineProcessFailed(errors) asset.processed_source = output.decode('utf-8') def get_args(self): return [self.executable] + self.params
Fix unicode support in ExecEngine
Fix unicode support in ExecEngine
Python
isc
gears/gears,gears/gears,gears/gears
import subprocess from functools import wraps class EngineProcessFailed(Exception): pass class BaseEngine(object): result_mimetype = None @classmethod def as_engine(cls, **initkwargs): @wraps(cls, updated=()) def engine(asset): instance = engine.engine_class(**initkwargs) return instance.process(asset) engine.engine_class = cls engine.result_mimetype = cls.result_mimetype return engine def process(self, asset): raise NotImplementedError() class ExecEngine(BaseEngine): executable = None params = [] def __init__(self, executable=None): if executable is not None: self.executable = executable def process(self, asset): self.asset = asset p = subprocess.Popen( args=self.get_args(), stdin=subprocess.PIPE, stdout=subprocess.PIPE, stderr=subprocess.PIPE) - output, errors = p.communicate(input=asset.processed_source) + output, errors = p.communicate(input=asset.processed_source.encode('utf-8')) if p.returncode != 0: raise EngineProcessFailed(errors) - asset.processed_source = output + asset.processed_source = output.decode('utf-8') def get_args(self): return [self.executable] + self.params
Fix unicode support in ExecEngine
## Code Before: import subprocess from functools import wraps class EngineProcessFailed(Exception): pass class BaseEngine(object): result_mimetype = None @classmethod def as_engine(cls, **initkwargs): @wraps(cls, updated=()) def engine(asset): instance = engine.engine_class(**initkwargs) return instance.process(asset) engine.engine_class = cls engine.result_mimetype = cls.result_mimetype return engine def process(self, asset): raise NotImplementedError() class ExecEngine(BaseEngine): executable = None params = [] def __init__(self, executable=None): if executable is not None: self.executable = executable def process(self, asset): self.asset = asset p = subprocess.Popen( args=self.get_args(), stdin=subprocess.PIPE, stdout=subprocess.PIPE, stderr=subprocess.PIPE) output, errors = p.communicate(input=asset.processed_source) if p.returncode != 0: raise EngineProcessFailed(errors) asset.processed_source = output def get_args(self): return [self.executable] + self.params ## Instruction: Fix unicode support in ExecEngine ## Code After: import subprocess from functools import wraps class EngineProcessFailed(Exception): pass class BaseEngine(object): result_mimetype = None @classmethod def as_engine(cls, **initkwargs): @wraps(cls, updated=()) def engine(asset): instance = engine.engine_class(**initkwargs) return instance.process(asset) engine.engine_class = cls engine.result_mimetype = cls.result_mimetype return engine def process(self, asset): raise NotImplementedError() class ExecEngine(BaseEngine): executable = None params = [] def __init__(self, executable=None): if executable is not None: self.executable = executable def process(self, asset): self.asset = asset p = subprocess.Popen( args=self.get_args(), stdin=subprocess.PIPE, stdout=subprocess.PIPE, stderr=subprocess.PIPE) output, errors = p.communicate(input=asset.processed_source.encode('utf-8')) if p.returncode != 0: raise EngineProcessFailed(errors) asset.processed_source = output.decode('utf-8') def get_args(self): return [self.executable] + self.params
# ... existing code ... stderr=subprocess.PIPE) output, errors = p.communicate(input=asset.processed_source.encode('utf-8')) if p.returncode != 0: # ... modified code ... raise EngineProcessFailed(errors) asset.processed_source = output.decode('utf-8') # ... rest of the code ...
03a54ea1920a3716e9c8d326f5c4c408f45b7d08
apps/api/urls.py
apps/api/urls.py
from django.conf.urls import url, include from rest_framework.routers import DefaultRouter from rest_framework_jwt import views as jwt_views from . import views router = DefaultRouter() router.register(r'quotes', views.QuoteViewSet) router.register(r'authors', views.AuthorViewSet) router.register(r'categories', views.CategoryViewSet) router.register(r'tags', views.TagViewSet) urlpatterns = [ url(r'^docs/$', views.schema_view), url(r'^', include(router.urls)), url(r'^token/new/$', jwt_views.obtain_jwt_token), url(r'^token/refresh/$', jwt_views.refresh_jwt_token), url(r'^filters/$', views.FiltersOptionsView.as_view()), url(r'^templates/(?P<page>[-\w]+.html)/$', views.AngularTemplateView.as_view()), ]
from django.conf.urls import url, include from rest_framework.routers import DefaultRouter from rest_framework_jwt import views as jwt_views from . import views router = DefaultRouter() router.register(r'quotes', views.QuoteViewSet) router.register(r'authors', views.AuthorViewSet) router.register(r'categories', views.CategoryViewSet) router.register(r'tags', views.TagViewSet) urlpatterns = [ url(r'^docs/$', views.schema_view), url(r'^', include(router.urls)), url(r'^token/new/$', jwt_views.ObtainJSONWebToken.as_view()), url(r'^token/refresh/$', jwt_views.RefreshJSONWebToken.as_view()), url(r'^token/verify/$', jwt_views.VerifyJSONWebToken.as_view()), url(r'^filters/$', views.FiltersOptionsView.as_view()), url(r'^templates/(?P<page>[-\w]+.html)/$', views.AngularTemplateView.as_view()), ]
Add option to verify jwt token
Add option to verify jwt token
Python
bsd-3-clause
lucifurtun/myquotes,lucifurtun/myquotes,lucifurtun/myquotes,lucifurtun/myquotes
from django.conf.urls import url, include from rest_framework.routers import DefaultRouter from rest_framework_jwt import views as jwt_views from . import views router = DefaultRouter() router.register(r'quotes', views.QuoteViewSet) router.register(r'authors', views.AuthorViewSet) router.register(r'categories', views.CategoryViewSet) router.register(r'tags', views.TagViewSet) urlpatterns = [ url(r'^docs/$', views.schema_view), url(r'^', include(router.urls)), - url(r'^token/new/$', jwt_views.obtain_jwt_token), + url(r'^token/new/$', jwt_views.ObtainJSONWebToken.as_view()), - url(r'^token/refresh/$', jwt_views.refresh_jwt_token), + url(r'^token/refresh/$', jwt_views.RefreshJSONWebToken.as_view()), + url(r'^token/verify/$', jwt_views.VerifyJSONWebToken.as_view()), url(r'^filters/$', views.FiltersOptionsView.as_view()), url(r'^templates/(?P<page>[-\w]+.html)/$', views.AngularTemplateView.as_view()), ]
Add option to verify jwt token
## Code Before: from django.conf.urls import url, include from rest_framework.routers import DefaultRouter from rest_framework_jwt import views as jwt_views from . import views router = DefaultRouter() router.register(r'quotes', views.QuoteViewSet) router.register(r'authors', views.AuthorViewSet) router.register(r'categories', views.CategoryViewSet) router.register(r'tags', views.TagViewSet) urlpatterns = [ url(r'^docs/$', views.schema_view), url(r'^', include(router.urls)), url(r'^token/new/$', jwt_views.obtain_jwt_token), url(r'^token/refresh/$', jwt_views.refresh_jwt_token), url(r'^filters/$', views.FiltersOptionsView.as_view()), url(r'^templates/(?P<page>[-\w]+.html)/$', views.AngularTemplateView.as_view()), ] ## Instruction: Add option to verify jwt token ## Code After: from django.conf.urls import url, include from rest_framework.routers import DefaultRouter from rest_framework_jwt import views as jwt_views from . import views router = DefaultRouter() router.register(r'quotes', views.QuoteViewSet) router.register(r'authors', views.AuthorViewSet) router.register(r'categories', views.CategoryViewSet) router.register(r'tags', views.TagViewSet) urlpatterns = [ url(r'^docs/$', views.schema_view), url(r'^', include(router.urls)), url(r'^token/new/$', jwt_views.ObtainJSONWebToken.as_view()), url(r'^token/refresh/$', jwt_views.RefreshJSONWebToken.as_view()), url(r'^token/verify/$', jwt_views.VerifyJSONWebToken.as_view()), url(r'^filters/$', views.FiltersOptionsView.as_view()), url(r'^templates/(?P<page>[-\w]+.html)/$', views.AngularTemplateView.as_view()), ]
# ... existing code ... url(r'^', include(router.urls)), url(r'^token/new/$', jwt_views.ObtainJSONWebToken.as_view()), url(r'^token/refresh/$', jwt_views.RefreshJSONWebToken.as_view()), url(r'^token/verify/$', jwt_views.VerifyJSONWebToken.as_view()), url(r'^filters/$', views.FiltersOptionsView.as_view()), # ... rest of the code ...
6f6f6e183b574f8505b53ddb7651c8766992b953
pywikibot/families/lingualibre_family.py
pywikibot/families/lingualibre_family.py
"""Family module for Lingua Libre.""" # # (C) Pywikibot team, 2021 # # Distributed under the terms of the MIT license. # from pywikibot import family # The Lingua Libre family class Family(family.WikimediaFamily): """Family class for Lingua Libre. *New in version 6.5.* """ name = 'lingualibre' langs = { 'lingualibre': 'lingualibre.org' } interwiki_forward = 'wikipedia' def scriptpath(self, code): """Return the script path for this family.""" return ''
"""Family module for Lingua Libre.""" # # (C) Pywikibot team, 2021 # # Distributed under the terms of the MIT license. # from pywikibot import family # The Lingua Libre family class Family(family.WikimediaFamily): """Family class for Lingua Libre. *New in version 6.5.* """ name = 'lingualibre' langs = { 'lingualibre': 'lingualibre.org' } interwiki_forward = 'wikipedia' def scriptpath(self, code): """Return the script path for this family.""" return '' def interface(self, code): """Return 'DataSite'.""" return 'DataSite'
Allow to request for item on Lingua Libre
Allow to request for item on Lingua Libre Bug: T286303 Change-Id: Ic0d8824d1bf326f2182fbb64d0cc2ed77f82fd4c
Python
mit
wikimedia/pywikibot-core,wikimedia/pywikibot-core
"""Family module for Lingua Libre.""" # # (C) Pywikibot team, 2021 # # Distributed under the terms of the MIT license. # from pywikibot import family # The Lingua Libre family class Family(family.WikimediaFamily): """Family class for Lingua Libre. *New in version 6.5.* """ name = 'lingualibre' langs = { 'lingualibre': 'lingualibre.org' } interwiki_forward = 'wikipedia' def scriptpath(self, code): """Return the script path for this family.""" return '' + def interface(self, code): + """Return 'DataSite'.""" + return 'DataSite' +
Allow to request for item on Lingua Libre
## Code Before: """Family module for Lingua Libre.""" # # (C) Pywikibot team, 2021 # # Distributed under the terms of the MIT license. # from pywikibot import family # The Lingua Libre family class Family(family.WikimediaFamily): """Family class for Lingua Libre. *New in version 6.5.* """ name = 'lingualibre' langs = { 'lingualibre': 'lingualibre.org' } interwiki_forward = 'wikipedia' def scriptpath(self, code): """Return the script path for this family.""" return '' ## Instruction: Allow to request for item on Lingua Libre ## Code After: """Family module for Lingua Libre.""" # # (C) Pywikibot team, 2021 # # Distributed under the terms of the MIT license. # from pywikibot import family # The Lingua Libre family class Family(family.WikimediaFamily): """Family class for Lingua Libre. *New in version 6.5.* """ name = 'lingualibre' langs = { 'lingualibre': 'lingualibre.org' } interwiki_forward = 'wikipedia' def scriptpath(self, code): """Return the script path for this family.""" return '' def interface(self, code): """Return 'DataSite'.""" return 'DataSite'
// ... existing code ... return '' def interface(self, code): """Return 'DataSite'.""" return 'DataSite' // ... rest of the code ...
faf35a814d045ce3d71921ed0d4ac268d5a9811c
app/notify_client/provider_client.py
app/notify_client/provider_client.py
from app.notify_client import _attach_current_user, NotifyAdminAPIClient class ProviderClient(NotifyAdminAPIClient): def __init__(self): super().__init__("a", "b", "c") def init_app(self, app): self.base_url = app.config['API_HOST_NAME'] self.service_id = app.config['ADMIN_CLIENT_USER_NAME'] self.api_key = app.config['ADMIN_CLIENT_SECRET'] def get_all_providers(self): return self.get( url='/provider-details' ) def get_provider_by_id(self, provider_id): return self.get( url='/provider-details/{}'.format(provider_id) ) def update_provider(self, provider_id, priority): data = { "priority": priority } data = _attach_current_user(data) return self.post(url='/provider-details/{}'.format(provider_id), data=data)
from app.notify_client import _attach_current_user, NotifyAdminAPIClient class ProviderClient(NotifyAdminAPIClient): def __init__(self): super().__init__("a", "b", "c") def init_app(self, app): self.base_url = app.config['API_HOST_NAME'] self.service_id = app.config['ADMIN_CLIENT_USER_NAME'] self.api_key = app.config['ADMIN_CLIENT_SECRET'] def get_all_providers(self): return self.get( url='/provider-details' ) def get_provider_by_id(self, provider_id): return self.get( url='/provider-details/{}'.format(provider_id) ) def get_provider_versions(self, provider_id): return self.get( url='/provider-details/{}/versions'.format(provider_id) ) def update_provider(self, provider_id, priority): data = { "priority": priority } data = _attach_current_user(data) return self.post(url='/provider-details/{}'.format(provider_id), data=data)
Add provider client method to get provider version history
Add provider client method to get provider version history
Python
mit
gov-cjwaszczuk/notifications-admin,alphagov/notifications-admin,alphagov/notifications-admin,alphagov/notifications-admin,alphagov/notifications-admin,gov-cjwaszczuk/notifications-admin,gov-cjwaszczuk/notifications-admin,gov-cjwaszczuk/notifications-admin
from app.notify_client import _attach_current_user, NotifyAdminAPIClient class ProviderClient(NotifyAdminAPIClient): def __init__(self): super().__init__("a", "b", "c") def init_app(self, app): self.base_url = app.config['API_HOST_NAME'] self.service_id = app.config['ADMIN_CLIENT_USER_NAME'] self.api_key = app.config['ADMIN_CLIENT_SECRET'] def get_all_providers(self): return self.get( url='/provider-details' ) def get_provider_by_id(self, provider_id): return self.get( url='/provider-details/{}'.format(provider_id) ) + def get_provider_versions(self, provider_id): + return self.get( + url='/provider-details/{}/versions'.format(provider_id) + ) + def update_provider(self, provider_id, priority): data = { "priority": priority } data = _attach_current_user(data) return self.post(url='/provider-details/{}'.format(provider_id), data=data)
Add provider client method to get provider version history
## Code Before: from app.notify_client import _attach_current_user, NotifyAdminAPIClient class ProviderClient(NotifyAdminAPIClient): def __init__(self): super().__init__("a", "b", "c") def init_app(self, app): self.base_url = app.config['API_HOST_NAME'] self.service_id = app.config['ADMIN_CLIENT_USER_NAME'] self.api_key = app.config['ADMIN_CLIENT_SECRET'] def get_all_providers(self): return self.get( url='/provider-details' ) def get_provider_by_id(self, provider_id): return self.get( url='/provider-details/{}'.format(provider_id) ) def update_provider(self, provider_id, priority): data = { "priority": priority } data = _attach_current_user(data) return self.post(url='/provider-details/{}'.format(provider_id), data=data) ## Instruction: Add provider client method to get provider version history ## Code After: from app.notify_client import _attach_current_user, NotifyAdminAPIClient class ProviderClient(NotifyAdminAPIClient): def __init__(self): super().__init__("a", "b", "c") def init_app(self, app): self.base_url = app.config['API_HOST_NAME'] self.service_id = app.config['ADMIN_CLIENT_USER_NAME'] self.api_key = app.config['ADMIN_CLIENT_SECRET'] def get_all_providers(self): return self.get( url='/provider-details' ) def get_provider_by_id(self, provider_id): return self.get( url='/provider-details/{}'.format(provider_id) ) def get_provider_versions(self, provider_id): return self.get( url='/provider-details/{}/versions'.format(provider_id) ) def update_provider(self, provider_id, priority): data = { "priority": priority } data = _attach_current_user(data) return self.post(url='/provider-details/{}'.format(provider_id), data=data)
// ... existing code ... def get_provider_versions(self, provider_id): return self.get( url='/provider-details/{}/versions'.format(provider_id) ) def update_provider(self, provider_id, priority): // ... rest of the code ...
b42dddaa45a8915a653f4b145f2a58eb6996f28a
home/openbox/lib/helpers.py
home/openbox/lib/helpers.py
import os.path, os def run_as_browser(fn): def wrapped(*args, **kwargs): rv = fn(*args, **kwargs) return 'sudo -Hiu browser %s' % rv return wrapped class Helpers: @property @run_as_browser def default_firefox_bin(self): candidates = [ '/usr/local/lib/firefox/firefox-bin', '/usr/local/lib/firefox3/firefox-bin', '/usr/bin/iceweasel', ] return self._pick(candidates, os.path.exists) @property @run_as_browser def default_firefox_wrapper(self): candidates = [ 'firefox', 'firefox3' ] return self._pick(candidates, self._wrapper_tester) default_firefox = default_firefox_wrapper @property def as_browser(self): return 'sudo -Hiu browser' @property def opera(self): return 'sudo -Hiu browser opera' @property def chrome(self): return 'sudo -Hiu browser chrome' def have_bin(self, basename): return self._wrapper_tester(basename) def _wrapper_tester(self, candidate): dirs = os.environ['PATH'].split(':') for dir in dirs: path = os.path.join(dir, candidate) if os.path.exists(path): return True return False def _pick(self, candidates, tester): for candidate in candidates: if tester(candidate): return candidate # consider raising here return None
import os.path, os def run_as_browser(fn): def wrapped(*args, **kwargs): rv = fn(*args, **kwargs) return 'sudo -Hiu browser env XAUTHORITY=/home/browser/.Xauthority %s' % rv return wrapped class Helpers: @property @run_as_browser def default_firefox_bin(self): candidates = [ '/usr/local/lib/firefox/firefox-bin', '/usr/local/lib/firefox3/firefox-bin', '/usr/bin/iceweasel', ] return self._pick(candidates, os.path.exists) @property @run_as_browser def default_firefox_wrapper(self): candidates = [ 'firefox', 'firefox3' ] return self._pick(candidates, self._wrapper_tester) default_firefox = default_firefox_wrapper @property def as_browser(self): return 'sudo -Hiu browser' @property def opera(self): return 'sudo -Hiu browser opera' @property def chrome(self): return 'sudo -Hiu browser chrome' def have_bin(self, basename): return self._wrapper_tester(basename) def _wrapper_tester(self, candidate): dirs = os.environ['PATH'].split(':') for dir in dirs: path = os.path.join(dir, candidate) if os.path.exists(path): return True return False def _pick(self, candidates, tester): for candidate in candidates: if tester(candidate): return candidate # consider raising here return None
Fix firefox invocation as browser
Fix firefox invocation as browser
Python
bsd-2-clause
p/pubfiles,p/pubfiles,p/pubfiles,p/pubfiles,p/pubfiles
import os.path, os def run_as_browser(fn): def wrapped(*args, **kwargs): rv = fn(*args, **kwargs) - return 'sudo -Hiu browser %s' % rv + return 'sudo -Hiu browser env XAUTHORITY=/home/browser/.Xauthority %s' % rv return wrapped class Helpers: @property @run_as_browser def default_firefox_bin(self): candidates = [ '/usr/local/lib/firefox/firefox-bin', '/usr/local/lib/firefox3/firefox-bin', '/usr/bin/iceweasel', ] return self._pick(candidates, os.path.exists) @property @run_as_browser def default_firefox_wrapper(self): candidates = [ 'firefox', 'firefox3' ] return self._pick(candidates, self._wrapper_tester) default_firefox = default_firefox_wrapper @property def as_browser(self): return 'sudo -Hiu browser' @property def opera(self): return 'sudo -Hiu browser opera' @property def chrome(self): return 'sudo -Hiu browser chrome' def have_bin(self, basename): return self._wrapper_tester(basename) def _wrapper_tester(self, candidate): dirs = os.environ['PATH'].split(':') for dir in dirs: path = os.path.join(dir, candidate) if os.path.exists(path): return True return False def _pick(self, candidates, tester): for candidate in candidates: if tester(candidate): return candidate # consider raising here return None
Fix firefox invocation as browser
## Code Before: import os.path, os def run_as_browser(fn): def wrapped(*args, **kwargs): rv = fn(*args, **kwargs) return 'sudo -Hiu browser %s' % rv return wrapped class Helpers: @property @run_as_browser def default_firefox_bin(self): candidates = [ '/usr/local/lib/firefox/firefox-bin', '/usr/local/lib/firefox3/firefox-bin', '/usr/bin/iceweasel', ] return self._pick(candidates, os.path.exists) @property @run_as_browser def default_firefox_wrapper(self): candidates = [ 'firefox', 'firefox3' ] return self._pick(candidates, self._wrapper_tester) default_firefox = default_firefox_wrapper @property def as_browser(self): return 'sudo -Hiu browser' @property def opera(self): return 'sudo -Hiu browser opera' @property def chrome(self): return 'sudo -Hiu browser chrome' def have_bin(self, basename): return self._wrapper_tester(basename) def _wrapper_tester(self, candidate): dirs = os.environ['PATH'].split(':') for dir in dirs: path = os.path.join(dir, candidate) if os.path.exists(path): return True return False def _pick(self, candidates, tester): for candidate in candidates: if tester(candidate): return candidate # consider raising here return None ## Instruction: Fix firefox invocation as browser ## Code After: import os.path, os def run_as_browser(fn): def wrapped(*args, **kwargs): rv = fn(*args, **kwargs) return 'sudo -Hiu browser env XAUTHORITY=/home/browser/.Xauthority %s' % rv return wrapped class Helpers: @property @run_as_browser def default_firefox_bin(self): candidates = [ '/usr/local/lib/firefox/firefox-bin', '/usr/local/lib/firefox3/firefox-bin', '/usr/bin/iceweasel', ] return self._pick(candidates, os.path.exists) @property @run_as_browser def default_firefox_wrapper(self): candidates = [ 'firefox', 'firefox3' ] return self._pick(candidates, self._wrapper_tester) default_firefox = default_firefox_wrapper @property def as_browser(self): return 'sudo -Hiu browser' @property def opera(self): return 'sudo -Hiu browser opera' @property def chrome(self): return 'sudo -Hiu browser chrome' def have_bin(self, basename): return self._wrapper_tester(basename) def _wrapper_tester(self, candidate): dirs = os.environ['PATH'].split(':') for dir in dirs: path = os.path.join(dir, candidate) if os.path.exists(path): return True return False def _pick(self, candidates, tester): for candidate in candidates: if tester(candidate): return candidate # consider raising here return None
# ... existing code ... rv = fn(*args, **kwargs) return 'sudo -Hiu browser env XAUTHORITY=/home/browser/.Xauthority %s' % rv return wrapped # ... rest of the code ...
fc7ba9019b42f056713b81bfee70f9e780b4aab5
models/rasmachine/twitter_client.py
models/rasmachine/twitter_client.py
from __future__ import absolute_import, print_function, unicode_literals from builtins import dict, str import tweepy def get_oauth(auth_file): try: fh = open(auth_file, 'rt') except IOError: print('Could not get Twitter credentials.') return None lines = [l.strip() for l in fh.readlines()] oauth = tweepy.OAuthHandler(lines[0], lines[1]) oauth.set_access_token(lines[2], lines[3]) fh.close() return oauth def update_status(msg, auth_file='twitter_cred.txt'): twitter_auth = get_oauth(auth_file) if twitter_auth is None: return twitter_api = tweepy.API(twitter_auth) twitter_api.update_status(msg)
from __future__ import absolute_import, print_function, unicode_literals from builtins import dict, str import tweepy def get_oauth_file(auth_file): try: fh = open(auth_file, 'rt') except IOError: print('Could not get Twitter credentials.') return None lines = [l.strip() for l in fh.readlines()] oauth = tweepy.OAuthHandler(lines[0], lines[1]) oauth.set_access_token(lines[2], lines[3]) fh.close() return oauth def get_oauth_dict(auth_dict): oauth = tweepy.OAuthHandler(auth_dict.get('consumer_token'), auth_dict.get('consumer_secred')) oauth.set_access_token(auth_dict.get('access_token'), auth_dict.get('access_secret')) return oauth def update_status(msg, twitter_cred): twitter_auth = get_oauth_dict(twitter_cred) if twitter_auth is None: return twitter_api = tweepy.API(twitter_auth) twitter_api.update_status(msg)
Implement dict credentials in Twitter client
Implement dict credentials in Twitter client
Python
bsd-2-clause
pvtodorov/indra,johnbachman/belpy,jmuhlich/indra,jmuhlich/indra,bgyori/indra,johnbachman/belpy,pvtodorov/indra,sorgerlab/belpy,johnbachman/belpy,pvtodorov/indra,johnbachman/indra,bgyori/indra,pvtodorov/indra,johnbachman/indra,sorgerlab/belpy,johnbachman/indra,sorgerlab/indra,bgyori/indra,sorgerlab/belpy,sorgerlab/indra,jmuhlich/indra,sorgerlab/indra
from __future__ import absolute_import, print_function, unicode_literals from builtins import dict, str import tweepy - def get_oauth(auth_file): + def get_oauth_file(auth_file): try: fh = open(auth_file, 'rt') except IOError: print('Could not get Twitter credentials.') return None lines = [l.strip() for l in fh.readlines()] oauth = tweepy.OAuthHandler(lines[0], lines[1]) oauth.set_access_token(lines[2], lines[3]) fh.close() return oauth + def get_oauth_dict(auth_dict): + oauth = tweepy.OAuthHandler(auth_dict.get('consumer_token'), + auth_dict.get('consumer_secred')) + oauth.set_access_token(auth_dict.get('access_token'), + auth_dict.get('access_secret')) + return oauth + - def update_status(msg, auth_file='twitter_cred.txt'): + def update_status(msg, twitter_cred): - twitter_auth = get_oauth(auth_file) + twitter_auth = get_oauth_dict(twitter_cred) if twitter_auth is None: return twitter_api = tweepy.API(twitter_auth) twitter_api.update_status(msg)
Implement dict credentials in Twitter client
## Code Before: from __future__ import absolute_import, print_function, unicode_literals from builtins import dict, str import tweepy def get_oauth(auth_file): try: fh = open(auth_file, 'rt') except IOError: print('Could not get Twitter credentials.') return None lines = [l.strip() for l in fh.readlines()] oauth = tweepy.OAuthHandler(lines[0], lines[1]) oauth.set_access_token(lines[2], lines[3]) fh.close() return oauth def update_status(msg, auth_file='twitter_cred.txt'): twitter_auth = get_oauth(auth_file) if twitter_auth is None: return twitter_api = tweepy.API(twitter_auth) twitter_api.update_status(msg) ## Instruction: Implement dict credentials in Twitter client ## Code After: from __future__ import absolute_import, print_function, unicode_literals from builtins import dict, str import tweepy def get_oauth_file(auth_file): try: fh = open(auth_file, 'rt') except IOError: print('Could not get Twitter credentials.') return None lines = [l.strip() for l in fh.readlines()] oauth = tweepy.OAuthHandler(lines[0], lines[1]) oauth.set_access_token(lines[2], lines[3]) fh.close() return oauth def get_oauth_dict(auth_dict): oauth = tweepy.OAuthHandler(auth_dict.get('consumer_token'), auth_dict.get('consumer_secred')) oauth.set_access_token(auth_dict.get('access_token'), auth_dict.get('access_secret')) return oauth def update_status(msg, twitter_cred): twitter_auth = get_oauth_dict(twitter_cred) if twitter_auth is None: return twitter_api = tweepy.API(twitter_auth) twitter_api.update_status(msg)
... def get_oauth_file(auth_file): try: ... def get_oauth_dict(auth_dict): oauth = tweepy.OAuthHandler(auth_dict.get('consumer_token'), auth_dict.get('consumer_secred')) oauth.set_access_token(auth_dict.get('access_token'), auth_dict.get('access_secret')) return oauth def update_status(msg, twitter_cred): twitter_auth = get_oauth_dict(twitter_cred) if twitter_auth is None: ...
6a3fbb7280c1078b574736eae3c6a3e4e42d3f46
seaborn/__init__.py
seaborn/__init__.py
import matplotlib as mpl _orig_rc_params = mpl.rcParams.copy() # Import seaborn objects from .rcmod import * from .utils import * from .palettes import * from .relational import * from .regression import * from .categorical import * from .distributions import * from .timeseries import * from .matrix import * from .miscplot import * from .axisgrid import * from .widgets import * from .colors import xkcd_rgb, crayons from . import cm __version__ = "0.9.1.dev0"
import matplotlib as mpl _orig_rc_params = mpl.rcParams.copy() # Import seaborn objects from .rcmod import * from .utils import * from .palettes import * from .relational import * from .regression import * from .categorical import * from .distributions import * from .matrix import * from .miscplot import * from .axisgrid import * from .widgets import * from .colors import xkcd_rgb, crayons from . import cm __version__ = "0.9.1.dev0"
Remove top-level import of timeseries module
Remove top-level import of timeseries module
Python
bsd-3-clause
arokem/seaborn,mwaskom/seaborn,mwaskom/seaborn,arokem/seaborn,anntzer/seaborn,anntzer/seaborn
import matplotlib as mpl _orig_rc_params = mpl.rcParams.copy() # Import seaborn objects from .rcmod import * from .utils import * from .palettes import * from .relational import * from .regression import * from .categorical import * from .distributions import * - from .timeseries import * from .matrix import * from .miscplot import * from .axisgrid import * from .widgets import * from .colors import xkcd_rgb, crayons from . import cm __version__ = "0.9.1.dev0"
Remove top-level import of timeseries module
## Code Before: import matplotlib as mpl _orig_rc_params = mpl.rcParams.copy() # Import seaborn objects from .rcmod import * from .utils import * from .palettes import * from .relational import * from .regression import * from .categorical import * from .distributions import * from .timeseries import * from .matrix import * from .miscplot import * from .axisgrid import * from .widgets import * from .colors import xkcd_rgb, crayons from . import cm __version__ = "0.9.1.dev0" ## Instruction: Remove top-level import of timeseries module ## Code After: import matplotlib as mpl _orig_rc_params = mpl.rcParams.copy() # Import seaborn objects from .rcmod import * from .utils import * from .palettes import * from .relational import * from .regression import * from .categorical import * from .distributions import * from .matrix import * from .miscplot import * from .axisgrid import * from .widgets import * from .colors import xkcd_rgb, crayons from . import cm __version__ = "0.9.1.dev0"
... from .distributions import * from .matrix import * ...
592c6550255793772add694cb941a0db0883713b
kamboo/core.py
kamboo/core.py
import logging import botocore from kotocore.session import Session log = logging.getLogger(__name__) class KambooConnection(object): """ Kamboo connection with botocore session initialized """ session = botocore.session.get_session() def __init__(self, service_name="ec2", region_name="us-east-1", account_id=None, credentials=None): self.region = region_name self.account_id = account_id self.credentials = credentials if self.credentials: self.session.set_credentials(**self.credentials) Connection = Session(session=self.session).get_connection(service_name) self.conn = Connection(region_name=self.region)
import logging import botocore from kotocore.session import Session log = logging.getLogger(__name__) class KambooConnection(object): """ Kamboo connection with botocore session initialized """ def __init__(self, service_name="ec2", region_name="us-east-1", account_id=None, credentials=None): self.session = botocore.session.get_session() self.service = service_name self.region = region_name self.account_id = account_id self.credentials = credentials if self.credentials: self.session.set_credentials(**self.credentials) Connection = Session(session=self.session).get_connection(service_name) self.conn = Connection(region_name=self.region) def __repr__(self): return "KambooConnection: [%s, %s, %s]" % (self.account_id, self.region, self.service)
Fix the issue: "session" shared in different connections
Fix the issue: "session" shared in different connections
Python
apache-2.0
henrysher/kamboo,henrysher/kamboo
import logging import botocore from kotocore.session import Session log = logging.getLogger(__name__) class KambooConnection(object): """ Kamboo connection with botocore session initialized """ - session = botocore.session.get_session() def __init__(self, service_name="ec2", region_name="us-east-1", account_id=None, credentials=None): + self.session = botocore.session.get_session() + self.service = service_name self.region = region_name self.account_id = account_id self.credentials = credentials if self.credentials: self.session.set_credentials(**self.credentials) Connection = Session(session=self.session).get_connection(service_name) self.conn = Connection(region_name=self.region) + def __repr__(self): + return "KambooConnection: [%s, %s, %s]" % (self.account_id, + self.region, + self.service) +
Fix the issue: "session" shared in different connections
## Code Before: import logging import botocore from kotocore.session import Session log = logging.getLogger(__name__) class KambooConnection(object): """ Kamboo connection with botocore session initialized """ session = botocore.session.get_session() def __init__(self, service_name="ec2", region_name="us-east-1", account_id=None, credentials=None): self.region = region_name self.account_id = account_id self.credentials = credentials if self.credentials: self.session.set_credentials(**self.credentials) Connection = Session(session=self.session).get_connection(service_name) self.conn = Connection(region_name=self.region) ## Instruction: Fix the issue: "session" shared in different connections ## Code After: import logging import botocore from kotocore.session import Session log = logging.getLogger(__name__) class KambooConnection(object): """ Kamboo connection with botocore session initialized """ def __init__(self, service_name="ec2", region_name="us-east-1", account_id=None, credentials=None): self.session = botocore.session.get_session() self.service = service_name self.region = region_name self.account_id = account_id self.credentials = credentials if self.credentials: self.session.set_credentials(**self.credentials) Connection = Session(session=self.session).get_connection(service_name) self.conn = Connection(region_name=self.region) def __repr__(self): return "KambooConnection: [%s, %s, %s]" % (self.account_id, self.region, self.service)
# ... existing code ... """ # ... modified code ... credentials=None): self.session = botocore.session.get_session() self.service = service_name self.region = region_name ... self.conn = Connection(region_name=self.region) def __repr__(self): return "KambooConnection: [%s, %s, %s]" % (self.account_id, self.region, self.service) # ... rest of the code ...
3e40f0bd3941a48e3b792573c0058f1cc339d5db
fabfile.py
fabfile.py
from fabric.api import * import slackbot_settings as settings from urllib import request, parse env.hosts = settings.DEPLOY_HOSTS def deploy(): slack("Deploy Started") with cd("/var/bot/slack-shogi"): run("git pull") run("supervisorctl reload") slack("Deploy Finished") def slack(text): if settings.WEBHOOK_URL: payload = ("payload={\"text\": \"" + parse.quote(text) + "\", \"username\": \"Mr.deploy\"}").encode("utf-8") request.urlopen(url=settings.WEBHOOK_URL, data=payload)
from fabric.api import * import slackbot_settings as settings from urllib import request, parse env.hosts = settings.DEPLOY_HOSTS def deploy(): slack("Deploy Started") try: with cd("/var/bot/slack-shogi"): run("git pull") run("supervisorctl reload") slack("Deploy Finished") except: slack("Deploy Failed") def slack(text): if settings.WEBHOOK_URL: payload = ("payload={\"text\": \"" + parse.quote(text) + "\", \"username\": \"Mr.deploy\"}").encode("utf-8") request.urlopen(url=settings.WEBHOOK_URL, data=payload)
Add fallback slack message for deploy script
Add fallback slack message for deploy script
Python
mit
setokinto/slack-shogi
from fabric.api import * import slackbot_settings as settings from urllib import request, parse env.hosts = settings.DEPLOY_HOSTS def deploy(): slack("Deploy Started") + try: - with cd("/var/bot/slack-shogi"): + with cd("/var/bot/slack-shogi"): - run("git pull") + run("git pull") - run("supervisorctl reload") + run("supervisorctl reload") - slack("Deploy Finished") + slack("Deploy Finished") + except: + slack("Deploy Failed") def slack(text): if settings.WEBHOOK_URL: payload = ("payload={\"text\": \"" + parse.quote(text) + "\", \"username\": \"Mr.deploy\"}").encode("utf-8") request.urlopen(url=settings.WEBHOOK_URL, data=payload) +
Add fallback slack message for deploy script
## Code Before: from fabric.api import * import slackbot_settings as settings from urllib import request, parse env.hosts = settings.DEPLOY_HOSTS def deploy(): slack("Deploy Started") with cd("/var/bot/slack-shogi"): run("git pull") run("supervisorctl reload") slack("Deploy Finished") def slack(text): if settings.WEBHOOK_URL: payload = ("payload={\"text\": \"" + parse.quote(text) + "\", \"username\": \"Mr.deploy\"}").encode("utf-8") request.urlopen(url=settings.WEBHOOK_URL, data=payload) ## Instruction: Add fallback slack message for deploy script ## Code After: from fabric.api import * import slackbot_settings as settings from urllib import request, parse env.hosts = settings.DEPLOY_HOSTS def deploy(): slack("Deploy Started") try: with cd("/var/bot/slack-shogi"): run("git pull") run("supervisorctl reload") slack("Deploy Finished") except: slack("Deploy Failed") def slack(text): if settings.WEBHOOK_URL: payload = ("payload={\"text\": \"" + parse.quote(text) + "\", \"username\": \"Mr.deploy\"}").encode("utf-8") request.urlopen(url=settings.WEBHOOK_URL, data=payload)
# ... existing code ... slack("Deploy Started") try: with cd("/var/bot/slack-shogi"): run("git pull") run("supervisorctl reload") slack("Deploy Finished") except: slack("Deploy Failed") # ... modified code ... request.urlopen(url=settings.WEBHOOK_URL, data=payload) # ... rest of the code ...
fa3ec9a764ca0d646588e908395367ce553981e1
tca/chat/views.py
tca/chat/views.py
from django.shortcuts import render from rest_framework import viewsets from chat.models import Member from chat.models import ChatRoom from chat.serializers import MemberSerializer from chat.serializers import ChatRoomSerializer class MemberViewSet(viewsets.ModelViewSet): model = Member serializer_class = MemberSerializer class ChatRoomViewSet(viewsets.ModelViewSet): model = ChatRoom serializer_class = ChatRoomSerializer
from django.shortcuts import render from django.shortcuts import get_object_or_404 from rest_framework import viewsets from rest_framework import status from rest_framework.decorators import action from rest_framework.response import Response from chat.models import Member from chat.models import ChatRoom from chat.serializers import MemberSerializer from chat.serializers import ChatRoomSerializer class MemberViewSet(viewsets.ModelViewSet): model = Member serializer_class = MemberSerializer class ChatRoomViewSet(viewsets.ModelViewSet): model = ChatRoom serializer_class = ChatRoomSerializer @action() def add_member(self, request, pk=None): chat_room = self.get_object() if 'lrz_id' not in request.DATA: # Invalid request return Response(status=status.HTTP_400_BAD_REQUEST) member = get_object_or_404(Member, lrz_id=request.DATA['lrz_id']) chat_room.members.add(member) return Response({ 'status': 'success', })
Add an action for adding members to a chat room
Add an action for adding members to a chat room Even though django-rest-framework supports a Ruby-on-Rails style of updating existing resources by issuing a PATCH or PUT request, such updates are unsafe and can cause race-conditions to lose some state. The implementation of this action isn't fully RESTful, but neither is the PATCH-style used here, so it is still a good choice for solving a possible race condition situation.
Python
bsd-3-clause
mlalic/TumCampusAppBackend,mlalic/TumCampusAppBackend
from django.shortcuts import render + from django.shortcuts import get_object_or_404 from rest_framework import viewsets + from rest_framework import status + from rest_framework.decorators import action + from rest_framework.response import Response from chat.models import Member from chat.models import ChatRoom from chat.serializers import MemberSerializer from chat.serializers import ChatRoomSerializer class MemberViewSet(viewsets.ModelViewSet): model = Member serializer_class = MemberSerializer class ChatRoomViewSet(viewsets.ModelViewSet): model = ChatRoom serializer_class = ChatRoomSerializer + @action() + def add_member(self, request, pk=None): + chat_room = self.get_object() + if 'lrz_id' not in request.DATA: + # Invalid request + return Response(status=status.HTTP_400_BAD_REQUEST) + + member = get_object_or_404(Member, lrz_id=request.DATA['lrz_id']) + chat_room.members.add(member) + + return Response({ + 'status': 'success', + }) +
Add an action for adding members to a chat room
## Code Before: from django.shortcuts import render from rest_framework import viewsets from chat.models import Member from chat.models import ChatRoom from chat.serializers import MemberSerializer from chat.serializers import ChatRoomSerializer class MemberViewSet(viewsets.ModelViewSet): model = Member serializer_class = MemberSerializer class ChatRoomViewSet(viewsets.ModelViewSet): model = ChatRoom serializer_class = ChatRoomSerializer ## Instruction: Add an action for adding members to a chat room ## Code After: from django.shortcuts import render from django.shortcuts import get_object_or_404 from rest_framework import viewsets from rest_framework import status from rest_framework.decorators import action from rest_framework.response import Response from chat.models import Member from chat.models import ChatRoom from chat.serializers import MemberSerializer from chat.serializers import ChatRoomSerializer class MemberViewSet(viewsets.ModelViewSet): model = Member serializer_class = MemberSerializer class ChatRoomViewSet(viewsets.ModelViewSet): model = ChatRoom serializer_class = ChatRoomSerializer @action() def add_member(self, request, pk=None): chat_room = self.get_object() if 'lrz_id' not in request.DATA: # Invalid request return Response(status=status.HTTP_400_BAD_REQUEST) member = get_object_or_404(Member, lrz_id=request.DATA['lrz_id']) chat_room.members.add(member) return Response({ 'status': 'success', })
// ... existing code ... from django.shortcuts import render from django.shortcuts import get_object_or_404 // ... modified code ... from rest_framework import viewsets from rest_framework import status from rest_framework.decorators import action from rest_framework.response import Response ... serializer_class = ChatRoomSerializer @action() def add_member(self, request, pk=None): chat_room = self.get_object() if 'lrz_id' not in request.DATA: # Invalid request return Response(status=status.HTTP_400_BAD_REQUEST) member = get_object_or_404(Member, lrz_id=request.DATA['lrz_id']) chat_room.members.add(member) return Response({ 'status': 'success', }) // ... rest of the code ...
ff85fc05e179e451dabb1f20781dfc5a90314d71
scripts/adb-wrapper.py
scripts/adb-wrapper.py
import subprocess import sys import re # Note: no output will be printed until the entire test suite has finished result = subprocess.run(sys.argv[1], shell=True, stdout=subprocess.PIPE, stderr=subprocess.PIPE, universal_newlines=True) successRegex = re.compile('OK \(\d+ tests\)') print(result.stderr) print(result.stdout) if successRegex.search(result.stderr + result.stdout): sys.exit(0) else: sys.exit(1)
import subprocess import sys import re # Note: no output will be printed until the entire test suite has finished p = subprocess.Popen(sys.argv[1], shell=True, stdout=subprocess.PIPE, stderr=subprocess.PIPE, universal_newlines=True) stdoutResult, stderrResult = p.communicate() successRegex = re.compile('OK \(\d+ tests\)') print(stdoutResult) print(stderrResult) if successRegex.search(stderrResult + stdoutResult): sys.exit(0) else: sys.exit(1)
Refactor the python wrapper script because apparently apt-get doesn't install 3.5, and subprocess.run() is only in 3.5
Refactor the python wrapper script because apparently apt-get doesn't install 3.5, and subprocess.run() is only in 3.5
Python
apache-2.0
sbosley/squidb,yahoo/squidb,yahoo/squidb,sbosley/squidb,sbosley/squidb,sbosley/squidb,sbosley/squidb,yahoo/squidb,yahoo/squidb,yahoo/squidb
import subprocess import sys import re # Note: no output will be printed until the entire test suite has finished - result = subprocess.run(sys.argv[1], shell=True, stdout=subprocess.PIPE, stderr=subprocess.PIPE, universal_newlines=True) + p = subprocess.Popen(sys.argv[1], shell=True, stdout=subprocess.PIPE, stderr=subprocess.PIPE, universal_newlines=True) + stdoutResult, stderrResult = p.communicate() successRegex = re.compile('OK \(\d+ tests\)') - print(result.stderr) - print(result.stdout) + print(stdoutResult) + print(stderrResult) - if successRegex.search(result.stderr + result.stdout): + if successRegex.search(stderrResult + stdoutResult): sys.exit(0) else: sys.exit(1)
Refactor the python wrapper script because apparently apt-get doesn't install 3.5, and subprocess.run() is only in 3.5
## Code Before: import subprocess import sys import re # Note: no output will be printed until the entire test suite has finished result = subprocess.run(sys.argv[1], shell=True, stdout=subprocess.PIPE, stderr=subprocess.PIPE, universal_newlines=True) successRegex = re.compile('OK \(\d+ tests\)') print(result.stderr) print(result.stdout) if successRegex.search(result.stderr + result.stdout): sys.exit(0) else: sys.exit(1) ## Instruction: Refactor the python wrapper script because apparently apt-get doesn't install 3.5, and subprocess.run() is only in 3.5 ## Code After: import subprocess import sys import re # Note: no output will be printed until the entire test suite has finished p = subprocess.Popen(sys.argv[1], shell=True, stdout=subprocess.PIPE, stderr=subprocess.PIPE, universal_newlines=True) stdoutResult, stderrResult = p.communicate() successRegex = re.compile('OK \(\d+ tests\)') print(stdoutResult) print(stderrResult) if successRegex.search(stderrResult + stdoutResult): sys.exit(0) else: sys.exit(1)
... # Note: no output will be printed until the entire test suite has finished p = subprocess.Popen(sys.argv[1], shell=True, stdout=subprocess.PIPE, stderr=subprocess.PIPE, universal_newlines=True) stdoutResult, stderrResult = p.communicate() ... print(stdoutResult) print(stderrResult) if successRegex.search(stderrResult + stdoutResult): sys.exit(0) ...
c3a6554ce24781a3eaa9229f9609a09e4e018069
lcd_restful/__main__.py
lcd_restful/__main__.py
import sys from getopt import getopt, GetoptError from .api import Server from .fake import FakeLcdApi USAGE = """\ Usage %s [-h|--help] \t-h or --help\tThis help message """ def get_args(args): try: opts, args = getopt(args[1:], 'hf', ['help', 'fake']) except GetoptError as e: print('GetoptError %s' % e) sys.exit(2) ret_args = {} ret_args['fake'] = False for opt, arg in opts: if opt in ['-h', '--help']: print(USAGE % args[0]) sys.exit(0) elif opt in ['-f', '--fake']: ret_args['fake'] = True else: print(USAGE % args[0]) sys.exit(1) return ret_args def main_serv(clargs=sys.argv): opts = get_args(clargs) lcd = None if opts['fake']: lcd = FakeLcdApi() s = Server(lcd=lcd) s.run() return 0 if __name__ == "__main__": sys.exit(main_serv())
import sys from getopt import getopt, GetoptError from .api import Server from .lcd import Lcd USAGE = """\ Usage %s [-h|--help] [-f|--fake] \t-h or --help\tThis help message \t-f or --fake\tIf on RPi, use FakeHw """ def get_args(args): arg0 = args[0] try: opts, args = getopt(args[1:], 'hf', ['help', 'fake']) except GetoptError as e: print('GetoptError %s' % e) sys.exit(2) ret_args = {} ret_args['fake'] = False for opt, arg in opts: if opt in ['-h', '--help']: print(USAGE % arg0) sys.exit(0) elif opt in ['-f', '--fake']: ret_args['fake'] = True else: print(USAGE % arg0) sys.exit(1) return ret_args def main_serv(clargs=sys.argv): opts = get_args(clargs) s = Server(lcd=Lcd(opts['fake'])) s.run() return 0 if __name__ == "__main__": sys.exit(main_serv())
Update main Lcd import and init, and fix help msg
Update main Lcd import and init, and fix help msg
Python
mit
rfarley3/lcd-restful,rfarley3/lcd-restful
import sys from getopt import getopt, GetoptError from .api import Server - from .fake import FakeLcdApi + from .lcd import Lcd USAGE = """\ - Usage %s [-h|--help] + Usage %s [-h|--help] [-f|--fake] \t-h or --help\tThis help message + \t-f or --fake\tIf on RPi, use FakeHw """ def get_args(args): + arg0 = args[0] try: opts, args = getopt(args[1:], 'hf', ['help', 'fake']) except GetoptError as e: print('GetoptError %s' % e) sys.exit(2) ret_args = {} ret_args['fake'] = False for opt, arg in opts: if opt in ['-h', '--help']: - print(USAGE % args[0]) + print(USAGE % arg0) sys.exit(0) elif opt in ['-f', '--fake']: ret_args['fake'] = True else: - print(USAGE % args[0]) + print(USAGE % arg0) sys.exit(1) return ret_args def main_serv(clargs=sys.argv): opts = get_args(clargs) + s = Server(lcd=Lcd(opts['fake'])) - lcd = None - if opts['fake']: - lcd = FakeLcdApi() - s = Server(lcd=lcd) s.run() return 0 if __name__ == "__main__": sys.exit(main_serv()) +
Update main Lcd import and init, and fix help msg
## Code Before: import sys from getopt import getopt, GetoptError from .api import Server from .fake import FakeLcdApi USAGE = """\ Usage %s [-h|--help] \t-h or --help\tThis help message """ def get_args(args): try: opts, args = getopt(args[1:], 'hf', ['help', 'fake']) except GetoptError as e: print('GetoptError %s' % e) sys.exit(2) ret_args = {} ret_args['fake'] = False for opt, arg in opts: if opt in ['-h', '--help']: print(USAGE % args[0]) sys.exit(0) elif opt in ['-f', '--fake']: ret_args['fake'] = True else: print(USAGE % args[0]) sys.exit(1) return ret_args def main_serv(clargs=sys.argv): opts = get_args(clargs) lcd = None if opts['fake']: lcd = FakeLcdApi() s = Server(lcd=lcd) s.run() return 0 if __name__ == "__main__": sys.exit(main_serv()) ## Instruction: Update main Lcd import and init, and fix help msg ## Code After: import sys from getopt import getopt, GetoptError from .api import Server from .lcd import Lcd USAGE = """\ Usage %s [-h|--help] [-f|--fake] \t-h or --help\tThis help message \t-f or --fake\tIf on RPi, use FakeHw """ def get_args(args): arg0 = args[0] try: opts, args = getopt(args[1:], 'hf', ['help', 'fake']) except GetoptError as e: print('GetoptError %s' % e) sys.exit(2) ret_args = {} ret_args['fake'] = False for opt, arg in opts: if opt in ['-h', '--help']: print(USAGE % arg0) sys.exit(0) elif opt in ['-f', '--fake']: ret_args['fake'] = True else: print(USAGE % arg0) sys.exit(1) return ret_args def main_serv(clargs=sys.argv): opts = get_args(clargs) s = Server(lcd=Lcd(opts['fake'])) s.run() return 0 if __name__ == "__main__": sys.exit(main_serv())
// ... existing code ... from .api import Server from .lcd import Lcd // ... modified code ... USAGE = """\ Usage %s [-h|--help] [-f|--fake] \t-h or --help\tThis help message \t-f or --fake\tIf on RPi, use FakeHw ... def get_args(args): arg0 = args[0] try: ... if opt in ['-h', '--help']: print(USAGE % arg0) sys.exit(0) ... else: print(USAGE % arg0) sys.exit(1) ... opts = get_args(clargs) s = Server(lcd=Lcd(opts['fake'])) s.run() ... sys.exit(main_serv()) // ... rest of the code ...
bc5fa08e84cd11349dc44c3065b7b5380d60ebd9
raven/contrib/django/handlers.py
raven/contrib/django/handlers.py
from __future__ import absolute_import import logging from raven.handlers.logging import SentryHandler as BaseSentryHandler class SentryHandler(BaseSentryHandler): def __init__(self): logging.Handler.__init__(self) def _get_client(self): from raven.contrib.django.models import client return client client = property(_get_client) def _emit(self, record): request = getattr(record, 'request', None) return super(SentryHandler, self)._emit(record, request=request)
from __future__ import absolute_import import logging from raven.handlers.logging import SentryHandler as BaseSentryHandler class SentryHandler(BaseSentryHandler): def __init__(self, level=logging.NOTSET): logging.Handler.__init__(self, level=level) def _get_client(self): from raven.contrib.django.models import client return client client = property(_get_client) def _emit(self, record): request = getattr(record, 'request', None) return super(SentryHandler, self)._emit(record, request=request)
Allow level param in Django SentryHandler.__init__
Allow level param in Django SentryHandler.__init__ For consistency with superclass and with logging.Handler
Python
bsd-3-clause
arthurlogilab/raven-python,inspirehep/raven-python,ewdurbin/raven-python,jbarbuto/raven-python,getsentry/raven-python,dbravender/raven-python,jbarbuto/raven-python,Photonomie/raven-python,percipient/raven-python,akheron/raven-python,icereval/raven-python,nikolas/raven-python,akalipetis/raven-python,someonehan/raven-python,nikolas/raven-python,ewdurbin/raven-python,inspirehep/raven-python,icereval/raven-python,danriti/raven-python,danriti/raven-python,Goldmund-Wyldebeast-Wunderliebe/raven-python,Goldmund-Wyldebeast-Wunderliebe/raven-python,someonehan/raven-python,danriti/raven-python,Goldmund-Wyldebeast-Wunderliebe/raven-python,akheron/raven-python,akheron/raven-python,arthurlogilab/raven-python,hzy/raven-python,nikolas/raven-python,arthurlogilab/raven-python,jbarbuto/raven-python,inspirehep/raven-python,smarkets/raven-python,arthurlogilab/raven-python,smarkets/raven-python,johansteffner/raven-python,akalipetis/raven-python,lepture/raven-python,getsentry/raven-python,akalipetis/raven-python,ronaldevers/raven-python,johansteffner/raven-python,percipient/raven-python,icereval/raven-python,jmp0xf/raven-python,jmagnusson/raven-python,lepture/raven-python,smarkets/raven-python,johansteffner/raven-python,Photonomie/raven-python,someonehan/raven-python,ronaldevers/raven-python,dbravender/raven-python,ronaldevers/raven-python,lepture/raven-python,percipient/raven-python,jbarbuto/raven-python,jmagnusson/raven-python,jmp0xf/raven-python,smarkets/raven-python,recht/raven-python,jmp0xf/raven-python,icereval/raven-python,jmagnusson/raven-python,Photonomie/raven-python,Goldmund-Wyldebeast-Wunderliebe/raven-python,dbravender/raven-python,recht/raven-python,nikolas/raven-python,getsentry/raven-python,hzy/raven-python,inspirehep/raven-python,recht/raven-python,ewdurbin/raven-python,hzy/raven-python
from __future__ import absolute_import import logging from raven.handlers.logging import SentryHandler as BaseSentryHandler class SentryHandler(BaseSentryHandler): - def __init__(self): + def __init__(self, level=logging.NOTSET): - logging.Handler.__init__(self) + logging.Handler.__init__(self, level=level) def _get_client(self): from raven.contrib.django.models import client return client client = property(_get_client) def _emit(self, record): request = getattr(record, 'request', None) return super(SentryHandler, self)._emit(record, request=request)
Allow level param in Django SentryHandler.__init__
## Code Before: from __future__ import absolute_import import logging from raven.handlers.logging import SentryHandler as BaseSentryHandler class SentryHandler(BaseSentryHandler): def __init__(self): logging.Handler.__init__(self) def _get_client(self): from raven.contrib.django.models import client return client client = property(_get_client) def _emit(self, record): request = getattr(record, 'request', None) return super(SentryHandler, self)._emit(record, request=request) ## Instruction: Allow level param in Django SentryHandler.__init__ ## Code After: from __future__ import absolute_import import logging from raven.handlers.logging import SentryHandler as BaseSentryHandler class SentryHandler(BaseSentryHandler): def __init__(self, level=logging.NOTSET): logging.Handler.__init__(self, level=level) def _get_client(self): from raven.contrib.django.models import client return client client = property(_get_client) def _emit(self, record): request = getattr(record, 'request', None) return super(SentryHandler, self)._emit(record, request=request)
// ... existing code ... class SentryHandler(BaseSentryHandler): def __init__(self, level=logging.NOTSET): logging.Handler.__init__(self, level=level) // ... rest of the code ...
d32e1d8349c115027e3095d61f8fa882fca1ab52
functions/test_lambda.py
functions/test_lambda.py
""" Explore how python works with lambda expressions. """ import unittest import string import random class TestGetWebsites(unittest.TestCase): def test_closure(self, m): """ See that python supports closures similar to JavaScript """ def gibberish(): """ Some random string """ return ''.join([random.choice(string.ascii_letters) for i in range(20)]) value1 = gibberish() value2 = gibberish() result = (lambda x: lambda y: x)(value1)(value2) self.assertEqual(result, value1) if __name__ == "__main__": unittest.main()
""" Explore how python works with lambda expressions. """ import unittest import string import random class TestGetWebsites(unittest.TestCase): def test_closure(self): """ See that python supports closures similar to JavaScript """ def gibberish(): """ Some random string """ return ''.join([random.choice(string.ascii_letters) for i in range(20)]) value1 = gibberish() result = (lambda x: lambda: x)(value1)() self.assertEqual(result, value1) def test_closure_ids(self): """ Show how a variable passed to a function remains in scope even after it returns """ def make_lambdas(num): """ Build a lambda generator """ for i in range(num): func = lambda x: lambda: x yield func(i) functions = list(make_lambdas(random.randint(1, 10))) random_index = random.randint(0, len(functions)-1) random_function = functions[random_index]() print("{0} equals {1}".format(random_index, random_function)) self.assertEqual(random_function, random_index) if __name__ == "__main__": unittest.main()
Add a test to generate a range of functions with closures. Check if it returns the expected value
Add a test to generate a range of functions with closures. Check if it returns the expected value
Python
mit
b-ritter/python-notes,b-ritter/python-notes
""" Explore how python works with lambda expressions. """ import unittest import string import random class TestGetWebsites(unittest.TestCase): - def test_closure(self, m): + def test_closure(self): """ See that python supports closures similar to JavaScript """ def gibberish(): """ Some random string """ return ''.join([random.choice(string.ascii_letters) for i in range(20)]) value1 = gibberish() - value2 = gibberish() - result = (lambda x: lambda y: x)(value1)(value2) + result = (lambda x: lambda: x)(value1)() self.assertEqual(result, value1) + + def test_closure_ids(self): + """ Show how a variable passed to a function remains in scope + even after it returns """ + def make_lambdas(num): + """ Build a lambda generator """ + for i in range(num): + func = lambda x: lambda: x + yield func(i) + functions = list(make_lambdas(random.randint(1, 10))) + random_index = random.randint(0, len(functions)-1) + random_function = functions[random_index]() + print("{0} equals {1}".format(random_index, random_function)) + self.assertEqual(random_function, random_index) if __name__ == "__main__": unittest.main()
Add a test to generate a range of functions with closures. Check if it returns the expected value
## Code Before: """ Explore how python works with lambda expressions. """ import unittest import string import random class TestGetWebsites(unittest.TestCase): def test_closure(self, m): """ See that python supports closures similar to JavaScript """ def gibberish(): """ Some random string """ return ''.join([random.choice(string.ascii_letters) for i in range(20)]) value1 = gibberish() value2 = gibberish() result = (lambda x: lambda y: x)(value1)(value2) self.assertEqual(result, value1) if __name__ == "__main__": unittest.main() ## Instruction: Add a test to generate a range of functions with closures. Check if it returns the expected value ## Code After: """ Explore how python works with lambda expressions. """ import unittest import string import random class TestGetWebsites(unittest.TestCase): def test_closure(self): """ See that python supports closures similar to JavaScript """ def gibberish(): """ Some random string """ return ''.join([random.choice(string.ascii_letters) for i in range(20)]) value1 = gibberish() result = (lambda x: lambda: x)(value1)() self.assertEqual(result, value1) def test_closure_ids(self): """ Show how a variable passed to a function remains in scope even after it returns """ def make_lambdas(num): """ Build a lambda generator """ for i in range(num): func = lambda x: lambda: x yield func(i) functions = list(make_lambdas(random.randint(1, 10))) random_index = random.randint(0, len(functions)-1) random_function = functions[random_index]() print("{0} equals {1}".format(random_index, random_function)) self.assertEqual(random_function, random_index) if __name__ == "__main__": unittest.main()
// ... existing code ... def test_closure(self): """ See that python supports closures similar to JavaScript """ // ... modified code ... value1 = gibberish() result = (lambda x: lambda: x)(value1)() self.assertEqual(result, value1) def test_closure_ids(self): """ Show how a variable passed to a function remains in scope even after it returns """ def make_lambdas(num): """ Build a lambda generator """ for i in range(num): func = lambda x: lambda: x yield func(i) functions = list(make_lambdas(random.randint(1, 10))) random_index = random.randint(0, len(functions)-1) random_function = functions[random_index]() print("{0} equals {1}".format(random_index, random_function)) self.assertEqual(random_function, random_index) // ... rest of the code ...
42e16bf376a64995a8b70a91829a82d7b0f3e1a1
gameanalysis/__main__.py
gameanalysis/__main__.py
"""Command line module""" import argparse import pkgutil import sys import gameanalysis from gameanalysis import script def create_parser(): """Create the default parser""" modules = [imp.find_module(name).load_module(name) for imp, name, _ in pkgutil.iter_modules(script.__path__)] parser = argparse.ArgumentParser( description="""Command line access to the game analysis toolkit.""") parser.add_argument('-V', '--version', action='version', version='%(prog)s {}'.format(gameanalysis.__version__)) subparsers = parser.add_subparsers( title='commands', dest='command', metavar='<command>', help="""The commands to execute. Available commands are:""") for module in modules: subparser = module.add_parser(subparsers) subparser.main = module.main return parser, subparsers.choices def amain(*argv): """Entry point for game analysis""" parser, commands = create_parser() args = parser.parse_args(argv) if args.command is None: parser.print_help() sys.exit(1) else: commands[args.command].main(args) def main(): """Entry point for game analysis""" amain(*sys.argv[1:]) if __name__ == '__main__': main()
"""Command line module""" import argparse import logging import pkgutil import sys import gameanalysis from gameanalysis import script def create_parser(): """Create the default parser""" modules = [imp.find_module(name).load_module(name) for imp, name, _ in pkgutil.iter_modules(script.__path__)] parser = argparse.ArgumentParser( description="""Command line access to the game analysis toolkit.""") parser.add_argument('-V', '--version', action='version', version='%(prog)s {}'.format(gameanalysis.__version__)) parser.add_argument( '-v', '--verbose', action='count', default=0, help="""Set the verbosity level depending on the number of times specified, up to a maximum of three.""") subparsers = parser.add_subparsers( title='commands', dest='command', metavar='<command>', help="""The commands to execute. Available commands are:""") for module in modules: subparser = module.add_parser(subparsers) subparser.main = module.main return parser, subparsers.choices def amain(*argv): """Entry point for game analysis""" parser, commands = create_parser() args = parser.parse_args(argv) if args.command is None: parser.print_help() sys.exit(1) else: logging.basicConfig(level=40 - 10 * min(args.verbose, 3)) commands[args.command].main(args) def main(): """Entry point for game analysis""" amain(*sys.argv[1:]) if __name__ == '__main__': main()
Add logging verbosity to game analysis
Add logging verbosity to game analysis
Python
apache-2.0
egtaonline/GameAnalysis
"""Command line module""" import argparse + import logging import pkgutil import sys import gameanalysis from gameanalysis import script def create_parser(): """Create the default parser""" modules = [imp.find_module(name).load_module(name) for imp, name, _ in pkgutil.iter_modules(script.__path__)] parser = argparse.ArgumentParser( description="""Command line access to the game analysis toolkit.""") parser.add_argument('-V', '--version', action='version', version='%(prog)s {}'.format(gameanalysis.__version__)) + parser.add_argument( + '-v', '--verbose', action='count', default=0, help="""Set the verbosity + level depending on the number of times specified, up to a maximum of + three.""") subparsers = parser.add_subparsers( title='commands', dest='command', metavar='<command>', help="""The commands to execute. Available commands are:""") for module in modules: subparser = module.add_parser(subparsers) subparser.main = module.main return parser, subparsers.choices def amain(*argv): """Entry point for game analysis""" parser, commands = create_parser() args = parser.parse_args(argv) if args.command is None: parser.print_help() sys.exit(1) else: + logging.basicConfig(level=40 - 10 * min(args.verbose, 3)) commands[args.command].main(args) def main(): """Entry point for game analysis""" amain(*sys.argv[1:]) if __name__ == '__main__': main()
Add logging verbosity to game analysis
## Code Before: """Command line module""" import argparse import pkgutil import sys import gameanalysis from gameanalysis import script def create_parser(): """Create the default parser""" modules = [imp.find_module(name).load_module(name) for imp, name, _ in pkgutil.iter_modules(script.__path__)] parser = argparse.ArgumentParser( description="""Command line access to the game analysis toolkit.""") parser.add_argument('-V', '--version', action='version', version='%(prog)s {}'.format(gameanalysis.__version__)) subparsers = parser.add_subparsers( title='commands', dest='command', metavar='<command>', help="""The commands to execute. Available commands are:""") for module in modules: subparser = module.add_parser(subparsers) subparser.main = module.main return parser, subparsers.choices def amain(*argv): """Entry point for game analysis""" parser, commands = create_parser() args = parser.parse_args(argv) if args.command is None: parser.print_help() sys.exit(1) else: commands[args.command].main(args) def main(): """Entry point for game analysis""" amain(*sys.argv[1:]) if __name__ == '__main__': main() ## Instruction: Add logging verbosity to game analysis ## Code After: """Command line module""" import argparse import logging import pkgutil import sys import gameanalysis from gameanalysis import script def create_parser(): """Create the default parser""" modules = [imp.find_module(name).load_module(name) for imp, name, _ in pkgutil.iter_modules(script.__path__)] parser = argparse.ArgumentParser( description="""Command line access to the game analysis toolkit.""") parser.add_argument('-V', '--version', action='version', version='%(prog)s {}'.format(gameanalysis.__version__)) parser.add_argument( '-v', '--verbose', action='count', default=0, help="""Set the verbosity level depending on the number of times specified, up to a maximum of three.""") subparsers = parser.add_subparsers( title='commands', dest='command', metavar='<command>', help="""The commands to execute. Available commands are:""") for module in modules: subparser = module.add_parser(subparsers) subparser.main = module.main return parser, subparsers.choices def amain(*argv): """Entry point for game analysis""" parser, commands = create_parser() args = parser.parse_args(argv) if args.command is None: parser.print_help() sys.exit(1) else: logging.basicConfig(level=40 - 10 * min(args.verbose, 3)) commands[args.command].main(args) def main(): """Entry point for game analysis""" amain(*sys.argv[1:]) if __name__ == '__main__': main()
# ... existing code ... import argparse import logging import pkgutil # ... modified code ... version='%(prog)s {}'.format(gameanalysis.__version__)) parser.add_argument( '-v', '--verbose', action='count', default=0, help="""Set the verbosity level depending on the number of times specified, up to a maximum of three.""") subparsers = parser.add_subparsers( ... else: logging.basicConfig(level=40 - 10 * min(args.verbose, 3)) commands[args.command].main(args) # ... rest of the code ...
d902045e991cc778dabe31e34a6dcd119e19ccd0
attributes/license/main.py
attributes/license/main.py
from utilities import url_to_json def run(project_id, repo_path, cursor, **options): query = 'SELECT url FROM projects WHERE id = ' + str(project_id) cursor.execute(query) record = cursor.fetchone() full_url = record[0].rstrip() json_response = url_to_json(full_url, headers={ 'Accept': 'application/vnd.github.drax-preview+json' } ) if 'license' in json_response: result = 1 else: result = 0 return result if __name__ == '__main__': print("Attribute plugins are not meant to be executed directly.")
from core import tokenize from utilities import url_to_json def run(project_id, repo_path, cursor, **options): query = 'SELECT url FROM projects WHERE id = ' + str(project_id) cursor.execute(query) record = cursor.fetchone() full_url = tokenize(record[0].rstrip()) json_response = url_to_json(full_url, headers={ 'Accept': 'application/vnd.github.drax-preview+json' } ) result = 'license' in json_response return result, int(result) if __name__ == '__main__': print("Attribute plugins are not meant to be executed directly.")
Update license attribute to return binary and raw result
Update license attribute to return binary and raw result
Python
apache-2.0
RepoReapers/reaper,RepoReapers/reaper,RepoReapers/reaper,RepoReapers/reaper
+ from core import tokenize from utilities import url_to_json def run(project_id, repo_path, cursor, **options): query = 'SELECT url FROM projects WHERE id = ' + str(project_id) cursor.execute(query) record = cursor.fetchone() - full_url = record[0].rstrip() + full_url = tokenize(record[0].rstrip()) json_response = url_to_json(full_url, headers={ 'Accept': 'application/vnd.github.drax-preview+json' } ) - if 'license' in json_response: + result = 'license' in json_response + return result, int(result) - result = 1 - else: - result = 0 - - return result if __name__ == '__main__': print("Attribute plugins are not meant to be executed directly.")
Update license attribute to return binary and raw result
## Code Before: from utilities import url_to_json def run(project_id, repo_path, cursor, **options): query = 'SELECT url FROM projects WHERE id = ' + str(project_id) cursor.execute(query) record = cursor.fetchone() full_url = record[0].rstrip() json_response = url_to_json(full_url, headers={ 'Accept': 'application/vnd.github.drax-preview+json' } ) if 'license' in json_response: result = 1 else: result = 0 return result if __name__ == '__main__': print("Attribute plugins are not meant to be executed directly.") ## Instruction: Update license attribute to return binary and raw result ## Code After: from core import tokenize from utilities import url_to_json def run(project_id, repo_path, cursor, **options): query = 'SELECT url FROM projects WHERE id = ' + str(project_id) cursor.execute(query) record = cursor.fetchone() full_url = tokenize(record[0].rstrip()) json_response = url_to_json(full_url, headers={ 'Accept': 'application/vnd.github.drax-preview+json' } ) result = 'license' in json_response return result, int(result) if __name__ == '__main__': print("Attribute plugins are not meant to be executed directly.")
... from core import tokenize from utilities import url_to_json ... full_url = tokenize(record[0].rstrip()) json_response = url_to_json(full_url, headers={ ... result = 'license' in json_response return result, int(result) ...
5977eb82f2614efe8cde843913db62a93c7978f5
navigation_extensions.py
navigation_extensions.py
from django.utils.text import capfirst from django.utils.translation import ugettext_lazy as _ from feincms.module.page.extensions.navigation import NavigationExtension, PagePretender class ZivinetzNavigationExtension(NavigationExtension): name = _('Zivinetz navigation extension') def children(self, page, **kwargs): request = kwargs.get('request') if request.user.is_authenticated() and request.user.is_staff: urls = [ (_('scheduling'), 'admin/scheduling/'), (_('waitlist'), 'admin/waitlist/'), (_('drudges'), 'admin/drudges/'), (_('assignments'), 'admin/assignments/'), (_('job references'), 'admin/jobreferences/'), (_('expense reports'), 'admin/expense_reports/'), (_('regional offices'), 'admin/regional_offices/'), (_('scope statements'), 'admin/scope_statements/'), (_('specifications'), 'admin/specifications/'), ] else: urls = [ (_('dashboard'), 'dashboard/'), (_('profile'), 'profile/'), ] return [PagePretender( title=capfirst(title), url='%s%s' % (page.get_navigation_url(), url), level=page.level+1, tree_id=page.tree_id, ) for title, url in urls]
from django.utils.text import capfirst from django.utils.translation import ugettext_lazy as _ from feincms.module.page.extensions.navigation import NavigationExtension, PagePretender class ZivinetzNavigationExtension(NavigationExtension): name = _('Zivinetz navigation extension') def children(self, page, **kwargs): request = kwargs.get('request') if request.user.is_authenticated() and request.user.is_staff: urls = [ (_('scheduling'), 'admin/scheduling/'), (_('waitlist'), 'admin/waitlist/'), (_('drudges'), 'admin/drudges/'), (_('assignments'), 'admin/assignments/'), (_('expense reports'), 'admin/expense_reports/'), (_('regional offices'), 'admin/regional_offices/'), (_('scope statements'), 'admin/scope_statements/'), (_('specifications'), 'admin/specifications/'), ] else: urls = [ (_('dashboard'), 'dashboard/'), (_('profile'), 'profile/'), ] return [PagePretender( title=capfirst(title), url='%s%s' % (page.get_navigation_url(), url), level=page.level+1, tree_id=page.tree_id, ) for title, url in urls]
Remove job references from navigation
Remove job references from navigation
Python
mit
matthiask/zivinetz,matthiask/zivinetz,matthiask/zivinetz,matthiask/zivinetz
from django.utils.text import capfirst from django.utils.translation import ugettext_lazy as _ from feincms.module.page.extensions.navigation import NavigationExtension, PagePretender class ZivinetzNavigationExtension(NavigationExtension): name = _('Zivinetz navigation extension') def children(self, page, **kwargs): request = kwargs.get('request') if request.user.is_authenticated() and request.user.is_staff: urls = [ (_('scheduling'), 'admin/scheduling/'), (_('waitlist'), 'admin/waitlist/'), (_('drudges'), 'admin/drudges/'), (_('assignments'), 'admin/assignments/'), - (_('job references'), 'admin/jobreferences/'), (_('expense reports'), 'admin/expense_reports/'), (_('regional offices'), 'admin/regional_offices/'), (_('scope statements'), 'admin/scope_statements/'), (_('specifications'), 'admin/specifications/'), ] else: urls = [ (_('dashboard'), 'dashboard/'), (_('profile'), 'profile/'), ] return [PagePretender( title=capfirst(title), url='%s%s' % (page.get_navigation_url(), url), level=page.level+1, tree_id=page.tree_id, ) for title, url in urls]
Remove job references from navigation
## Code Before: from django.utils.text import capfirst from django.utils.translation import ugettext_lazy as _ from feincms.module.page.extensions.navigation import NavigationExtension, PagePretender class ZivinetzNavigationExtension(NavigationExtension): name = _('Zivinetz navigation extension') def children(self, page, **kwargs): request = kwargs.get('request') if request.user.is_authenticated() and request.user.is_staff: urls = [ (_('scheduling'), 'admin/scheduling/'), (_('waitlist'), 'admin/waitlist/'), (_('drudges'), 'admin/drudges/'), (_('assignments'), 'admin/assignments/'), (_('job references'), 'admin/jobreferences/'), (_('expense reports'), 'admin/expense_reports/'), (_('regional offices'), 'admin/regional_offices/'), (_('scope statements'), 'admin/scope_statements/'), (_('specifications'), 'admin/specifications/'), ] else: urls = [ (_('dashboard'), 'dashboard/'), (_('profile'), 'profile/'), ] return [PagePretender( title=capfirst(title), url='%s%s' % (page.get_navigation_url(), url), level=page.level+1, tree_id=page.tree_id, ) for title, url in urls] ## Instruction: Remove job references from navigation ## Code After: from django.utils.text import capfirst from django.utils.translation import ugettext_lazy as _ from feincms.module.page.extensions.navigation import NavigationExtension, PagePretender class ZivinetzNavigationExtension(NavigationExtension): name = _('Zivinetz navigation extension') def children(self, page, **kwargs): request = kwargs.get('request') if request.user.is_authenticated() and request.user.is_staff: urls = [ (_('scheduling'), 'admin/scheduling/'), (_('waitlist'), 'admin/waitlist/'), (_('drudges'), 'admin/drudges/'), (_('assignments'), 'admin/assignments/'), (_('expense reports'), 'admin/expense_reports/'), (_('regional offices'), 'admin/regional_offices/'), (_('scope statements'), 'admin/scope_statements/'), (_('specifications'), 'admin/specifications/'), ] else: urls = [ (_('dashboard'), 'dashboard/'), (_('profile'), 'profile/'), ] return [PagePretender( title=capfirst(title), url='%s%s' % (page.get_navigation_url(), url), level=page.level+1, tree_id=page.tree_id, ) for title, url in urls]
... (_('assignments'), 'admin/assignments/'), (_('expense reports'), 'admin/expense_reports/'), ...
267c6591ef7ab3354b0965902585203fbfe04dee
pybot/http_client.py
pybot/http_client.py
import requests, json from resources.urls import FACEBOOK_MESSAGES_POST_URL class HttpClient(): """ Client which excutes the call to facebook's messenger api """ def submit_request(self, path, method, payload, completion): assert len(path) > 0 path = self.get_api_url(path) headers = self.get_headers() if method == 'GET': request = requests.get(path, headers=headers) if request.status_code is not 200: error = self.get_error_from_request(request) completion(None, error) else: json_ = self.get_json(request.text) completion(json_, None) elif method == 'POST': raise NotImplementedError def get_error_from_request(self, request): return { 'error': self.get_json(request.text) } def get_json(self, string): return json.loads(string) def get_api_url(self, path): return FACEBOOK_MESSAGES_POST_URL + path def get_headers(self): return {}
import requests, json from resources.urls import FACEBOOK_MESSAGES_POST_URL class HttpClient(): """ Client which excutes the call to facebook's messenger api """ def submit_request(self, path, method, payload, completion): assert len(path) > 0 path = self.get_api_url(path) headers = self.get_headers() if method == 'GET': response = requests.get(path, headers=headers) if response.status_code is not 200: error = self.get_error_from_request(response) completion(None, error) else: json_ = self.get_json(response.text) completion(json_, None) elif method == 'POST': re def get_error_from_response(self, response): return { 'error': self.get_json(response.text) } def get_json(self, string): return json.loads(string) def get_api_url(self, path): return FACEBOOK_MESSAGES_POST_URL + path def get_headers(self): return {}
Change wording to use response
Change wording to use response
Python
mit
ben-cunningham/pybot,ben-cunningham/python-messenger-bot
import requests, json from resources.urls import FACEBOOK_MESSAGES_POST_URL class HttpClient(): """ Client which excutes the call to facebook's messenger api """ def submit_request(self, path, method, payload, completion): assert len(path) > 0 path = self.get_api_url(path) headers = self.get_headers() if method == 'GET': - request = requests.get(path, headers=headers) + response = requests.get(path, headers=headers) - if request.status_code is not 200: + if response.status_code is not 200: - error = self.get_error_from_request(request) + error = self.get_error_from_request(response) completion(None, error) else: - json_ = self.get_json(request.text) + json_ = self.get_json(response.text) completion(json_, None) elif method == 'POST': - raise NotImplementedError + re - def get_error_from_request(self, request): + def get_error_from_response(self, response): return { - 'error': self.get_json(request.text) + 'error': self.get_json(response.text) } def get_json(self, string): return json.loads(string) def get_api_url(self, path): return FACEBOOK_MESSAGES_POST_URL + path def get_headers(self): return {}
Change wording to use response
## Code Before: import requests, json from resources.urls import FACEBOOK_MESSAGES_POST_URL class HttpClient(): """ Client which excutes the call to facebook's messenger api """ def submit_request(self, path, method, payload, completion): assert len(path) > 0 path = self.get_api_url(path) headers = self.get_headers() if method == 'GET': request = requests.get(path, headers=headers) if request.status_code is not 200: error = self.get_error_from_request(request) completion(None, error) else: json_ = self.get_json(request.text) completion(json_, None) elif method == 'POST': raise NotImplementedError def get_error_from_request(self, request): return { 'error': self.get_json(request.text) } def get_json(self, string): return json.loads(string) def get_api_url(self, path): return FACEBOOK_MESSAGES_POST_URL + path def get_headers(self): return {} ## Instruction: Change wording to use response ## Code After: import requests, json from resources.urls import FACEBOOK_MESSAGES_POST_URL class HttpClient(): """ Client which excutes the call to facebook's messenger api """ def submit_request(self, path, method, payload, completion): assert len(path) > 0 path = self.get_api_url(path) headers = self.get_headers() if method == 'GET': response = requests.get(path, headers=headers) if response.status_code is not 200: error = self.get_error_from_request(response) completion(None, error) else: json_ = self.get_json(response.text) completion(json_, None) elif method == 'POST': re def get_error_from_response(self, response): return { 'error': self.get_json(response.text) } def get_json(self, string): return json.loads(string) def get_api_url(self, path): return FACEBOOK_MESSAGES_POST_URL + path def get_headers(self): return {}
... if method == 'GET': response = requests.get(path, headers=headers) if response.status_code is not 200: error = self.get_error_from_request(response) completion(None, error) ... else: json_ = self.get_json(response.text) completion(json_, None) ... elif method == 'POST': re def get_error_from_response(self, response): ... return { 'error': self.get_json(response.text) } ...
5786942c88420be913705790489676780dcd9fc0
nlppln/utils.py
nlppln/utils.py
"""NLP pipeline utility functionality""" import os def remove_ext(fname): """Removes the extension from a filename """ bn = os.path.basename(fname) return os.path.splitext(bn)[0] def create_dirs(fname): """Create (output) directories if they don't exist """ fname = os.path.dirname(fname) if not os.path.exists(fname): os.makedirs(fname) def out_file_name(out_dir, fname, ext): fname = remove_ext(fname) return os.path.join(out_dir, '{}.{}'.format(fname, ext))
"""NLP pipeline utility functionality""" import os def remove_ext(fname): """Removes the extension from a filename """ bn = os.path.basename(fname) return os.path.splitext(bn)[0] def create_dirs(fname): """Create (output) directories if they don't exist """ fname = os.path.dirname(os.path.abspath(fname)) if not os.path.exists(fname): os.makedirs(fname) def out_file_name(out_dir, fname, ext): fname = remove_ext(fname) return os.path.join(out_dir, '{}.{}'.format(fname, ext))
Update createdirs to determine the absolute path of files
Update createdirs to determine the absolute path of files
Python
apache-2.0
WhatWorksWhenForWhom/nlppln,WhatWorksWhenForWhom/nlppln,WhatWorksWhenForWhom/nlppln
"""NLP pipeline utility functionality""" import os def remove_ext(fname): """Removes the extension from a filename """ bn = os.path.basename(fname) return os.path.splitext(bn)[0] def create_dirs(fname): """Create (output) directories if they don't exist """ - fname = os.path.dirname(fname) + fname = os.path.dirname(os.path.abspath(fname)) if not os.path.exists(fname): os.makedirs(fname) def out_file_name(out_dir, fname, ext): fname = remove_ext(fname) return os.path.join(out_dir, '{}.{}'.format(fname, ext))
Update createdirs to determine the absolute path of files
## Code Before: """NLP pipeline utility functionality""" import os def remove_ext(fname): """Removes the extension from a filename """ bn = os.path.basename(fname) return os.path.splitext(bn)[0] def create_dirs(fname): """Create (output) directories if they don't exist """ fname = os.path.dirname(fname) if not os.path.exists(fname): os.makedirs(fname) def out_file_name(out_dir, fname, ext): fname = remove_ext(fname) return os.path.join(out_dir, '{}.{}'.format(fname, ext)) ## Instruction: Update createdirs to determine the absolute path of files ## Code After: """NLP pipeline utility functionality""" import os def remove_ext(fname): """Removes the extension from a filename """ bn = os.path.basename(fname) return os.path.splitext(bn)[0] def create_dirs(fname): """Create (output) directories if they don't exist """ fname = os.path.dirname(os.path.abspath(fname)) if not os.path.exists(fname): os.makedirs(fname) def out_file_name(out_dir, fname, ext): fname = remove_ext(fname) return os.path.join(out_dir, '{}.{}'.format(fname, ext))
// ... existing code ... """ fname = os.path.dirname(os.path.abspath(fname)) // ... rest of the code ...
29aa1a440a9ff225d3f9a4773f9097a5efcbd0de
test/integration/test_output.py
test/integration/test_output.py
from ..helpers import * def test_honcho_start_joins_stderr_into_stdout(): ret, out, err = get_honcho_output(['-f', 'Procfile.output', 'start']) assert_equal(ret, 0) assert_in('some normal output', out) assert_in('and then write to stderr', out) assert_equal(err, '') def test_honcho_run_keeps_stderr_and_stdout_separate(): ret, out, err = get_honcho_output(['run', 'python', 'output.py']) assert_equal(ret, 0) assert_equal(out, 'some normal output\n') assert_equal(err, 'and then write to stderr\n')
from ..helpers import * def test_honcho_start_joins_stderr_into_stdout(): ret, out, err = get_honcho_output(['-f', 'Procfile.output', 'start']) assert_equal(ret, 0) assert_regexp_matches(out, r'some normal output') assert_regexp_matches(out, r'and then write to stderr') assert_equal(err, '') def test_honcho_run_keeps_stderr_and_stdout_separate(): ret, out, err = get_honcho_output(['run', 'python', 'output.py']) assert_equal(ret, 0) assert_equal(out, 'some normal output\n') assert_equal(err, 'and then write to stderr\n')
Rewrite assertions for Python 2.6 compatibility
Rewrite assertions for Python 2.6 compatibility
Python
mit
janusnic/honcho,xarisd/honcho,myyk/honcho,gratipay/honcho,nickstenning/honcho,nickstenning/honcho,gratipay/honcho
from ..helpers import * def test_honcho_start_joins_stderr_into_stdout(): ret, out, err = get_honcho_output(['-f', 'Procfile.output', 'start']) assert_equal(ret, 0) - assert_in('some normal output', out) - assert_in('and then write to stderr', out) + assert_regexp_matches(out, r'some normal output') + assert_regexp_matches(out, r'and then write to stderr') assert_equal(err, '') def test_honcho_run_keeps_stderr_and_stdout_separate(): ret, out, err = get_honcho_output(['run', 'python', 'output.py']) assert_equal(ret, 0) assert_equal(out, 'some normal output\n') assert_equal(err, 'and then write to stderr\n')
Rewrite assertions for Python 2.6 compatibility
## Code Before: from ..helpers import * def test_honcho_start_joins_stderr_into_stdout(): ret, out, err = get_honcho_output(['-f', 'Procfile.output', 'start']) assert_equal(ret, 0) assert_in('some normal output', out) assert_in('and then write to stderr', out) assert_equal(err, '') def test_honcho_run_keeps_stderr_and_stdout_separate(): ret, out, err = get_honcho_output(['run', 'python', 'output.py']) assert_equal(ret, 0) assert_equal(out, 'some normal output\n') assert_equal(err, 'and then write to stderr\n') ## Instruction: Rewrite assertions for Python 2.6 compatibility ## Code After: from ..helpers import * def test_honcho_start_joins_stderr_into_stdout(): ret, out, err = get_honcho_output(['-f', 'Procfile.output', 'start']) assert_equal(ret, 0) assert_regexp_matches(out, r'some normal output') assert_regexp_matches(out, r'and then write to stderr') assert_equal(err, '') def test_honcho_run_keeps_stderr_and_stdout_separate(): ret, out, err = get_honcho_output(['run', 'python', 'output.py']) assert_equal(ret, 0) assert_equal(out, 'some normal output\n') assert_equal(err, 'and then write to stderr\n')
... assert_regexp_matches(out, r'some normal output') assert_regexp_matches(out, r'and then write to stderr') assert_equal(err, '') ...
ee9df63aeaabb4111cea3698a4f0e30b4502e519
test/disable_captcha.py
test/disable_captcha.py
import owebunit import urlparse from wolis_test_case import WolisTestCase class AcpLoginTestCase(WolisTestCase): def test_disable_captcha(self): self.login('morpheus', 'morpheus') self.acp_login('morpheus', 'morpheus') start_url = '/adm/index.php' self.get_with_sid(start_url) self.assert_status(200) assert 'Board statistics' in self.response.body url = self.link_href_by_text('Spambot countermeasures') # already has sid self.get(urlparse.urljoin(start_url, url)) self.assert_status(200) assert 'Enable spambot countermeasures' in self.response.body assert len(self.response.forms) == 1 form = self.response.forms[0] params = { 'enable_confirm': '0', } params = owebunit.extend_params(form.params.list, params) self.post(form.computed_action, body=params) self.assert_status(200) assert 'Configuration updated successfully' in self.response.body if __name__ == '__main__': import unittest unittest.main()
import owebunit import urlparse from wolis_test_case import WolisTestCase class AcpLoginTestCase(WolisTestCase): def test_disable_captcha(self): self.login('morpheus', 'morpheus') self.acp_login('morpheus', 'morpheus') self.change_acp_knob( link_text='Spambot countermeasures', check_page_text='Enable spambot countermeasures', name='enable_confirm', value='0', ) def change_acp_knob(self, link_text, check_page_text, name, value): start_url = '/adm/index.php' self.get_with_sid(start_url) self.assert_status(200) assert 'Board statistics' in self.response.body url = self.link_href_by_text(link_text) # already has sid self.get(urlparse.urljoin(start_url, url)) self.assert_status(200) assert check_page_text in self.response.body assert len(self.response.forms) == 1 form = self.response.forms[0] params = { name: value, } params = owebunit.extend_params(form.params.list, params) self.post(form.computed_action, body=params) self.assert_status(200) assert 'Configuration updated successfully' in self.response.body if __name__ == '__main__': import unittest unittest.main()
Generalize to allow editing other configuration
Generalize to allow editing other configuration
Python
bsd-2-clause
p/wolis-phpbb,p/wolis-phpbb
import owebunit import urlparse from wolis_test_case import WolisTestCase class AcpLoginTestCase(WolisTestCase): def test_disable_captcha(self): self.login('morpheus', 'morpheus') self.acp_login('morpheus', 'morpheus') + self.change_acp_knob( + link_text='Spambot countermeasures', + check_page_text='Enable spambot countermeasures', + name='enable_confirm', + value='0', + ) + + def change_acp_knob(self, link_text, check_page_text, name, value): start_url = '/adm/index.php' self.get_with_sid(start_url) self.assert_status(200) assert 'Board statistics' in self.response.body - url = self.link_href_by_text('Spambot countermeasures') + url = self.link_href_by_text(link_text) # already has sid self.get(urlparse.urljoin(start_url, url)) self.assert_status(200) - assert 'Enable spambot countermeasures' in self.response.body + assert check_page_text in self.response.body assert len(self.response.forms) == 1 form = self.response.forms[0] params = { - 'enable_confirm': '0', + name: value, } params = owebunit.extend_params(form.params.list, params) self.post(form.computed_action, body=params) self.assert_status(200) assert 'Configuration updated successfully' in self.response.body if __name__ == '__main__': import unittest unittest.main()
Generalize to allow editing other configuration
## Code Before: import owebunit import urlparse from wolis_test_case import WolisTestCase class AcpLoginTestCase(WolisTestCase): def test_disable_captcha(self): self.login('morpheus', 'morpheus') self.acp_login('morpheus', 'morpheus') start_url = '/adm/index.php' self.get_with_sid(start_url) self.assert_status(200) assert 'Board statistics' in self.response.body url = self.link_href_by_text('Spambot countermeasures') # already has sid self.get(urlparse.urljoin(start_url, url)) self.assert_status(200) assert 'Enable spambot countermeasures' in self.response.body assert len(self.response.forms) == 1 form = self.response.forms[0] params = { 'enable_confirm': '0', } params = owebunit.extend_params(form.params.list, params) self.post(form.computed_action, body=params) self.assert_status(200) assert 'Configuration updated successfully' in self.response.body if __name__ == '__main__': import unittest unittest.main() ## Instruction: Generalize to allow editing other configuration ## Code After: import owebunit import urlparse from wolis_test_case import WolisTestCase class AcpLoginTestCase(WolisTestCase): def test_disable_captcha(self): self.login('morpheus', 'morpheus') self.acp_login('morpheus', 'morpheus') self.change_acp_knob( link_text='Spambot countermeasures', check_page_text='Enable spambot countermeasures', name='enable_confirm', value='0', ) def change_acp_knob(self, link_text, check_page_text, name, value): start_url = '/adm/index.php' self.get_with_sid(start_url) self.assert_status(200) assert 'Board statistics' in self.response.body url = self.link_href_by_text(link_text) # already has sid self.get(urlparse.urljoin(start_url, url)) self.assert_status(200) assert check_page_text in self.response.body assert len(self.response.forms) == 1 form = self.response.forms[0] params = { name: value, } params = owebunit.extend_params(form.params.list, params) self.post(form.computed_action, body=params) self.assert_status(200) assert 'Configuration updated successfully' in self.response.body if __name__ == '__main__': import unittest unittest.main()
... self.change_acp_knob( link_text='Spambot countermeasures', check_page_text='Enable spambot countermeasures', name='enable_confirm', value='0', ) def change_acp_knob(self, link_text, check_page_text, name, value): start_url = '/adm/index.php' ... url = self.link_href_by_text(link_text) ... assert check_page_text in self.response.body ... params = { name: value, } ...
73e5fe29074f52e0b769fd2a6c40669040bef330
app/notify_client/invite_api_client.py
app/notify_client/invite_api_client.py
from notifications_python_client.base import BaseAPIClient from app.notify_client.models import User class InviteApiClient(BaseAPIClient): def __init__(self, base_url=None, client_id=None, secret=None): super(self.__class__, self).__init__(base_url=base_url or 'base_url', client_id=client_id or 'client_id', secret=secret or 'secret') def init_app(self, app): self.base_url = app.config['API_HOST_NAME'] self.client_id = app.config['ADMIN_CLIENT_USER_NAME'] self.secret = app.config['ADMIN_CLIENT_SECRET'] def create_invite(self, invite_from_id, service_id, email_address, permissions): data = { 'service': str(service_id), 'email_address': email_address, 'from_user': invite_from_id, 'permissions': permissions } resp = self.post(url='/service/{}/invite'.format(service_id), data=data) return resp['data'] def get_invites_for_service(self, service_id): endpoint = '/service/{}/invite'.format(service_id) resp = self.get(endpoint) return [User(data) for data in resp['data']] def cancel_invited_user(self, service_id, invited_user_id): data = {'status': 'cancelled'} resp = self.post(url='/service/{0}/invite/{1}'.format(service_id, invited_user_id), data=data) return resp['data']
from notifications_python_client.base import BaseAPIClient from app.notify_client.models import User class InviteApiClient(BaseAPIClient): def __init__(self, base_url=None, client_id=None, secret=None): super(self.__class__, self).__init__(base_url=base_url or 'base_url', client_id=client_id or 'client_id', secret=secret or 'secret') def init_app(self, app): self.base_url = app.config['API_HOST_NAME'] self.client_id = app.config['ADMIN_CLIENT_USER_NAME'] self.secret = app.config['ADMIN_CLIENT_SECRET'] def create_invite(self, invite_from_id, service_id, email_address, permissions): data = { 'service': str(service_id), 'email_address': email_address, 'from_user': invite_from_id, 'permissions': permissions } resp = self.post(url='/service/{}/invite'.format(service_id), data=data) return resp['data'] def get_invites_for_service(self, service_id): endpoint = '/service/{}/invite'.format(service_id) resp = self.get(endpoint) return [User(data) for data in resp['data']] def cancel_invited_user(self, service_id, invited_user_id): data = {'status': 'cancelled'} self.post(url='/service/{0}/invite/{1}'.format(service_id, invited_user_id), data=data)
Change cancel_invited_user client to not return anything.
Change cancel_invited_user client to not return anything.
Python
mit
alphagov/notifications-admin,gov-cjwaszczuk/notifications-admin,alphagov/notifications-admin,gov-cjwaszczuk/notifications-admin,gov-cjwaszczuk/notifications-admin,alphagov/notifications-admin,alphagov/notifications-admin,gov-cjwaszczuk/notifications-admin
from notifications_python_client.base import BaseAPIClient from app.notify_client.models import User class InviteApiClient(BaseAPIClient): def __init__(self, base_url=None, client_id=None, secret=None): super(self.__class__, self).__init__(base_url=base_url or 'base_url', client_id=client_id or 'client_id', secret=secret or 'secret') def init_app(self, app): self.base_url = app.config['API_HOST_NAME'] self.client_id = app.config['ADMIN_CLIENT_USER_NAME'] self.secret = app.config['ADMIN_CLIENT_SECRET'] def create_invite(self, invite_from_id, service_id, email_address, permissions): data = { 'service': str(service_id), 'email_address': email_address, 'from_user': invite_from_id, 'permissions': permissions } resp = self.post(url='/service/{}/invite'.format(service_id), data=data) return resp['data'] def get_invites_for_service(self, service_id): endpoint = '/service/{}/invite'.format(service_id) resp = self.get(endpoint) return [User(data) for data in resp['data']] def cancel_invited_user(self, service_id, invited_user_id): data = {'status': 'cancelled'} - resp = self.post(url='/service/{0}/invite/{1}'.format(service_id, invited_user_id), + self.post(url='/service/{0}/invite/{1}'.format(service_id, invited_user_id), - data=data) + data=data) - return resp['data']
Change cancel_invited_user client to not return anything.
## Code Before: from notifications_python_client.base import BaseAPIClient from app.notify_client.models import User class InviteApiClient(BaseAPIClient): def __init__(self, base_url=None, client_id=None, secret=None): super(self.__class__, self).__init__(base_url=base_url or 'base_url', client_id=client_id or 'client_id', secret=secret or 'secret') def init_app(self, app): self.base_url = app.config['API_HOST_NAME'] self.client_id = app.config['ADMIN_CLIENT_USER_NAME'] self.secret = app.config['ADMIN_CLIENT_SECRET'] def create_invite(self, invite_from_id, service_id, email_address, permissions): data = { 'service': str(service_id), 'email_address': email_address, 'from_user': invite_from_id, 'permissions': permissions } resp = self.post(url='/service/{}/invite'.format(service_id), data=data) return resp['data'] def get_invites_for_service(self, service_id): endpoint = '/service/{}/invite'.format(service_id) resp = self.get(endpoint) return [User(data) for data in resp['data']] def cancel_invited_user(self, service_id, invited_user_id): data = {'status': 'cancelled'} resp = self.post(url='/service/{0}/invite/{1}'.format(service_id, invited_user_id), data=data) return resp['data'] ## Instruction: Change cancel_invited_user client to not return anything. ## Code After: from notifications_python_client.base import BaseAPIClient from app.notify_client.models import User class InviteApiClient(BaseAPIClient): def __init__(self, base_url=None, client_id=None, secret=None): super(self.__class__, self).__init__(base_url=base_url or 'base_url', client_id=client_id or 'client_id', secret=secret or 'secret') def init_app(self, app): self.base_url = app.config['API_HOST_NAME'] self.client_id = app.config['ADMIN_CLIENT_USER_NAME'] self.secret = app.config['ADMIN_CLIENT_SECRET'] def create_invite(self, invite_from_id, service_id, email_address, permissions): data = { 'service': str(service_id), 'email_address': email_address, 'from_user': invite_from_id, 'permissions': permissions } resp = self.post(url='/service/{}/invite'.format(service_id), data=data) return resp['data'] def get_invites_for_service(self, service_id): endpoint = '/service/{}/invite'.format(service_id) resp = self.get(endpoint) return [User(data) for data in resp['data']] def cancel_invited_user(self, service_id, invited_user_id): data = {'status': 'cancelled'} self.post(url='/service/{0}/invite/{1}'.format(service_id, invited_user_id), data=data)
... data = {'status': 'cancelled'} self.post(url='/service/{0}/invite/{1}'.format(service_id, invited_user_id), data=data) ...
60625877a23e26e66c2c97cbeb4f139ede717eda
B.py
B.py
from collections import namedtuple import matplotlib.pyplot as plt BCand = namedtuple('BCand', ['m', 'merr', 'pt', 'p']) bs = [] with open('B.txt') as f: for line in f.readlines()[1:]: bs.append(BCand(*[float(v) for v in line.strip().split(',')])) masses = [b.m for b in bs] plt.hist(masses, 60, histtype='stepfilled') plt.xlabel(r'$m_B / \mathrm{GeV}$') plt.savefig('mass.pdf')
from collections import namedtuple import matplotlib.pyplot as plt import numpy as np BCand = namedtuple('BCand', ['m', 'merr', 'pt', 'p']) bs = [BCand(*b) for b in np.genfromtxt('B.txt', skip_header=1, delimiter=',')] masses = [b.m for b in bs] ns, bins, _ = plt.hist(masses, 60, histtype='stepfilled', facecolor='r', edgecolor='none') centers = bins[:-1] + (bins[1:] - bins[:-1]) / 2 merr = np.sqrt(ns) plt.errorbar(centers, ns, yerr=merr, fmt='b+') plt.xlabel(r'$m_B / \mathrm{GeV}$') plt.savefig('mass.pdf')
Use numpy for readin and add errorbars.
Use numpy for readin and add errorbars.
Python
mit
bixel/python-introduction
from collections import namedtuple import matplotlib.pyplot as plt + import numpy as np BCand = namedtuple('BCand', ['m', 'merr', 'pt', 'p']) + bs = [BCand(*b) for b in np.genfromtxt('B.txt', skip_header=1, delimiter=',')] - bs = [] - - with open('B.txt') as f: - for line in f.readlines()[1:]: - bs.append(BCand(*[float(v) for v in line.strip().split(',')])) masses = [b.m for b in bs] - plt.hist(masses, 60, histtype='stepfilled') + ns, bins, _ = plt.hist(masses, 60, histtype='stepfilled', facecolor='r', + edgecolor='none') + centers = bins[:-1] + (bins[1:] - bins[:-1]) / 2 + merr = np.sqrt(ns) + plt.errorbar(centers, ns, yerr=merr, fmt='b+') plt.xlabel(r'$m_B / \mathrm{GeV}$') plt.savefig('mass.pdf')
Use numpy for readin and add errorbars.
## Code Before: from collections import namedtuple import matplotlib.pyplot as plt BCand = namedtuple('BCand', ['m', 'merr', 'pt', 'p']) bs = [] with open('B.txt') as f: for line in f.readlines()[1:]: bs.append(BCand(*[float(v) for v in line.strip().split(',')])) masses = [b.m for b in bs] plt.hist(masses, 60, histtype='stepfilled') plt.xlabel(r'$m_B / \mathrm{GeV}$') plt.savefig('mass.pdf') ## Instruction: Use numpy for readin and add errorbars. ## Code After: from collections import namedtuple import matplotlib.pyplot as plt import numpy as np BCand = namedtuple('BCand', ['m', 'merr', 'pt', 'p']) bs = [BCand(*b) for b in np.genfromtxt('B.txt', skip_header=1, delimiter=',')] masses = [b.m for b in bs] ns, bins, _ = plt.hist(masses, 60, histtype='stepfilled', facecolor='r', edgecolor='none') centers = bins[:-1] + (bins[1:] - bins[:-1]) / 2 merr = np.sqrt(ns) plt.errorbar(centers, ns, yerr=merr, fmt='b+') plt.xlabel(r'$m_B / \mathrm{GeV}$') plt.savefig('mass.pdf')
... import matplotlib.pyplot as plt import numpy as np ... bs = [BCand(*b) for b in np.genfromtxt('B.txt', skip_header=1, delimiter=',')] ... ns, bins, _ = plt.hist(masses, 60, histtype='stepfilled', facecolor='r', edgecolor='none') centers = bins[:-1] + (bins[1:] - bins[:-1]) / 2 merr = np.sqrt(ns) plt.errorbar(centers, ns, yerr=merr, fmt='b+') plt.xlabel(r'$m_B / \mathrm{GeV}$') ...
f814e945d3e62c87c5f86ef5ac37c5feb733b83d
tests/test_ext.py
tests/test_ext.py
from __future__ import absolute_import, unicode_literals import unittest from mopidy import config, ext class ExtensionTest(unittest.TestCase): def setUp(self): # noqa: N802 self.ext = ext.Extension() def test_dist_name_is_none(self): self.assertIsNone(self.ext.dist_name) def test_ext_name_is_none(self): self.assertIsNone(self.ext.ext_name) def test_version_is_none(self): self.assertIsNone(self.ext.version) def test_get_default_config_raises_not_implemented(self): with self.assertRaises(NotImplementedError): self.ext.get_default_config() def test_get_config_schema_returns_extension_schema(self): schema = self.ext.get_config_schema() self.assertIsInstance(schema['enabled'], config.Boolean) def test_validate_environment_does_nothing_by_default(self): self.assertIsNone(self.ext.validate_environment()) def test_setup_raises_not_implemented(self): with self.assertRaises(NotImplementedError): self.ext.setup(None)
from __future__ import absolute_import, unicode_literals import pytest from mopidy import config, ext @pytest.fixture def extension(): return ext.Extension() def test_dist_name_is_none(extension): assert extension.dist_name is None def test_ext_name_is_none(extension): assert extension.ext_name is None def test_version_is_none(extension): assert extension.version is None def test_get_default_config_raises_not_implemented(extension): with pytest.raises(NotImplementedError): extension.get_default_config() def test_get_config_schema_returns_extension_schema(extension): schema = extension.get_config_schema() assert isinstance(schema['enabled'], config.Boolean) def test_validate_environment_does_nothing_by_default(extension): assert extension.validate_environment() is None def test_setup_raises_not_implemented(extension): with pytest.raises(NotImplementedError): extension.setup(None)
Convert ext test to pytests
tests: Convert ext test to pytests
Python
apache-2.0
mokieyue/mopidy,bencevans/mopidy,ZenithDK/mopidy,jodal/mopidy,quartz55/mopidy,pacificIT/mopidy,pacificIT/mopidy,quartz55/mopidy,ali/mopidy,swak/mopidy,tkem/mopidy,bencevans/mopidy,mopidy/mopidy,SuperStarPL/mopidy,dbrgn/mopidy,hkariti/mopidy,glogiotatidis/mopidy,mokieyue/mopidy,ali/mopidy,bacontext/mopidy,glogiotatidis/mopidy,diandiankan/mopidy,dbrgn/mopidy,mopidy/mopidy,jmarsik/mopidy,glogiotatidis/mopidy,bacontext/mopidy,vrs01/mopidy,kingosticks/mopidy,SuperStarPL/mopidy,diandiankan/mopidy,vrs01/mopidy,dbrgn/mopidy,jcass77/mopidy,glogiotatidis/mopidy,swak/mopidy,bencevans/mopidy,mokieyue/mopidy,adamcik/mopidy,tkem/mopidy,rawdlite/mopidy,bacontext/mopidy,pacificIT/mopidy,ali/mopidy,jcass77/mopidy,rawdlite/mopidy,mopidy/mopidy,vrs01/mopidy,tkem/mopidy,dbrgn/mopidy,SuperStarPL/mopidy,diandiankan/mopidy,tkem/mopidy,jodal/mopidy,diandiankan/mopidy,jmarsik/mopidy,quartz55/mopidy,ZenithDK/mopidy,jodal/mopidy,ZenithDK/mopidy,jmarsik/mopidy,swak/mopidy,ZenithDK/mopidy,bacontext/mopidy,kingosticks/mopidy,quartz55/mopidy,pacificIT/mopidy,SuperStarPL/mopidy,hkariti/mopidy,ali/mopidy,hkariti/mopidy,adamcik/mopidy,vrs01/mopidy,rawdlite/mopidy,hkariti/mopidy,rawdlite/mopidy,mokieyue/mopidy,adamcik/mopidy,swak/mopidy,kingosticks/mopidy,bencevans/mopidy,jmarsik/mopidy,jcass77/mopidy
from __future__ import absolute_import, unicode_literals - import unittest + import pytest from mopidy import config, ext - class ExtensionTest(unittest.TestCase): + @pytest.fixture + def extension(): + return ext.Extension() - def setUp(self): # noqa: N802 - self.ext = ext.Extension() - def test_dist_name_is_none(self): + def test_dist_name_is_none(extension): - self.assertIsNone(self.ext.dist_name) + assert extension.dist_name is None - def test_ext_name_is_none(self): - self.assertIsNone(self.ext.ext_name) - def test_version_is_none(self): - self.assertIsNone(self.ext.version) + def test_ext_name_is_none(extension): + assert extension.ext_name is None - def test_get_default_config_raises_not_implemented(self): - with self.assertRaises(NotImplementedError): - self.ext.get_default_config() + def test_version_is_none(extension): + assert extension.version is None - def test_get_config_schema_returns_extension_schema(self): - schema = self.ext.get_config_schema() - self.assertIsInstance(schema['enabled'], config.Boolean) - def test_validate_environment_does_nothing_by_default(self): - self.assertIsNone(self.ext.validate_environment()) - def test_setup_raises_not_implemented(self): + def test_get_default_config_raises_not_implemented(extension): - with self.assertRaises(NotImplementedError): + with pytest.raises(NotImplementedError): - self.ext.setup(None) + extension.get_default_config() + + def test_get_config_schema_returns_extension_schema(extension): + schema = extension.get_config_schema() + assert isinstance(schema['enabled'], config.Boolean) + + + def test_validate_environment_does_nothing_by_default(extension): + assert extension.validate_environment() is None + + + def test_setup_raises_not_implemented(extension): + with pytest.raises(NotImplementedError): + extension.setup(None) +
Convert ext test to pytests
## Code Before: from __future__ import absolute_import, unicode_literals import unittest from mopidy import config, ext class ExtensionTest(unittest.TestCase): def setUp(self): # noqa: N802 self.ext = ext.Extension() def test_dist_name_is_none(self): self.assertIsNone(self.ext.dist_name) def test_ext_name_is_none(self): self.assertIsNone(self.ext.ext_name) def test_version_is_none(self): self.assertIsNone(self.ext.version) def test_get_default_config_raises_not_implemented(self): with self.assertRaises(NotImplementedError): self.ext.get_default_config() def test_get_config_schema_returns_extension_schema(self): schema = self.ext.get_config_schema() self.assertIsInstance(schema['enabled'], config.Boolean) def test_validate_environment_does_nothing_by_default(self): self.assertIsNone(self.ext.validate_environment()) def test_setup_raises_not_implemented(self): with self.assertRaises(NotImplementedError): self.ext.setup(None) ## Instruction: Convert ext test to pytests ## Code After: from __future__ import absolute_import, unicode_literals import pytest from mopidy import config, ext @pytest.fixture def extension(): return ext.Extension() def test_dist_name_is_none(extension): assert extension.dist_name is None def test_ext_name_is_none(extension): assert extension.ext_name is None def test_version_is_none(extension): assert extension.version is None def test_get_default_config_raises_not_implemented(extension): with pytest.raises(NotImplementedError): extension.get_default_config() def test_get_config_schema_returns_extension_schema(extension): schema = extension.get_config_schema() assert isinstance(schema['enabled'], config.Boolean) def test_validate_environment_does_nothing_by_default(extension): assert extension.validate_environment() is None def test_setup_raises_not_implemented(extension): with pytest.raises(NotImplementedError): extension.setup(None)
// ... existing code ... import pytest // ... modified code ... @pytest.fixture def extension(): return ext.Extension() def test_dist_name_is_none(extension): assert extension.dist_name is None def test_ext_name_is_none(extension): assert extension.ext_name is None def test_version_is_none(extension): assert extension.version is None def test_get_default_config_raises_not_implemented(extension): with pytest.raises(NotImplementedError): extension.get_default_config() def test_get_config_schema_returns_extension_schema(extension): schema = extension.get_config_schema() assert isinstance(schema['enabled'], config.Boolean) def test_validate_environment_does_nothing_by_default(extension): assert extension.validate_environment() is None def test_setup_raises_not_implemented(extension): with pytest.raises(NotImplementedError): extension.setup(None) // ... rest of the code ...
8c551fe51ed142305945c0cef530ac84ed3e7eb9
nodeconductor/logging/perms.py
nodeconductor/logging/perms.py
from nodeconductor.core.permissions import StaffPermissionLogic PERMISSION_LOGICS = ( ('logging.Alert', StaffPermissionLogic(any_permission=True)), ('logging.SystemNotification', StaffPermissionLogic(any_permission=True)), )
from nodeconductor.core.permissions import StaffPermissionLogic PERMISSION_LOGICS = ( ('logging.Alert', StaffPermissionLogic(any_permission=True)), ('logging.WebHook', StaffPermissionLogic(any_permission=True)), ('logging.PushHook', StaffPermissionLogic(any_permission=True)), ('logging.EmailHook', StaffPermissionLogic(any_permission=True)), ('logging.SystemNotification', StaffPermissionLogic(any_permission=True)), )
Allow staff user to manage hooks.
Allow staff user to manage hooks.
Python
mit
opennode/nodeconductor,opennode/nodeconductor,opennode/nodeconductor
from nodeconductor.core.permissions import StaffPermissionLogic PERMISSION_LOGICS = ( ('logging.Alert', StaffPermissionLogic(any_permission=True)), + ('logging.WebHook', StaffPermissionLogic(any_permission=True)), + ('logging.PushHook', StaffPermissionLogic(any_permission=True)), + ('logging.EmailHook', StaffPermissionLogic(any_permission=True)), ('logging.SystemNotification', StaffPermissionLogic(any_permission=True)), )
Allow staff user to manage hooks.
## Code Before: from nodeconductor.core.permissions import StaffPermissionLogic PERMISSION_LOGICS = ( ('logging.Alert', StaffPermissionLogic(any_permission=True)), ('logging.SystemNotification', StaffPermissionLogic(any_permission=True)), ) ## Instruction: Allow staff user to manage hooks. ## Code After: from nodeconductor.core.permissions import StaffPermissionLogic PERMISSION_LOGICS = ( ('logging.Alert', StaffPermissionLogic(any_permission=True)), ('logging.WebHook', StaffPermissionLogic(any_permission=True)), ('logging.PushHook', StaffPermissionLogic(any_permission=True)), ('logging.EmailHook', StaffPermissionLogic(any_permission=True)), ('logging.SystemNotification', StaffPermissionLogic(any_permission=True)), )
// ... existing code ... ('logging.Alert', StaffPermissionLogic(any_permission=True)), ('logging.WebHook', StaffPermissionLogic(any_permission=True)), ('logging.PushHook', StaffPermissionLogic(any_permission=True)), ('logging.EmailHook', StaffPermissionLogic(any_permission=True)), ('logging.SystemNotification', StaffPermissionLogic(any_permission=True)), // ... rest of the code ...
9c16b71ecbb38115f107c7baba56304fb9630ec5
ocds/export/__init__.py
ocds/export/__init__.py
from .models import ( Release, ReleasePackage, Record, RecordPackage ) from .schema import Tender from .helpers import ( mode_test, get_ocid ) def release_tender(tender, prefix): """ returns Release object created from `tender` with ocid `prefix` """ date = tender.get('dateModified', '') ocid = get_ocid(prefix, tender['tenderID']) return Release(dict(tender=tender, ocid=ocid, date=date)) def release_tenders(tenders, prefix): """ returns list of Release object created from `tenders` with amendment info and ocid `prefix` """ prev_tender = next(tenders) for tender in tenders: yield Tender.with_diff(prev_tender, tender) prev_tender = tender def package_tenders(tenders, params): data = {} for field in ReleasePackage._fields: if field in params: data[field] = params.get(field, '') data['releases'] = [release_tender(tender, params.get('prefix')) for tender in tenders] return ReleasePackage(dict(**data)).serialize()
from .models import ( Release, ReleasePackage, Record, RecordPackage ) from .schema import Tender from .helpers import ( mode_test, get_ocid ) def release_tender(tender, prefix): """ returns Release object created from `tender` with ocid `prefix` """ date = tender.get('dateModified', '') ocid = get_ocid(prefix, tender['tenderID']) return Release(dict(tender=tender, ocid=ocid, date=date)) def release_tenders(tenders, prefix): """ returns list of Release object created from `tenders` with amendment info and ocid `prefix` """ prev_tender = next(tenders) for tender in tenders: data = {} for field in ['tender', 'awards', 'contracts']: model = getattr(Release, field).model_class if field in tender: data[field] = model.fromDiff(prev_tender.get(field, ''), new_tender.get(field, '')) elif field == 'tender': data['tender'] = model.fromDiff(prev_tender, tender) yield Release(data) prev_tender = tender def package_tenders(tenders, params): data = {} for field in ReleasePackage._fields: if field in params: data[field] = params.get(field, '') data['releases'] = [release_tender(tender, params.get('prefix')) for tender in tenders] return ReleasePackage(dict(**data)).serialize()
Update helpers for generating releases
Update helpers for generating releases
Python
apache-2.0
yshalenyk/openprocurement.ocds.export,yshalenyk/openprocurement.ocds.export,yshalenyk/ocds.export
from .models import ( Release, ReleasePackage, Record, RecordPackage ) from .schema import Tender from .helpers import ( mode_test, get_ocid ) def release_tender(tender, prefix): """ returns Release object created from `tender` with ocid `prefix` """ date = tender.get('dateModified', '') ocid = get_ocid(prefix, tender['tenderID']) return Release(dict(tender=tender, ocid=ocid, date=date)) def release_tenders(tenders, prefix): """ returns list of Release object created from `tenders` with amendment info and ocid `prefix` """ prev_tender = next(tenders) for tender in tenders: - yield Tender.with_diff(prev_tender, tender) + data = {} + for field in ['tender', 'awards', 'contracts']: + model = getattr(Release, field).model_class + if field in tender: + data[field] = model.fromDiff(prev_tender.get(field, ''), new_tender.get(field, '')) + elif field == 'tender': + data['tender'] = model.fromDiff(prev_tender, tender) + yield Release(data) prev_tender = tender + def package_tenders(tenders, params): data = {} for field in ReleasePackage._fields: if field in params: data[field] = params.get(field, '') data['releases'] = [release_tender(tender, params.get('prefix')) for tender in tenders] return ReleasePackage(dict(**data)).serialize()
Update helpers for generating releases
## Code Before: from .models import ( Release, ReleasePackage, Record, RecordPackage ) from .schema import Tender from .helpers import ( mode_test, get_ocid ) def release_tender(tender, prefix): """ returns Release object created from `tender` with ocid `prefix` """ date = tender.get('dateModified', '') ocid = get_ocid(prefix, tender['tenderID']) return Release(dict(tender=tender, ocid=ocid, date=date)) def release_tenders(tenders, prefix): """ returns list of Release object created from `tenders` with amendment info and ocid `prefix` """ prev_tender = next(tenders) for tender in tenders: yield Tender.with_diff(prev_tender, tender) prev_tender = tender def package_tenders(tenders, params): data = {} for field in ReleasePackage._fields: if field in params: data[field] = params.get(field, '') data['releases'] = [release_tender(tender, params.get('prefix')) for tender in tenders] return ReleasePackage(dict(**data)).serialize() ## Instruction: Update helpers for generating releases ## Code After: from .models import ( Release, ReleasePackage, Record, RecordPackage ) from .schema import Tender from .helpers import ( mode_test, get_ocid ) def release_tender(tender, prefix): """ returns Release object created from `tender` with ocid `prefix` """ date = tender.get('dateModified', '') ocid = get_ocid(prefix, tender['tenderID']) return Release(dict(tender=tender, ocid=ocid, date=date)) def release_tenders(tenders, prefix): """ returns list of Release object created from `tenders` with amendment info and ocid `prefix` """ prev_tender = next(tenders) for tender in tenders: data = {} for field in ['tender', 'awards', 'contracts']: model = getattr(Release, field).model_class if field in tender: data[field] = model.fromDiff(prev_tender.get(field, ''), new_tender.get(field, '')) elif field == 'tender': data['tender'] = model.fromDiff(prev_tender, tender) yield Release(data) prev_tender = tender def package_tenders(tenders, params): data = {} for field in ReleasePackage._fields: if field in params: data[field] = params.get(field, '') data['releases'] = [release_tender(tender, params.get('prefix')) for tender in tenders] return ReleasePackage(dict(**data)).serialize()
# ... existing code ... for tender in tenders: data = {} for field in ['tender', 'awards', 'contracts']: model = getattr(Release, field).model_class if field in tender: data[field] = model.fromDiff(prev_tender.get(field, ''), new_tender.get(field, '')) elif field == 'tender': data['tender'] = model.fromDiff(prev_tender, tender) yield Release(data) prev_tender = tender # ... rest of the code ...
a7328bd229070126ca5b09bb1c9fe4c5e319bb04
members/urls.py
members/urls.py
from django.conf.urls import patterns, url from django.contrib import auth urlpatterns = patterns('members.views', url(r'^login/$', 'login', name='login'), url(r'^logout/$', 'logout', name='logout'), url(r'^search/(?P<name>.*)/$', 'search', name='search'), url(r'^archive/$', 'archive_student_council', name='archive_student_council'), )
from django.conf.urls import patterns, url from django.contrib import auth urlpatterns = patterns('members.views', url(r'^login/$', 'login', name='login'), url(r'^logout/$', 'logout', name='logout'), url(r'^search/(?P<name>.*)/$', 'search', name='search'), url(r'^archive/$', 'archive_student_council', name='archive_student_council'), url(r'^profile/$', 'user_projects', name='user-projects'), )
Add url for user's profile
Add url for user's profile
Python
mit
Hackfmi/Diaphanum,Hackfmi/Diaphanum
from django.conf.urls import patterns, url from django.contrib import auth urlpatterns = patterns('members.views', url(r'^login/$', 'login', name='login'), url(r'^logout/$', 'logout', name='logout'), url(r'^search/(?P<name>.*)/$', 'search', name='search'), url(r'^archive/$', 'archive_student_council', name='archive_student_council'), + url(r'^profile/$', 'user_projects', name='user-projects'), )
Add url for user's profile
## Code Before: from django.conf.urls import patterns, url from django.contrib import auth urlpatterns = patterns('members.views', url(r'^login/$', 'login', name='login'), url(r'^logout/$', 'logout', name='logout'), url(r'^search/(?P<name>.*)/$', 'search', name='search'), url(r'^archive/$', 'archive_student_council', name='archive_student_council'), ) ## Instruction: Add url for user's profile ## Code After: from django.conf.urls import patterns, url from django.contrib import auth urlpatterns = patterns('members.views', url(r'^login/$', 'login', name='login'), url(r'^logout/$', 'logout', name='logout'), url(r'^search/(?P<name>.*)/$', 'search', name='search'), url(r'^archive/$', 'archive_student_council', name='archive_student_council'), url(r'^profile/$', 'user_projects', name='user-projects'), )
# ... existing code ... url(r'^archive/$', 'archive_student_council', name='archive_student_council'), url(r'^profile/$', 'user_projects', name='user-projects'), ) # ... rest of the code ...
019d33092226d1ff8fe36897c03d25ddd48e34b1
serve.py
serve.py
import datetime as dt import sys import flask import sqlalchemy as sa import coils import tables import mapping app = flask.Flask(__name__) # Load configuration file. CONFIG = sys.argv[1] if len(sys.argv)>=2 else 'wabbit.cfg' config = coils.Config(CONFIG) @app.route('/') def index(): """Render the index page.""" return flask.render_template('index.html') @app.route('/info') def info(): """Return JSON of server info.""" # Connect to database engine. engine = sa.create_engine( 'mysql://{}:{}@{}/{}'.format( config['username'], config['password'], config['host'], config['db_name'])) Session = sa.orm.sessionmaker(bind=engine) session = Session() now = dt.datetime.now() datum = session.query(mapping.Datum).\ filter(mapping.Datum.name=='size')[0] return flask.jsonify(server_time=now, db_size=datum.value) if __name__ == '__main__': app.run()
import datetime as dt import sys import flask from flask.ext.sqlalchemy import SQLAlchemy import coils import mapping # Load configuration file. CONFIG = sys.argv[1] if len(sys.argv)>=2 else 'wabbit.cfg' config = coils.Config(CONFIG) # Initialize Flask and SQLAlchemy. app = flask.Flask(__name__) app.config['SQLALCHEMY_DATABASE_URI'] = 'mysql://{}:{}@{}/{}'.format( config['username'], config['password'], config['host'], config['db_name']) db = SQLAlchemy(app) @app.route('/') def index(): """Render the index page.""" return flask.render_template('index.html') @app.route('/info') def info(): """Return JSON of server info.""" now = dt.datetime.now() datum = db.session.query(mapping.Datum).\ filter(mapping.Datum.name=='size')[0] return flask.jsonify(server_time=now, db_size=datum.value) if __name__ == '__main__': app.run()
Use SQLAlchemy extension in Flask app.
Use SQLAlchemy extension in Flask app.
Python
mit
vmlaker/wabbit,vmlaker/wabbit,vmlaker/wabbit,vmlaker/wabbit
import datetime as dt import sys import flask - import sqlalchemy as sa + from flask.ext.sqlalchemy import SQLAlchemy import coils - import tables import mapping - - app = flask.Flask(__name__) # Load configuration file. CONFIG = sys.argv[1] if len(sys.argv)>=2 else 'wabbit.cfg' config = coils.Config(CONFIG) + + # Initialize Flask and SQLAlchemy. + app = flask.Flask(__name__) + app.config['SQLALCHEMY_DATABASE_URI'] = 'mysql://{}:{}@{}/{}'.format( + config['username'], config['password'], + config['host'], config['db_name']) + db = SQLAlchemy(app) @app.route('/') def index(): """Render the index page.""" return flask.render_template('index.html') @app.route('/info') def info(): """Return JSON of server info.""" - # Connect to database engine. - engine = sa.create_engine( - 'mysql://{}:{}@{}/{}'.format( - config['username'], config['password'], - config['host'], config['db_name'])) - Session = sa.orm.sessionmaker(bind=engine) - session = Session() now = dt.datetime.now() - datum = session.query(mapping.Datum).\ + datum = db.session.query(mapping.Datum).\ filter(mapping.Datum.name=='size')[0] return flask.jsonify(server_time=now, db_size=datum.value) if __name__ == '__main__': app.run()
Use SQLAlchemy extension in Flask app.
## Code Before: import datetime as dt import sys import flask import sqlalchemy as sa import coils import tables import mapping app = flask.Flask(__name__) # Load configuration file. CONFIG = sys.argv[1] if len(sys.argv)>=2 else 'wabbit.cfg' config = coils.Config(CONFIG) @app.route('/') def index(): """Render the index page.""" return flask.render_template('index.html') @app.route('/info') def info(): """Return JSON of server info.""" # Connect to database engine. engine = sa.create_engine( 'mysql://{}:{}@{}/{}'.format( config['username'], config['password'], config['host'], config['db_name'])) Session = sa.orm.sessionmaker(bind=engine) session = Session() now = dt.datetime.now() datum = session.query(mapping.Datum).\ filter(mapping.Datum.name=='size')[0] return flask.jsonify(server_time=now, db_size=datum.value) if __name__ == '__main__': app.run() ## Instruction: Use SQLAlchemy extension in Flask app. ## Code After: import datetime as dt import sys import flask from flask.ext.sqlalchemy import SQLAlchemy import coils import mapping # Load configuration file. CONFIG = sys.argv[1] if len(sys.argv)>=2 else 'wabbit.cfg' config = coils.Config(CONFIG) # Initialize Flask and SQLAlchemy. app = flask.Flask(__name__) app.config['SQLALCHEMY_DATABASE_URI'] = 'mysql://{}:{}@{}/{}'.format( config['username'], config['password'], config['host'], config['db_name']) db = SQLAlchemy(app) @app.route('/') def index(): """Render the index page.""" return flask.render_template('index.html') @app.route('/info') def info(): """Return JSON of server info.""" now = dt.datetime.now() datum = db.session.query(mapping.Datum).\ filter(mapping.Datum.name=='size')[0] return flask.jsonify(server_time=now, db_size=datum.value) if __name__ == '__main__': app.run()
... import flask from flask.ext.sqlalchemy import SQLAlchemy import coils import mapping ... config = coils.Config(CONFIG) # Initialize Flask and SQLAlchemy. app = flask.Flask(__name__) app.config['SQLALCHEMY_DATABASE_URI'] = 'mysql://{}:{}@{}/{}'.format( config['username'], config['password'], config['host'], config['db_name']) db = SQLAlchemy(app) ... """Return JSON of server info.""" now = dt.datetime.now() datum = db.session.query(mapping.Datum).\ filter(mapping.Datum.name=='size')[0] ...
8ed94e1fb93252eed47239d8c6a5f28796802a36
src/cclib/__init__.py
src/cclib/__init__.py
__version__ = "1.3" from . import parser from . import progress from . import method from . import bridge # The test module can be imported if it was installed with cclib. try: from . import test except ImportError: pass
# This file is part of cclib (http://cclib.sf.net), a library for parsing # and interpreting the results of computational chemistry packages. # # Copyright (C) 2006-2014 the cclib development team # # The library is free software, distributed under the terms of # the GNU Lesser General Public version 2.1 or later. You should have # received a copy of the license along with cclib. You can also access # the full license online at http://www.gnu.org/copyleft/lgpl.html. __version__ = "1.3" from . import parser from . import progress from . import method from . import bridge # The test module can be imported if it was installed with cclib. try: from . import test except ImportError: pass
Add a descriptive docstring to main cclib module
Add a descriptive docstring to main cclib module
Python
bsd-3-clause
berquist/cclib,jchodera/cclib,ghutchis/cclib,ben-albrecht/cclib,andersx/cclib,gaursagar/cclib,Clyde-fare/cclib,ghutchis/cclib,langner/cclib,andersx/cclib,cclib/cclib,Schamnad/cclib,ATenderholt/cclib,berquist/cclib,cclib/cclib,ATenderholt/cclib,langner/cclib,berquist/cclib,cclib/cclib,langner/cclib,gaursagar/cclib,jchodera/cclib,ben-albrecht/cclib,Schamnad/cclib,Clyde-fare/cclib
+ + # This file is part of cclib (http://cclib.sf.net), a library for parsing + # and interpreting the results of computational chemistry packages. + # + # Copyright (C) 2006-2014 the cclib development team + # + # The library is free software, distributed under the terms of + # the GNU Lesser General Public version 2.1 or later. You should have + # received a copy of the license along with cclib. You can also access + # the full license online at http://www.gnu.org/copyleft/lgpl.html. __version__ = "1.3" from . import parser from . import progress from . import method from . import bridge # The test module can be imported if it was installed with cclib. try: from . import test except ImportError: pass
Add a descriptive docstring to main cclib module
## Code Before: __version__ = "1.3" from . import parser from . import progress from . import method from . import bridge # The test module can be imported if it was installed with cclib. try: from . import test except ImportError: pass ## Instruction: Add a descriptive docstring to main cclib module ## Code After: # This file is part of cclib (http://cclib.sf.net), a library for parsing # and interpreting the results of computational chemistry packages. # # Copyright (C) 2006-2014 the cclib development team # # The library is free software, distributed under the terms of # the GNU Lesser General Public version 2.1 or later. You should have # received a copy of the license along with cclib. You can also access # the full license online at http://www.gnu.org/copyleft/lgpl.html. __version__ = "1.3" from . import parser from . import progress from . import method from . import bridge # The test module can be imported if it was installed with cclib. try: from . import test except ImportError: pass
# ... existing code ... # This file is part of cclib (http://cclib.sf.net), a library for parsing # and interpreting the results of computational chemistry packages. # # Copyright (C) 2006-2014 the cclib development team # # The library is free software, distributed under the terms of # the GNU Lesser General Public version 2.1 or later. You should have # received a copy of the license along with cclib. You can also access # the full license online at http://www.gnu.org/copyleft/lgpl.html. # ... rest of the code ...
838012c457d6c963707bb16259cd72d28c231672
cellcounter/accounts/decorators.py
cellcounter/accounts/decorators.py
__author__ = 'jvc26'
from functools import wraps from ratelimit.exceptions import Ratelimited from ratelimit.helpers import is_ratelimited def registration_ratelimit(ip=True, block=False, method=['POST'], field=None, rate='1/h', skip_if=None, keys=None): def decorator(fn): @wraps(fn) def _wrapped(request, *args, **kw): request.limited = getattr(request, 'limited', False) if skip_if is None or not skip_if(request): ratelimited = is_ratelimited(request=request, increment=False, ip=ip, method=method, field=field, rate=rate, keys=keys) if ratelimited and block: raise Ratelimited() return_val, success = fn(request, *args, **kw) if success: is_ratelimited(request=request, increment=True, ip=ip, method=method, field=field, rate=rate, keys=keys) return return_val return _wrapped return decorator
Use custom decorator to allow ratelimiting only on successful POST - prevents blocking form errors
Use custom decorator to allow ratelimiting only on successful POST - prevents blocking form errors
Python
mit
haematologic/cellcounter,cellcounter/cellcounter,cellcounter/cellcounter,haematologic/cellcounter,cellcounter/cellcounter,cellcounter/cellcounter,haematologic/cellcounter
- __author__ = 'jvc26' + from functools import wraps + from ratelimit.exceptions import Ratelimited + from ratelimit.helpers import is_ratelimited + + + def registration_ratelimit(ip=True, block=False, method=['POST'], field=None, rate='1/h', + skip_if=None, keys=None): + def decorator(fn): + @wraps(fn) + def _wrapped(request, *args, **kw): + request.limited = getattr(request, 'limited', False) + if skip_if is None or not skip_if(request): + ratelimited = is_ratelimited(request=request, increment=False, + ip=ip, method=method, field=field, + rate=rate, keys=keys) + if ratelimited and block: + raise Ratelimited() + return_val, success = fn(request, *args, **kw) + if success: + is_ratelimited(request=request, increment=True, ip=ip, + method=method, field=field, rate=rate, keys=keys) + return return_val + return _wrapped + return decorator
Use custom decorator to allow ratelimiting only on successful POST - prevents blocking form errors
## Code Before: __author__ = 'jvc26' ## Instruction: Use custom decorator to allow ratelimiting only on successful POST - prevents blocking form errors ## Code After: from functools import wraps from ratelimit.exceptions import Ratelimited from ratelimit.helpers import is_ratelimited def registration_ratelimit(ip=True, block=False, method=['POST'], field=None, rate='1/h', skip_if=None, keys=None): def decorator(fn): @wraps(fn) def _wrapped(request, *args, **kw): request.limited = getattr(request, 'limited', False) if skip_if is None or not skip_if(request): ratelimited = is_ratelimited(request=request, increment=False, ip=ip, method=method, field=field, rate=rate, keys=keys) if ratelimited and block: raise Ratelimited() return_val, success = fn(request, *args, **kw) if success: is_ratelimited(request=request, increment=True, ip=ip, method=method, field=field, rate=rate, keys=keys) return return_val return _wrapped return decorator
# ... existing code ... from functools import wraps from ratelimit.exceptions import Ratelimited from ratelimit.helpers import is_ratelimited def registration_ratelimit(ip=True, block=False, method=['POST'], field=None, rate='1/h', skip_if=None, keys=None): def decorator(fn): @wraps(fn) def _wrapped(request, *args, **kw): request.limited = getattr(request, 'limited', False) if skip_if is None or not skip_if(request): ratelimited = is_ratelimited(request=request, increment=False, ip=ip, method=method, field=field, rate=rate, keys=keys) if ratelimited and block: raise Ratelimited() return_val, success = fn(request, *args, **kw) if success: is_ratelimited(request=request, increment=True, ip=ip, method=method, field=field, rate=rate, keys=keys) return return_val return _wrapped return decorator # ... rest of the code ...
a90889b773010d2fe2ed1dff133f951c0b5baea4
demo/__init__.py
demo/__init__.py
"""Package for PythonTemplateDemo.""" __project__ = 'PythonTemplateDemo' __version__ = '0.0.0' VERSION = __project__ + '-' + __version__ PYTHON_VERSION = 2, 7 import sys if not sys.version_info >= PYTHON_VERSION: # pragma: no cover (manual test) exit("Python {}.{}+ is required.".format(*PYTHON_VERSION))
"""Package for PythonTemplateDemo.""" import sys __project__ = 'PythonTemplateDemo' __version__ = '0.0.0' VERSION = __project__ + '-' + __version__ PYTHON_VERSION = 2, 7 if not sys.version_info >= PYTHON_VERSION: # pragma: no cover (manual test) exit("Python {}.{}+ is required.".format(*PYTHON_VERSION))
Deploy Travis CI build 387 to GitHub
Deploy Travis CI build 387 to GitHub
Python
mit
jacebrowning/template-python-demo
"""Package for PythonTemplateDemo.""" + + import sys __project__ = 'PythonTemplateDemo' __version__ = '0.0.0' VERSION = __project__ + '-' + __version__ PYTHON_VERSION = 2, 7 - import sys if not sys.version_info >= PYTHON_VERSION: # pragma: no cover (manual test) exit("Python {}.{}+ is required.".format(*PYTHON_VERSION))
Deploy Travis CI build 387 to GitHub
## Code Before: """Package for PythonTemplateDemo.""" __project__ = 'PythonTemplateDemo' __version__ = '0.0.0' VERSION = __project__ + '-' + __version__ PYTHON_VERSION = 2, 7 import sys if not sys.version_info >= PYTHON_VERSION: # pragma: no cover (manual test) exit("Python {}.{}+ is required.".format(*PYTHON_VERSION)) ## Instruction: Deploy Travis CI build 387 to GitHub ## Code After: """Package for PythonTemplateDemo.""" import sys __project__ = 'PythonTemplateDemo' __version__ = '0.0.0' VERSION = __project__ + '-' + __version__ PYTHON_VERSION = 2, 7 if not sys.version_info >= PYTHON_VERSION: # pragma: no cover (manual test) exit("Python {}.{}+ is required.".format(*PYTHON_VERSION))
... """Package for PythonTemplateDemo.""" import sys ... if not sys.version_info >= PYTHON_VERSION: # pragma: no cover (manual test) ...
8eb66d72452d69d683a576c75cdf2be72b2370fa
tests/test_utils.py
tests/test_utils.py
import pytest from mangacork import utils @pytest.fixture def sample_page_bad_format(): sample_page = {'chapter': "chapter1", 'page': 3} return sample_page @pytest.fixture def sample_page_good_format(): sample_page = {'chapter':'manga_ch1', 'page':'x_v001-001'} return sample_page def test_build_img_path(sample_page_bad_format): chapter = sample_page_bad_format["chapter"] page = sample_page_bad_format["page"] expected_output = "/chapter1/3" assert utils.build_img_path(chapter,page) == expected_output def test_increment_page_number_bad_formate(sample_page_bad_format): with pytest.raises(ValueError): current_page = utils.build_img_path(sample_page_bad_format["chapter"], sample_page_bad_format["page"]) utils.increment_page_number(current_page)
import pytest from mangacork import utils @pytest.fixture def sample_page_bad_format(): sample_page = {'chapter': "chapter1", 'page': 3} return sample_page @pytest.fixture def sample_page_good_format(): sample_page = {'chapter':'manga_ch1', 'page':'x_v001-001'} return sample_page def test_build_img_path(sample_page_bad_format): chapter = sample_page_bad_format["chapter"] page = sample_page_bad_format["page"] expected_output = "/chapter1/3" assert utils.build_img_path(chapter,page) == expected_output def test_increment_page_number_bad_format(sample_page_bad_format): with pytest.raises(ValueError): current_page = utils.build_img_path(sample_page_bad_format["chapter"], sample_page_bad_format["page"]) utils.increment_page_number(current_page) def test_increment_page_number_good_format(sample_page_good_format): chapter = sample_page_good_format["chapter"] page = sample_page_good_format["page"] current_page = utils.build_img_path(chapter, page) next_page = utils.increment_page_number(current_page) expected_output = '/manga_ch1/x_v001-002' assert next_page == expected_output
Add test for inc page num, good format
Add test for inc page num, good format
Python
mit
ma3lstrom/manga-cork,ma3lstrom/manga-cork,ma3lstrom/manga-cork
import pytest from mangacork import utils @pytest.fixture def sample_page_bad_format(): sample_page = {'chapter': "chapter1", 'page': 3} return sample_page @pytest.fixture def sample_page_good_format(): sample_page = {'chapter':'manga_ch1', 'page':'x_v001-001'} return sample_page - def test_build_img_path(sample_page_bad_format): chapter = sample_page_bad_format["chapter"] page = sample_page_bad_format["page"] expected_output = "/chapter1/3" assert utils.build_img_path(chapter,page) == expected_output - def test_increment_page_number_bad_formate(sample_page_bad_format): + def test_increment_page_number_bad_format(sample_page_bad_format): with pytest.raises(ValueError): current_page = utils.build_img_path(sample_page_bad_format["chapter"], sample_page_bad_format["page"]) utils.increment_page_number(current_page) + + def test_increment_page_number_good_format(sample_page_good_format): + chapter = sample_page_good_format["chapter"] + page = sample_page_good_format["page"] + current_page = utils.build_img_path(chapter, page) + next_page = utils.increment_page_number(current_page) + expected_output = '/manga_ch1/x_v001-002' + assert next_page == expected_output +
Add test for inc page num, good format
## Code Before: import pytest from mangacork import utils @pytest.fixture def sample_page_bad_format(): sample_page = {'chapter': "chapter1", 'page': 3} return sample_page @pytest.fixture def sample_page_good_format(): sample_page = {'chapter':'manga_ch1', 'page':'x_v001-001'} return sample_page def test_build_img_path(sample_page_bad_format): chapter = sample_page_bad_format["chapter"] page = sample_page_bad_format["page"] expected_output = "/chapter1/3" assert utils.build_img_path(chapter,page) == expected_output def test_increment_page_number_bad_formate(sample_page_bad_format): with pytest.raises(ValueError): current_page = utils.build_img_path(sample_page_bad_format["chapter"], sample_page_bad_format["page"]) utils.increment_page_number(current_page) ## Instruction: Add test for inc page num, good format ## Code After: import pytest from mangacork import utils @pytest.fixture def sample_page_bad_format(): sample_page = {'chapter': "chapter1", 'page': 3} return sample_page @pytest.fixture def sample_page_good_format(): sample_page = {'chapter':'manga_ch1', 'page':'x_v001-001'} return sample_page def test_build_img_path(sample_page_bad_format): chapter = sample_page_bad_format["chapter"] page = sample_page_bad_format["page"] expected_output = "/chapter1/3" assert utils.build_img_path(chapter,page) == expected_output def test_increment_page_number_bad_format(sample_page_bad_format): with pytest.raises(ValueError): current_page = utils.build_img_path(sample_page_bad_format["chapter"], sample_page_bad_format["page"]) utils.increment_page_number(current_page) def test_increment_page_number_good_format(sample_page_good_format): chapter = sample_page_good_format["chapter"] page = sample_page_good_format["page"] current_page = utils.build_img_path(chapter, page) next_page = utils.increment_page_number(current_page) expected_output = '/manga_ch1/x_v001-002' assert next_page == expected_output
# ... existing code ... def test_build_img_path(sample_page_bad_format): # ... modified code ... def test_increment_page_number_bad_format(sample_page_bad_format): with pytest.raises(ValueError): ... utils.increment_page_number(current_page) def test_increment_page_number_good_format(sample_page_good_format): chapter = sample_page_good_format["chapter"] page = sample_page_good_format["page"] current_page = utils.build_img_path(chapter, page) next_page = utils.increment_page_number(current_page) expected_output = '/manga_ch1/x_v001-002' assert next_page == expected_output # ... rest of the code ...
66edf9f04c1b23681fae4234a8b297868e66b7aa
osmaxx-py/osmaxx/excerptexport/models/excerpt.py
osmaxx-py/osmaxx/excerptexport/models/excerpt.py
from django.db import models from django.contrib.auth.models import User from django.utils.translation import gettext_lazy as _ class Excerpt(models.Model): name = models.CharField(max_length=128, verbose_name=_('name'), blank=False) is_public = models.BooleanField(default=False, verbose_name=_('is public')) is_active = models.BooleanField(default=True, verbose_name=_('is active')) owner = models.ForeignKey(User, related_name='excerpts', verbose_name=_('owner')) bounding_geometry = models.OneToOneField('BoundingGeometry', verbose_name=_('bounding geometry')) @property def type_of_geometry(self): return self.bounding_geometry.type_of_geometry @property def extent(self): return self.bounding_geometry.extent def __str__(self): return self.name def _active_excerpts(): return Excerpt.objects.filter(is_active=True).filter( bounding_geometry__bboxboundinggeometry__isnull=False ) def private_user_excerpts(user): return _active_excerpts().filter(is_public=False, owner=user) def public_user_excerpts(user): return _active_excerpts().filter(is_public=True, owner=user) def other_users_public_excerpts(user): return _active_excerpts().filter(is_public=True).exclude(owner=user)
from django.db import models from django.contrib.auth.models import User from django.utils.translation import gettext_lazy as _ class Excerpt(models.Model): name = models.CharField(max_length=128, verbose_name=_('name')) is_public = models.BooleanField(default=False, verbose_name=_('is public')) is_active = models.BooleanField(default=True, verbose_name=_('is active')) owner = models.ForeignKey(User, related_name='excerpts', verbose_name=_('owner')) bounding_geometry = models.OneToOneField('BoundingGeometry', verbose_name=_('bounding geometry')) @property def type_of_geometry(self): return self.bounding_geometry.type_of_geometry @property def extent(self): return self.bounding_geometry.extent def __str__(self): return self.name def _active_excerpts(): return Excerpt.objects.filter(is_active=True).filter( bounding_geometry__bboxboundinggeometry__isnull=False ) def private_user_excerpts(user): return _active_excerpts().filter(is_public=False, owner=user) def public_user_excerpts(user): return _active_excerpts().filter(is_public=True, owner=user) def other_users_public_excerpts(user): return _active_excerpts().filter(is_public=True).exclude(owner=user)
Remove value which is already default
Remove value which is already default
Python
mit
geometalab/osmaxx,geometalab/drf-utm-zone-info,geometalab/osmaxx,geometalab/osmaxx-frontend,geometalab/osmaxx-frontend,geometalab/osmaxx-frontend,geometalab/osmaxx,geometalab/osmaxx-frontend,geometalab/drf-utm-zone-info,geometalab/osmaxx
from django.db import models from django.contrib.auth.models import User from django.utils.translation import gettext_lazy as _ class Excerpt(models.Model): - name = models.CharField(max_length=128, verbose_name=_('name'), blank=False) + name = models.CharField(max_length=128, verbose_name=_('name')) is_public = models.BooleanField(default=False, verbose_name=_('is public')) is_active = models.BooleanField(default=True, verbose_name=_('is active')) owner = models.ForeignKey(User, related_name='excerpts', verbose_name=_('owner')) bounding_geometry = models.OneToOneField('BoundingGeometry', verbose_name=_('bounding geometry')) @property def type_of_geometry(self): return self.bounding_geometry.type_of_geometry @property def extent(self): return self.bounding_geometry.extent def __str__(self): return self.name def _active_excerpts(): return Excerpt.objects.filter(is_active=True).filter( bounding_geometry__bboxboundinggeometry__isnull=False ) def private_user_excerpts(user): return _active_excerpts().filter(is_public=False, owner=user) def public_user_excerpts(user): return _active_excerpts().filter(is_public=True, owner=user) def other_users_public_excerpts(user): return _active_excerpts().filter(is_public=True).exclude(owner=user)
Remove value which is already default
## Code Before: from django.db import models from django.contrib.auth.models import User from django.utils.translation import gettext_lazy as _ class Excerpt(models.Model): name = models.CharField(max_length=128, verbose_name=_('name'), blank=False) is_public = models.BooleanField(default=False, verbose_name=_('is public')) is_active = models.BooleanField(default=True, verbose_name=_('is active')) owner = models.ForeignKey(User, related_name='excerpts', verbose_name=_('owner')) bounding_geometry = models.OneToOneField('BoundingGeometry', verbose_name=_('bounding geometry')) @property def type_of_geometry(self): return self.bounding_geometry.type_of_geometry @property def extent(self): return self.bounding_geometry.extent def __str__(self): return self.name def _active_excerpts(): return Excerpt.objects.filter(is_active=True).filter( bounding_geometry__bboxboundinggeometry__isnull=False ) def private_user_excerpts(user): return _active_excerpts().filter(is_public=False, owner=user) def public_user_excerpts(user): return _active_excerpts().filter(is_public=True, owner=user) def other_users_public_excerpts(user): return _active_excerpts().filter(is_public=True).exclude(owner=user) ## Instruction: Remove value which is already default ## Code After: from django.db import models from django.contrib.auth.models import User from django.utils.translation import gettext_lazy as _ class Excerpt(models.Model): name = models.CharField(max_length=128, verbose_name=_('name')) is_public = models.BooleanField(default=False, verbose_name=_('is public')) is_active = models.BooleanField(default=True, verbose_name=_('is active')) owner = models.ForeignKey(User, related_name='excerpts', verbose_name=_('owner')) bounding_geometry = models.OneToOneField('BoundingGeometry', verbose_name=_('bounding geometry')) @property def type_of_geometry(self): return self.bounding_geometry.type_of_geometry @property def extent(self): return self.bounding_geometry.extent def __str__(self): return self.name def _active_excerpts(): return Excerpt.objects.filter(is_active=True).filter( bounding_geometry__bboxboundinggeometry__isnull=False ) def private_user_excerpts(user): return _active_excerpts().filter(is_public=False, owner=user) def public_user_excerpts(user): return _active_excerpts().filter(is_public=True, owner=user) def other_users_public_excerpts(user): return _active_excerpts().filter(is_public=True).exclude(owner=user)
... class Excerpt(models.Model): name = models.CharField(max_length=128, verbose_name=_('name')) is_public = models.BooleanField(default=False, verbose_name=_('is public')) ...
03ee406800fb59ff3e7565397107fa9aad0d54d0
website/notifications/listeners.py
website/notifications/listeners.py
import logging from website.notifications.exceptions import InvalidSubscriptionError from website.notifications.utils import subscribe_user_to_notifications, subscribe_user_to_global_notifications from website.project.signals import contributor_added, project_created from framework.auth.signals import user_confirmed logger = logging.getLogger(__name__) @project_created.connect def subscribe_creator(node): if node.is_collection or node.is_deleted: return None try: subscribe_user_to_notifications(node, node.creator) except InvalidSubscriptionError as err: user = node.creator._id if node.creator else 'None' logger.warn('Skipping subscription of user {} to node {}'.format(user, node._id)) logger.warn('Reason: {}'.format(str(err))) @contributor_added.connect def subscribe_contributor(node, contributor, auth=None, *args, **kwargs): try: subscribe_user_to_notifications(node, contributor) except InvalidSubscriptionError as err: logger.warn('Skipping subscription of user {} to node {}'.format(contributor, node._id)) logger.warn('Reason: {}'.format(str(err))) @user_confirmed.connect def subscribe_confirmed_user(user): try: subscribe_user_to_global_notifications(user) except InvalidSubscriptionError as err: logger.warn('Skipping subscription of user {} to global subscriptions'.format(user)) logger.warn('Reason: {}'.format(str(err)))
import logging from website.notifications.exceptions import InvalidSubscriptionError from website.notifications.utils import subscribe_user_to_notifications, subscribe_user_to_global_notifications from website.project.signals import contributor_added, project_created from framework.auth.signals import user_confirmed logger = logging.getLogger(__name__) @project_created.connect def subscribe_creator(node): if node.institution_id or node.is_collection or node.is_deleted: return None try: subscribe_user_to_notifications(node, node.creator) except InvalidSubscriptionError as err: user = node.creator._id if node.creator else 'None' logger.warn('Skipping subscription of user {} to node {}'.format(user, node._id)) logger.warn('Reason: {}'.format(str(err))) @contributor_added.connect def subscribe_contributor(node, contributor, auth=None, *args, **kwargs): try: subscribe_user_to_notifications(node, contributor) except InvalidSubscriptionError as err: logger.warn('Skipping subscription of user {} to node {}'.format(contributor, node._id)) logger.warn('Reason: {}'.format(str(err))) @user_confirmed.connect def subscribe_confirmed_user(user): try: subscribe_user_to_global_notifications(user) except InvalidSubscriptionError as err: logger.warn('Skipping subscription of user {} to global subscriptions'.format(user)) logger.warn('Reason: {}'.format(str(err)))
Revert "Remove incorrect check for institution_id"
Revert "Remove incorrect check for institution_id" This reverts commit 617df13670573b858b6c23249f4287786807d8b6.
Python
apache-2.0
hmoco/osf.io,cslzchen/osf.io,monikagrabowska/osf.io,cslzchen/osf.io,cslzchen/osf.io,Nesiehr/osf.io,aaxelb/osf.io,CenterForOpenScience/osf.io,chrisseto/osf.io,chennan47/osf.io,crcresearch/osf.io,Nesiehr/osf.io,felliott/osf.io,Johnetordoff/osf.io,acshi/osf.io,baylee-d/osf.io,HalcyonChimera/osf.io,brianjgeiger/osf.io,aaxelb/osf.io,binoculars/osf.io,cwisecarver/osf.io,CenterForOpenScience/osf.io,mattclark/osf.io,sloria/osf.io,TomBaxter/osf.io,caneruguz/osf.io,adlius/osf.io,hmoco/osf.io,caneruguz/osf.io,chennan47/osf.io,chrisseto/osf.io,mattclark/osf.io,saradbowman/osf.io,aaxelb/osf.io,mfraezz/osf.io,felliott/osf.io,monikagrabowska/osf.io,sloria/osf.io,brianjgeiger/osf.io,CenterForOpenScience/osf.io,TomBaxter/osf.io,erinspace/osf.io,erinspace/osf.io,leb2dg/osf.io,adlius/osf.io,pattisdr/osf.io,laurenrevere/osf.io,caneruguz/osf.io,mfraezz/osf.io,CenterForOpenScience/osf.io,crcresearch/osf.io,adlius/osf.io,adlius/osf.io,cwisecarver/osf.io,leb2dg/osf.io,baylee-d/osf.io,mfraezz/osf.io,caseyrollins/osf.io,mattclark/osf.io,icereval/osf.io,brianjgeiger/osf.io,Nesiehr/osf.io,leb2dg/osf.io,crcresearch/osf.io,monikagrabowska/osf.io,felliott/osf.io,erinspace/osf.io,laurenrevere/osf.io,leb2dg/osf.io,binoculars/osf.io,HalcyonChimera/osf.io,Johnetordoff/osf.io,pattisdr/osf.io,chennan47/osf.io,hmoco/osf.io,acshi/osf.io,cslzchen/osf.io,monikagrabowska/osf.io,Nesiehr/osf.io,pattisdr/osf.io,icereval/osf.io,HalcyonChimera/osf.io,caseyrollins/osf.io,caseyrollins/osf.io,aaxelb/osf.io,laurenrevere/osf.io,TomBaxter/osf.io,monikagrabowska/osf.io,acshi/osf.io,Johnetordoff/osf.io,Johnetordoff/osf.io,cwisecarver/osf.io,chrisseto/osf.io,baylee-d/osf.io,icereval/osf.io,binoculars/osf.io,felliott/osf.io,mfraezz/osf.io,HalcyonChimera/osf.io,brianjgeiger/osf.io,hmoco/osf.io,sloria/osf.io,saradbowman/osf.io,acshi/osf.io,caneruguz/osf.io,acshi/osf.io,chrisseto/osf.io,cwisecarver/osf.io
import logging from website.notifications.exceptions import InvalidSubscriptionError from website.notifications.utils import subscribe_user_to_notifications, subscribe_user_to_global_notifications from website.project.signals import contributor_added, project_created from framework.auth.signals import user_confirmed logger = logging.getLogger(__name__) @project_created.connect def subscribe_creator(node): - if node.is_collection or node.is_deleted: + if node.institution_id or node.is_collection or node.is_deleted: return None try: subscribe_user_to_notifications(node, node.creator) except InvalidSubscriptionError as err: user = node.creator._id if node.creator else 'None' logger.warn('Skipping subscription of user {} to node {}'.format(user, node._id)) logger.warn('Reason: {}'.format(str(err))) @contributor_added.connect def subscribe_contributor(node, contributor, auth=None, *args, **kwargs): try: subscribe_user_to_notifications(node, contributor) except InvalidSubscriptionError as err: logger.warn('Skipping subscription of user {} to node {}'.format(contributor, node._id)) logger.warn('Reason: {}'.format(str(err))) @user_confirmed.connect def subscribe_confirmed_user(user): try: subscribe_user_to_global_notifications(user) except InvalidSubscriptionError as err: logger.warn('Skipping subscription of user {} to global subscriptions'.format(user)) logger.warn('Reason: {}'.format(str(err)))
Revert "Remove incorrect check for institution_id"
## Code Before: import logging from website.notifications.exceptions import InvalidSubscriptionError from website.notifications.utils import subscribe_user_to_notifications, subscribe_user_to_global_notifications from website.project.signals import contributor_added, project_created from framework.auth.signals import user_confirmed logger = logging.getLogger(__name__) @project_created.connect def subscribe_creator(node): if node.is_collection or node.is_deleted: return None try: subscribe_user_to_notifications(node, node.creator) except InvalidSubscriptionError as err: user = node.creator._id if node.creator else 'None' logger.warn('Skipping subscription of user {} to node {}'.format(user, node._id)) logger.warn('Reason: {}'.format(str(err))) @contributor_added.connect def subscribe_contributor(node, contributor, auth=None, *args, **kwargs): try: subscribe_user_to_notifications(node, contributor) except InvalidSubscriptionError as err: logger.warn('Skipping subscription of user {} to node {}'.format(contributor, node._id)) logger.warn('Reason: {}'.format(str(err))) @user_confirmed.connect def subscribe_confirmed_user(user): try: subscribe_user_to_global_notifications(user) except InvalidSubscriptionError as err: logger.warn('Skipping subscription of user {} to global subscriptions'.format(user)) logger.warn('Reason: {}'.format(str(err))) ## Instruction: Revert "Remove incorrect check for institution_id" ## Code After: import logging from website.notifications.exceptions import InvalidSubscriptionError from website.notifications.utils import subscribe_user_to_notifications, subscribe_user_to_global_notifications from website.project.signals import contributor_added, project_created from framework.auth.signals import user_confirmed logger = logging.getLogger(__name__) @project_created.connect def subscribe_creator(node): if node.institution_id or node.is_collection or node.is_deleted: return None try: subscribe_user_to_notifications(node, node.creator) except InvalidSubscriptionError as err: user = node.creator._id if node.creator else 'None' logger.warn('Skipping subscription of user {} to node {}'.format(user, node._id)) logger.warn('Reason: {}'.format(str(err))) @contributor_added.connect def subscribe_contributor(node, contributor, auth=None, *args, **kwargs): try: subscribe_user_to_notifications(node, contributor) except InvalidSubscriptionError as err: logger.warn('Skipping subscription of user {} to node {}'.format(contributor, node._id)) logger.warn('Reason: {}'.format(str(err))) @user_confirmed.connect def subscribe_confirmed_user(user): try: subscribe_user_to_global_notifications(user) except InvalidSubscriptionError as err: logger.warn('Skipping subscription of user {} to global subscriptions'.format(user)) logger.warn('Reason: {}'.format(str(err)))
// ... existing code ... def subscribe_creator(node): if node.institution_id or node.is_collection or node.is_deleted: return None // ... rest of the code ...
85814828d2caedd8612db6ce0ecec92025a34330
tests/test_main.py
tests/test_main.py
from cookiecutter.main import is_repo_url def test_is_repo_url(): """Verify is_repo_url works.""" assert is_repo_url('gitolite@server:team/repo') is True assert is_repo_url('[email protected]:audreyr/cookiecutter.git') is True assert is_repo_url('https://github.com/audreyr/cookiecutter.git') is True assert is_repo_url('gh:audreyr/cookiecutter-pypackage') is True assert is_repo_url('/audreyr/cookiecutter.git') is False assert is_repo_url('/home/audreyr/cookiecutter') is False
from cookiecutter.main import is_repo_url def test_is_repo_url(): """Verify is_repo_url works.""" assert is_repo_url('gitolite@server:team/repo') is True assert is_repo_url('[email protected]:audreyr/cookiecutter.git') is True assert is_repo_url('https://github.com/audreyr/cookiecutter.git') is True assert is_repo_url('gh:audreyr/cookiecutter-pypackage') is True assert is_repo_url('https://bitbucket.org/pokoli/cookiecutter.hg') is True assert is_repo_url('/audreyr/cookiecutter.git') is False assert is_repo_url('/home/audreyr/cookiecutter') is False
Add test for bitbucket domain
Add test for bitbucket domain
Python
bsd-3-clause
michaeljoseph/cookiecutter,Springerle/cookiecutter,Springerle/cookiecutter,venumech/cookiecutter,cguardia/cookiecutter,luzfcb/cookiecutter,pjbull/cookiecutter,agconti/cookiecutter,willingc/cookiecutter,audreyr/cookiecutter,audreyr/cookiecutter,venumech/cookiecutter,takeflight/cookiecutter,dajose/cookiecutter,takeflight/cookiecutter,agconti/cookiecutter,terryjbates/cookiecutter,cguardia/cookiecutter,stevepiercy/cookiecutter,pjbull/cookiecutter,hackebrot/cookiecutter,luzfcb/cookiecutter,christabor/cookiecutter,ramiroluz/cookiecutter,hackebrot/cookiecutter,christabor/cookiecutter,benthomasson/cookiecutter,benthomasson/cookiecutter,michaeljoseph/cookiecutter,ramiroluz/cookiecutter,moi65/cookiecutter,stevepiercy/cookiecutter,dajose/cookiecutter,terryjbates/cookiecutter,moi65/cookiecutter,willingc/cookiecutter
from cookiecutter.main import is_repo_url def test_is_repo_url(): """Verify is_repo_url works.""" assert is_repo_url('gitolite@server:team/repo') is True assert is_repo_url('[email protected]:audreyr/cookiecutter.git') is True assert is_repo_url('https://github.com/audreyr/cookiecutter.git') is True assert is_repo_url('gh:audreyr/cookiecutter-pypackage') is True + assert is_repo_url('https://bitbucket.org/pokoli/cookiecutter.hg') is True assert is_repo_url('/audreyr/cookiecutter.git') is False assert is_repo_url('/home/audreyr/cookiecutter') is False
Add test for bitbucket domain
## Code Before: from cookiecutter.main import is_repo_url def test_is_repo_url(): """Verify is_repo_url works.""" assert is_repo_url('gitolite@server:team/repo') is True assert is_repo_url('[email protected]:audreyr/cookiecutter.git') is True assert is_repo_url('https://github.com/audreyr/cookiecutter.git') is True assert is_repo_url('gh:audreyr/cookiecutter-pypackage') is True assert is_repo_url('/audreyr/cookiecutter.git') is False assert is_repo_url('/home/audreyr/cookiecutter') is False ## Instruction: Add test for bitbucket domain ## Code After: from cookiecutter.main import is_repo_url def test_is_repo_url(): """Verify is_repo_url works.""" assert is_repo_url('gitolite@server:team/repo') is True assert is_repo_url('[email protected]:audreyr/cookiecutter.git') is True assert is_repo_url('https://github.com/audreyr/cookiecutter.git') is True assert is_repo_url('gh:audreyr/cookiecutter-pypackage') is True assert is_repo_url('https://bitbucket.org/pokoli/cookiecutter.hg') is True assert is_repo_url('/audreyr/cookiecutter.git') is False assert is_repo_url('/home/audreyr/cookiecutter') is False
// ... existing code ... assert is_repo_url('gh:audreyr/cookiecutter-pypackage') is True assert is_repo_url('https://bitbucket.org/pokoli/cookiecutter.hg') is True // ... rest of the code ...
461019099c41ca4ef2fc7ccfec0141ed5b7e3bd6
tests/test_unicode.py
tests/test_unicode.py
import sys import pytest import jupytext from .utils import list_all_notebooks @pytest.mark.parametrize('nb_file', list_all_notebooks('.ipynb') + list_all_notebooks('.Rmd')) def test_notebook_contents_is_unicode(nb_file): nb = jupytext.readf(nb_file) for cell in nb.cells: if sys.version_info < (3, 0): assert cell.source == '' or isinstance(cell.source, unicode) else: assert isinstance(cell.source, str) def test_write_non_ascii(tmpdir): nb = jupytext.reads(u'Non-ascii contênt', ext='.Rmd') jupytext.writef(nb, str(tmpdir.join('notebook.Rmd'))) jupytext.writef(nb, str(tmpdir.join('notebook.ipynb')))
import sys import pytest import jupytext from .utils import list_all_notebooks try: unicode # Python 2 except NameError: unicode = str # Python 3 @pytest.mark.parametrize('nb_file', list_all_notebooks('.ipynb') + list_all_notebooks('.Rmd')) def test_notebook_contents_is_unicode(nb_file): nb = jupytext.readf(nb_file) for cell in nb.cells: if sys.version_info < (3, 0): assert cell.source == '' or isinstance(cell.source, unicode) else: assert isinstance(cell.source, str) def test_write_non_ascii(tmpdir): nb = jupytext.reads(u'Non-ascii contênt', ext='.Rmd') jupytext.writef(nb, str(tmpdir.join('notebook.Rmd'))) jupytext.writef(nb, str(tmpdir.join('notebook.ipynb')))
Define unicode in Python 3
Define unicode in Python 3 __unicode__ was removed in Python 3 because all __str__ are Unicode. [flake8](http://flake8.pycqa.org) testing of https://github.com/mwouts/jupytext on Python 3.7.0 $ __flake8 . --count --select=E901,E999,F821,F822,F823 --show-source --statistics__ ``` ./.jupyter/jupyter_notebook_config.py:1:1: F821 undefined name 'c' c.NotebookApp.contents_manager_class = 'jupytext.TextFileContentsManager' ^ ./tests/test_unicode.py:15:65: F821 undefined name 'unicode' assert cell.source == '' or isinstance(cell.source, unicode) ^ ./tests/mirror/jupyter_again.py:32:1: E999 SyntaxError: invalid syntax ?next ^ 1 E999 SyntaxError: invalid syntax 2 F821 undefined name 'c' 3 ```
Python
mit
mwouts/jupytext,mwouts/jupytext,mwouts/jupytext,mwouts/jupytext,mwouts/jupytext,mwouts/jupytext,mwouts/jupytext,mwouts/jupytext,mwouts/jupytext,mwouts/jupytext
import sys import pytest import jupytext from .utils import list_all_notebooks + + try: + unicode # Python 2 + except NameError: + unicode = str # Python 3 @pytest.mark.parametrize('nb_file', list_all_notebooks('.ipynb') + list_all_notebooks('.Rmd')) def test_notebook_contents_is_unicode(nb_file): nb = jupytext.readf(nb_file) for cell in nb.cells: if sys.version_info < (3, 0): assert cell.source == '' or isinstance(cell.source, unicode) else: assert isinstance(cell.source, str) def test_write_non_ascii(tmpdir): nb = jupytext.reads(u'Non-ascii contênt', ext='.Rmd') jupytext.writef(nb, str(tmpdir.join('notebook.Rmd'))) jupytext.writef(nb, str(tmpdir.join('notebook.ipynb')))
Define unicode in Python 3
## Code Before: import sys import pytest import jupytext from .utils import list_all_notebooks @pytest.mark.parametrize('nb_file', list_all_notebooks('.ipynb') + list_all_notebooks('.Rmd')) def test_notebook_contents_is_unicode(nb_file): nb = jupytext.readf(nb_file) for cell in nb.cells: if sys.version_info < (3, 0): assert cell.source == '' or isinstance(cell.source, unicode) else: assert isinstance(cell.source, str) def test_write_non_ascii(tmpdir): nb = jupytext.reads(u'Non-ascii contênt', ext='.Rmd') jupytext.writef(nb, str(tmpdir.join('notebook.Rmd'))) jupytext.writef(nb, str(tmpdir.join('notebook.ipynb'))) ## Instruction: Define unicode in Python 3 ## Code After: import sys import pytest import jupytext from .utils import list_all_notebooks try: unicode # Python 2 except NameError: unicode = str # Python 3 @pytest.mark.parametrize('nb_file', list_all_notebooks('.ipynb') + list_all_notebooks('.Rmd')) def test_notebook_contents_is_unicode(nb_file): nb = jupytext.readf(nb_file) for cell in nb.cells: if sys.version_info < (3, 0): assert cell.source == '' or isinstance(cell.source, unicode) else: assert isinstance(cell.source, str) def test_write_non_ascii(tmpdir): nb = jupytext.reads(u'Non-ascii contênt', ext='.Rmd') jupytext.writef(nb, str(tmpdir.join('notebook.Rmd'))) jupytext.writef(nb, str(tmpdir.join('notebook.ipynb')))
// ... existing code ... from .utils import list_all_notebooks try: unicode # Python 2 except NameError: unicode = str # Python 3 // ... rest of the code ...
3290d532f3dd9c1e24921c4b80aeb6e860bc86a8
spock/plugins/__init__.py
spock/plugins/__init__.py
from spock.plugins.core import auth, event, net, ticker, timer from spock.plugins.helpers import clientinfo, entities, interact, inventory,\ keepalive, movement, physics, respawn, start, world from spock.plugins.base import PluginBase # noqa core_plugins = [ ('auth', auth.AuthPlugin), ('event', event.EventPlugin), ('net', net.NetPlugin), ('ticker', ticker.TickerPlugin), ('timers', timer.TimerPlugin), ] helper_plugins = [ ('clientinfo', clientinfo.ClientInfoPlugin), ('entities', entities.EntitiesPlugin), ('interact', interact.InteractPlugin), ('inventory', inventory.InventoryPlugin), ('keepalive', keepalive.KeepalivePlugin), ('movement', movement.MovementPlugin), ('physics', physics.PhysicsPlugin), ('respawn', respawn.RespawnPlugin), ('start', start.StartPlugin), ('world', world.WorldPlugin), ] default_plugins = core_plugins + helper_plugins
from spock.plugins.core import auth, event, net, ticker, timer from spock.plugins.helpers import chat, clientinfo, entities, interact, \ inventory, keepalive, movement, physics, respawn, start, world from spock.plugins.base import PluginBase # noqa core_plugins = [ ('auth', auth.AuthPlugin), ('event', event.EventPlugin), ('net', net.NetPlugin), ('ticker', ticker.TickerPlugin), ('timers', timer.TimerPlugin), ] helper_plugins = [ ('chat', chat.ChatPlugin), ('clientinfo', clientinfo.ClientInfoPlugin), ('entities', entities.EntitiesPlugin), ('interact', interact.InteractPlugin), ('inventory', inventory.InventoryPlugin), ('keepalive', keepalive.KeepalivePlugin), ('movement', movement.MovementPlugin), ('physics', physics.PhysicsPlugin), ('respawn', respawn.RespawnPlugin), ('start', start.StartPlugin), ('world', world.WorldPlugin), ] default_plugins = core_plugins + helper_plugins
Fix removal of chat plugin
Fix removal of chat plugin
Python
mit
nickelpro/SpockBot,MrSwiss/SpockBot,Gjum/SpockBot,SpockBotMC/SpockBot,gamingrobot/SpockBot,luken/SpockBot
from spock.plugins.core import auth, event, net, ticker, timer - from spock.plugins.helpers import clientinfo, entities, interact, inventory,\ + from spock.plugins.helpers import chat, clientinfo, entities, interact, \ - keepalive, movement, physics, respawn, start, world + inventory, keepalive, movement, physics, respawn, start, world - from spock.plugins.base import PluginBase # noqa core_plugins = [ ('auth', auth.AuthPlugin), ('event', event.EventPlugin), ('net', net.NetPlugin), ('ticker', ticker.TickerPlugin), ('timers', timer.TimerPlugin), ] helper_plugins = [ + ('chat', chat.ChatPlugin), ('clientinfo', clientinfo.ClientInfoPlugin), ('entities', entities.EntitiesPlugin), ('interact', interact.InteractPlugin), ('inventory', inventory.InventoryPlugin), ('keepalive', keepalive.KeepalivePlugin), ('movement', movement.MovementPlugin), ('physics', physics.PhysicsPlugin), ('respawn', respawn.RespawnPlugin), ('start', start.StartPlugin), ('world', world.WorldPlugin), ] default_plugins = core_plugins + helper_plugins
Fix removal of chat plugin
## Code Before: from spock.plugins.core import auth, event, net, ticker, timer from spock.plugins.helpers import clientinfo, entities, interact, inventory,\ keepalive, movement, physics, respawn, start, world from spock.plugins.base import PluginBase # noqa core_plugins = [ ('auth', auth.AuthPlugin), ('event', event.EventPlugin), ('net', net.NetPlugin), ('ticker', ticker.TickerPlugin), ('timers', timer.TimerPlugin), ] helper_plugins = [ ('clientinfo', clientinfo.ClientInfoPlugin), ('entities', entities.EntitiesPlugin), ('interact', interact.InteractPlugin), ('inventory', inventory.InventoryPlugin), ('keepalive', keepalive.KeepalivePlugin), ('movement', movement.MovementPlugin), ('physics', physics.PhysicsPlugin), ('respawn', respawn.RespawnPlugin), ('start', start.StartPlugin), ('world', world.WorldPlugin), ] default_plugins = core_plugins + helper_plugins ## Instruction: Fix removal of chat plugin ## Code After: from spock.plugins.core import auth, event, net, ticker, timer from spock.plugins.helpers import chat, clientinfo, entities, interact, \ inventory, keepalive, movement, physics, respawn, start, world from spock.plugins.base import PluginBase # noqa core_plugins = [ ('auth', auth.AuthPlugin), ('event', event.EventPlugin), ('net', net.NetPlugin), ('ticker', ticker.TickerPlugin), ('timers', timer.TimerPlugin), ] helper_plugins = [ ('chat', chat.ChatPlugin), ('clientinfo', clientinfo.ClientInfoPlugin), ('entities', entities.EntitiesPlugin), ('interact', interact.InteractPlugin), ('inventory', inventory.InventoryPlugin), ('keepalive', keepalive.KeepalivePlugin), ('movement', movement.MovementPlugin), ('physics', physics.PhysicsPlugin), ('respawn', respawn.RespawnPlugin), ('start', start.StartPlugin), ('world', world.WorldPlugin), ] default_plugins = core_plugins + helper_plugins
... from spock.plugins.core import auth, event, net, ticker, timer from spock.plugins.helpers import chat, clientinfo, entities, interact, \ inventory, keepalive, movement, physics, respawn, start, world from spock.plugins.base import PluginBase # noqa ... helper_plugins = [ ('chat', chat.ChatPlugin), ('clientinfo', clientinfo.ClientInfoPlugin), ...
d5a3285b05d96ffc99049867256cdba87a5b420a
packages/mono_crypto.py
packages/mono_crypto.py
from mono_master import MonoMasterPackage from bockbuild.util.util import * class MonoMasterEncryptedPackage (MonoMasterPackage): def __init__(self): MonoMasterPackage.__init__ (self) self.configure_flags.extend(['--enable-extension-module=crypto --enable-native-types']) def prep(self): MonoMasterPackage.prep(self) retry (self.checkout_mono_extensions) def checkout_mono_extensions(self): ext = '[email protected]:xamarin/mono-extensions.git' dirname = os.path.join(self.profile.build_root, "mono-extensions") if not os.path.exists(dirname): self.sh('%' + '{git} clone --local --shared "%s" "%s"' % (ext, dirname)) self.pushd(dirname) try: self.sh('%{git} clean -xffd') self.sh('%{git} fetch --all --prune') if "pr/" not in self.git_branch: self.sh('%' + '{git} checkout origin/%s' % self.git_branch) else: self.sh('%{git} checkout origin/master') self.sh ('%{git} reset --hard') except Exception as e: self.popd () self.rm_if_exists (dirname) raise finally: info ('Mono crypto extensions (rev. %s)' % git_get_revision (self)) self.popd () MonoMasterEncryptedPackage()
from mono_master import MonoMasterPackage from bockbuild.util.util import * class MonoMasterEncryptedPackage (MonoMasterPackage): def __init__(self): MonoMasterPackage.__init__ (self) self.configure_flags.extend(['--enable-extension-module=crypto --enable-native-types']) def prep(self): MonoMasterPackage.prep(self) retry (self.checkout_mono_extensions) def checkout_mono_extensions(self): ext = '[email protected]:xamarin/mono-extensions.git' dirname = os.path.join(self.profile.build_root, "mono-extensions") if not os.path.exists(dirname): self.sh('%' + '{git} clone --local --shared "%s" "%s"' % (ext, dirname)) self.pushd(dirname) try: self.sh('%{git} clean -xffd') self.sh('%{git} fetch --all --prune') if 'pull/' in self.git_branch: # pull request self.sh('%{git} checkout origin/master') else: self.sh('%' + '{git} checkout origin/%s' % self.git_branch) self.sh ('%{git} reset --hard') except Exception as e: self.popd () self.rm_if_exists (dirname) raise finally: info ('Mono crypto extensions (rev. %s)' % git_get_revision (self)) self.popd () MonoMasterEncryptedPackage()
Fix mono-extensions checkout for PR branches ('origin/pull/N/merge')
Fix mono-extensions checkout for PR branches ('origin/pull/N/merge')
Python
mit
mono/bockbuild,mono/bockbuild
from mono_master import MonoMasterPackage from bockbuild.util.util import * class MonoMasterEncryptedPackage (MonoMasterPackage): def __init__(self): MonoMasterPackage.__init__ (self) self.configure_flags.extend(['--enable-extension-module=crypto --enable-native-types']) def prep(self): MonoMasterPackage.prep(self) retry (self.checkout_mono_extensions) def checkout_mono_extensions(self): ext = '[email protected]:xamarin/mono-extensions.git' dirname = os.path.join(self.profile.build_root, "mono-extensions") if not os.path.exists(dirname): self.sh('%' + '{git} clone --local --shared "%s" "%s"' % (ext, dirname)) self.pushd(dirname) try: self.sh('%{git} clean -xffd') self.sh('%{git} fetch --all --prune') - if "pr/" not in self.git_branch: + if 'pull/' in self.git_branch: # pull request + self.sh('%{git} checkout origin/master') + else: self.sh('%' + '{git} checkout origin/%s' % self.git_branch) + - else: - self.sh('%{git} checkout origin/master') self.sh ('%{git} reset --hard') except Exception as e: self.popd () self.rm_if_exists (dirname) raise finally: info ('Mono crypto extensions (rev. %s)' % git_get_revision (self)) self.popd () MonoMasterEncryptedPackage()
Fix mono-extensions checkout for PR branches ('origin/pull/N/merge')
## Code Before: from mono_master import MonoMasterPackage from bockbuild.util.util import * class MonoMasterEncryptedPackage (MonoMasterPackage): def __init__(self): MonoMasterPackage.__init__ (self) self.configure_flags.extend(['--enable-extension-module=crypto --enable-native-types']) def prep(self): MonoMasterPackage.prep(self) retry (self.checkout_mono_extensions) def checkout_mono_extensions(self): ext = '[email protected]:xamarin/mono-extensions.git' dirname = os.path.join(self.profile.build_root, "mono-extensions") if not os.path.exists(dirname): self.sh('%' + '{git} clone --local --shared "%s" "%s"' % (ext, dirname)) self.pushd(dirname) try: self.sh('%{git} clean -xffd') self.sh('%{git} fetch --all --prune') if "pr/" not in self.git_branch: self.sh('%' + '{git} checkout origin/%s' % self.git_branch) else: self.sh('%{git} checkout origin/master') self.sh ('%{git} reset --hard') except Exception as e: self.popd () self.rm_if_exists (dirname) raise finally: info ('Mono crypto extensions (rev. %s)' % git_get_revision (self)) self.popd () MonoMasterEncryptedPackage() ## Instruction: Fix mono-extensions checkout for PR branches ('origin/pull/N/merge') ## Code After: from mono_master import MonoMasterPackage from bockbuild.util.util import * class MonoMasterEncryptedPackage (MonoMasterPackage): def __init__(self): MonoMasterPackage.__init__ (self) self.configure_flags.extend(['--enable-extension-module=crypto --enable-native-types']) def prep(self): MonoMasterPackage.prep(self) retry (self.checkout_mono_extensions) def checkout_mono_extensions(self): ext = '[email protected]:xamarin/mono-extensions.git' dirname = os.path.join(self.profile.build_root, "mono-extensions") if not os.path.exists(dirname): self.sh('%' + '{git} clone --local --shared "%s" "%s"' % (ext, dirname)) self.pushd(dirname) try: self.sh('%{git} clean -xffd') self.sh('%{git} fetch --all --prune') if 'pull/' in self.git_branch: # pull request self.sh('%{git} checkout origin/master') else: self.sh('%' + '{git} checkout origin/%s' % self.git_branch) self.sh ('%{git} reset --hard') except Exception as e: self.popd () self.rm_if_exists (dirname) raise finally: info ('Mono crypto extensions (rev. %s)' % git_get_revision (self)) self.popd () MonoMasterEncryptedPackage()
... self.sh('%{git} fetch --all --prune') if 'pull/' in self.git_branch: # pull request self.sh('%{git} checkout origin/master') else: self.sh('%' + '{git} checkout origin/%s' % self.git_branch) self.sh ('%{git} reset --hard') ...
a92121cfdbb94d36d021fb8d1386031829ee86a2
patterns/solid.py
patterns/solid.py
import blinkypattern class Solid(blinkypattern.BlinkyPattern): def __init__(self, blinkytape, solid_color): super(Solid, self).__init__(blinkytape) self._pixels = [solid_color] * self._blinkytape.pixel_count def setup(self): super(Solid, self).setup() self._blinkytape.set_pixels(self._pixels) self._blinkytape.update()
class Solid(object): def __init__(self, pixel_count, color): self._pixels = [color] * pixel_count @property def pixels(self): return self._pixels
Update Solid pattern for refactor
Update Solid pattern for refactor
Python
mit
jonspeicher/blinkyfun
- import blinkypattern + class Solid(object): + def __init__(self, pixel_count, color): + self._pixels = [color] * pixel_count + @property + def pixels(self): + return self._pixels - class Solid(blinkypattern.BlinkyPattern): - def __init__(self, blinkytape, solid_color): - super(Solid, self).__init__(blinkytape) - self._pixels = [solid_color] * self._blinkytape.pixel_count - def setup(self): - super(Solid, self).setup() - self._blinkytape.set_pixels(self._pixels) - self._blinkytape.update() -
Update Solid pattern for refactor
## Code Before: import blinkypattern class Solid(blinkypattern.BlinkyPattern): def __init__(self, blinkytape, solid_color): super(Solid, self).__init__(blinkytape) self._pixels = [solid_color] * self._blinkytape.pixel_count def setup(self): super(Solid, self).setup() self._blinkytape.set_pixels(self._pixels) self._blinkytape.update() ## Instruction: Update Solid pattern for refactor ## Code After: class Solid(object): def __init__(self, pixel_count, color): self._pixels = [color] * pixel_count @property def pixels(self): return self._pixels
// ... existing code ... class Solid(object): def __init__(self, pixel_count, color): self._pixels = [color] * pixel_count @property def pixels(self): return self._pixels // ... rest of the code ...
b39518482da1d3e064cdbc34490e4a9924f6d5f1
quantecon/tests/test_ecdf.py
quantecon/tests/test_ecdf.py
import unittest import numpy as np from quantecon import ECDF class TestECDF(unittest.TestCase): @classmethod def setUpClass(cls): cls.obs = np.random.rand(40) # observations defining dist cls.ecdf = ECDF(cls.obs) def test_call_high(self): "ecdf: x above all obs give 1.0" # all of self.obs <= 1 so ecdf(1.1) should be 1 self.assertAlmostEqual(self.ecdf(1.1), 1.0) def test_call_low(self): "ecdf: x below all obs give 0.0" # all of self.obs <= 1 so ecdf(1.1) should be 1 self.assertAlmostEqual(self.ecdf(-0.1), 0.0) def test_ascending(self): "ecdf: larger values should return F(x) at least as big" x = np.random.rand() F_1 = self.ecdf(x) F_2 = self.ecdf(1.1 * x) self.assertGreaterEqual(F_2, F_1)
import unittest import numpy as np from quantecon import ECDF class TestECDF(unittest.TestCase): @classmethod def setUpClass(cls): cls.obs = np.random.rand(40) # observations defining dist cls.ecdf = ECDF(cls.obs) def test_call_high(self): "ecdf: x above all obs give 1.0" # all of self.obs <= 1 so ecdf(1.1) should be 1 self.assertAlmostEqual(self.ecdf(1.1), 1.0) def test_call_low(self): "ecdf: x below all obs give 0.0" # all of self.obs <= 1 so ecdf(1.1) should be 1 self.assertAlmostEqual(self.ecdf(-0.1), 0.0) def test_ascending(self): "ecdf: larger values should return F(x) at least as big" x = np.random.rand() F_1 = self.ecdf(x) F_2 = self.ecdf(1.1 * x) self.assertGreaterEqual(F_2, F_1) def test_vectorized(self): "ecdf: testing vectorized __call__ method" t = np.linspace(-1, 1, 100) self.assertEqual(t.shape, self.ecdf(t).shape) t = np.linspace(-1, 1, 100).reshape(2, 2, 25) self.assertEqual(t.shape, self.ecdf(t).shape)
Add a test for vectorized call
TST: Add a test for vectorized call
Python
bsd-3-clause
oyamad/QuantEcon.py,QuantEcon/QuantEcon.py,oyamad/QuantEcon.py,QuantEcon/QuantEcon.py
import unittest import numpy as np from quantecon import ECDF class TestECDF(unittest.TestCase): @classmethod def setUpClass(cls): cls.obs = np.random.rand(40) # observations defining dist cls.ecdf = ECDF(cls.obs) def test_call_high(self): "ecdf: x above all obs give 1.0" # all of self.obs <= 1 so ecdf(1.1) should be 1 self.assertAlmostEqual(self.ecdf(1.1), 1.0) def test_call_low(self): "ecdf: x below all obs give 0.0" # all of self.obs <= 1 so ecdf(1.1) should be 1 self.assertAlmostEqual(self.ecdf(-0.1), 0.0) def test_ascending(self): "ecdf: larger values should return F(x) at least as big" x = np.random.rand() F_1 = self.ecdf(x) F_2 = self.ecdf(1.1 * x) self.assertGreaterEqual(F_2, F_1) + def test_vectorized(self): + "ecdf: testing vectorized __call__ method" + t = np.linspace(-1, 1, 100) + self.assertEqual(t.shape, self.ecdf(t).shape) + t = np.linspace(-1, 1, 100).reshape(2, 2, 25) + self.assertEqual(t.shape, self.ecdf(t).shape) +
Add a test for vectorized call
## Code Before: import unittest import numpy as np from quantecon import ECDF class TestECDF(unittest.TestCase): @classmethod def setUpClass(cls): cls.obs = np.random.rand(40) # observations defining dist cls.ecdf = ECDF(cls.obs) def test_call_high(self): "ecdf: x above all obs give 1.0" # all of self.obs <= 1 so ecdf(1.1) should be 1 self.assertAlmostEqual(self.ecdf(1.1), 1.0) def test_call_low(self): "ecdf: x below all obs give 0.0" # all of self.obs <= 1 so ecdf(1.1) should be 1 self.assertAlmostEqual(self.ecdf(-0.1), 0.0) def test_ascending(self): "ecdf: larger values should return F(x) at least as big" x = np.random.rand() F_1 = self.ecdf(x) F_2 = self.ecdf(1.1 * x) self.assertGreaterEqual(F_2, F_1) ## Instruction: Add a test for vectorized call ## Code After: import unittest import numpy as np from quantecon import ECDF class TestECDF(unittest.TestCase): @classmethod def setUpClass(cls): cls.obs = np.random.rand(40) # observations defining dist cls.ecdf = ECDF(cls.obs) def test_call_high(self): "ecdf: x above all obs give 1.0" # all of self.obs <= 1 so ecdf(1.1) should be 1 self.assertAlmostEqual(self.ecdf(1.1), 1.0) def test_call_low(self): "ecdf: x below all obs give 0.0" # all of self.obs <= 1 so ecdf(1.1) should be 1 self.assertAlmostEqual(self.ecdf(-0.1), 0.0) def test_ascending(self): "ecdf: larger values should return F(x) at least as big" x = np.random.rand() F_1 = self.ecdf(x) F_2 = self.ecdf(1.1 * x) self.assertGreaterEqual(F_2, F_1) def test_vectorized(self): "ecdf: testing vectorized __call__ method" t = np.linspace(-1, 1, 100) self.assertEqual(t.shape, self.ecdf(t).shape) t = np.linspace(-1, 1, 100).reshape(2, 2, 25) self.assertEqual(t.shape, self.ecdf(t).shape)
// ... existing code ... self.assertGreaterEqual(F_2, F_1) def test_vectorized(self): "ecdf: testing vectorized __call__ method" t = np.linspace(-1, 1, 100) self.assertEqual(t.shape, self.ecdf(t).shape) t = np.linspace(-1, 1, 100).reshape(2, 2, 25) self.assertEqual(t.shape, self.ecdf(t).shape) // ... rest of the code ...
2cdb6a5eeb1730627cea2a812d590efed82d03fb
acceptance_tests/test_course_learners.py
acceptance_tests/test_course_learners.py
from unittest import skipUnless from bok_choy.web_app_test import WebAppTest from acceptance_tests import ENABLE_LEARNER_ANALYTICS from acceptance_tests.mixins import CoursePageTestsMixin from acceptance_tests.pages import CourseLearnersPage @skipUnless(ENABLE_LEARNER_ANALYTICS, 'Learner Analytics must be enabled to run CourseLearnersTests') class CourseLearnersTests(CoursePageTestsMixin, WebAppTest): def setUp(self): super(CourseLearnersTests, self).setUp() self.page = CourseLearnersPage(self.browser) def _test_data_update_message(self): # Don't test the update message for now, since it won't exist # until the SPA adds it to the page in AN-6205. pass def _get_data_update_message(self): # Don't test the update message for now, since it won't exist # until the SPA adds it to the page in AN-6205. return ''
from unittest import skipUnless from bok_choy.web_app_test import WebAppTest from acceptance_tests import ENABLE_LEARNER_ANALYTICS from acceptance_tests.mixins import CoursePageTestsMixin from acceptance_tests.pages import CourseLearnersPage @skipUnless(ENABLE_LEARNER_ANALYTICS, 'Learner Analytics must be enabled to run CourseLearnersTests') class CourseLearnersTests(CoursePageTestsMixin, WebAppTest): help_path = 'engagement/learners.html' def setUp(self): super(CourseLearnersTests, self).setUp() self.page = CourseLearnersPage(self.browser) def _test_data_update_message(self): # Don't test the update message for now, since it won't exist # until the SPA adds it to the page in AN-6205. pass def _get_data_update_message(self): # Don't test the update message for now, since it won't exist # until the SPA adds it to the page in AN-6205. return ''
Add test for learners help link
Add test for learners help link
Python
agpl-3.0
Stanford-Online/edx-analytics-dashboard,Stanford-Online/edx-analytics-dashboard,edx/edx-analytics-dashboard,Stanford-Online/edx-analytics-dashboard,edx/edx-analytics-dashboard,edx/edx-analytics-dashboard,Stanford-Online/edx-analytics-dashboard,edx/edx-analytics-dashboard
from unittest import skipUnless from bok_choy.web_app_test import WebAppTest from acceptance_tests import ENABLE_LEARNER_ANALYTICS from acceptance_tests.mixins import CoursePageTestsMixin from acceptance_tests.pages import CourseLearnersPage @skipUnless(ENABLE_LEARNER_ANALYTICS, 'Learner Analytics must be enabled to run CourseLearnersTests') class CourseLearnersTests(CoursePageTestsMixin, WebAppTest): + help_path = 'engagement/learners.html' + def setUp(self): super(CourseLearnersTests, self).setUp() self.page = CourseLearnersPage(self.browser) def _test_data_update_message(self): # Don't test the update message for now, since it won't exist # until the SPA adds it to the page in AN-6205. pass def _get_data_update_message(self): # Don't test the update message for now, since it won't exist # until the SPA adds it to the page in AN-6205. return ''
Add test for learners help link
## Code Before: from unittest import skipUnless from bok_choy.web_app_test import WebAppTest from acceptance_tests import ENABLE_LEARNER_ANALYTICS from acceptance_tests.mixins import CoursePageTestsMixin from acceptance_tests.pages import CourseLearnersPage @skipUnless(ENABLE_LEARNER_ANALYTICS, 'Learner Analytics must be enabled to run CourseLearnersTests') class CourseLearnersTests(CoursePageTestsMixin, WebAppTest): def setUp(self): super(CourseLearnersTests, self).setUp() self.page = CourseLearnersPage(self.browser) def _test_data_update_message(self): # Don't test the update message for now, since it won't exist # until the SPA adds it to the page in AN-6205. pass def _get_data_update_message(self): # Don't test the update message for now, since it won't exist # until the SPA adds it to the page in AN-6205. return '' ## Instruction: Add test for learners help link ## Code After: from unittest import skipUnless from bok_choy.web_app_test import WebAppTest from acceptance_tests import ENABLE_LEARNER_ANALYTICS from acceptance_tests.mixins import CoursePageTestsMixin from acceptance_tests.pages import CourseLearnersPage @skipUnless(ENABLE_LEARNER_ANALYTICS, 'Learner Analytics must be enabled to run CourseLearnersTests') class CourseLearnersTests(CoursePageTestsMixin, WebAppTest): help_path = 'engagement/learners.html' def setUp(self): super(CourseLearnersTests, self).setUp() self.page = CourseLearnersPage(self.browser) def _test_data_update_message(self): # Don't test the update message for now, since it won't exist # until the SPA adds it to the page in AN-6205. pass def _get_data_update_message(self): # Don't test the update message for now, since it won't exist # until the SPA adds it to the page in AN-6205. return ''
// ... existing code ... class CourseLearnersTests(CoursePageTestsMixin, WebAppTest): help_path = 'engagement/learners.html' def setUp(self): // ... rest of the code ...
3ecc978421e1bcceb30635e875333e52272e07a3
tests/providers/test_ovh.py
tests/providers/test_ovh.py
from unittest import TestCase from lexicon.providers.ovh import Provider from lexicon.common.options_handler import env_auth_options from integration_tests import IntegrationTests # Hook into testing framework by inheriting unittest.TestCase and reuse # the tests which *each and every* implementation of the interface must # pass, by inheritance from integration_tests.IntegrationTests class OvhProviderTests(TestCase, IntegrationTests): Provider = Provider provider_name = 'ovh' domain = 'elogium.net' def _filter_headers(self): return ['X-Ovh-Application', 'X-Ovh-Consumer', 'X-Ovh-Signature'] # Override _test_options to call env_auth_options and then import auth config from env variables def _test_options(self): cmd_options = env_auth_options(self.provider_name) cmd_options['domain'] = self.domain return cmd_options
from unittest import TestCase from lexicon.providers.ovh import Provider from lexicon.common.options_handler import env_auth_options from integration_tests import IntegrationTests # Hook into testing framework by inheriting unittest.TestCase and reuse # the tests which *each and every* implementation of the interface must # pass, by inheritance from integration_tests.IntegrationTests class OvhProviderTests(TestCase, IntegrationTests): Provider = Provider provider_name = 'ovh' domain = 'elogium.net' def _filter_headers(self): return ['X-Ovh-Application', 'X-Ovh-Consumer', 'X-Ovh-Signature'] # Override _test_options to call env_auth_options and then import auth config from env variables def _test_options(self): cmd_options = env_auth_options(self.provider_name) cmd_options['auth_entrypoint'] = 'ovh-eu' cmd_options['domain'] = self.domain return cmd_options
Select ovh-eu entrypoint for test integration
Select ovh-eu entrypoint for test integration
Python
mit
tnwhitwell/lexicon,AnalogJ/lexicon,AnalogJ/lexicon,tnwhitwell/lexicon
from unittest import TestCase from lexicon.providers.ovh import Provider from lexicon.common.options_handler import env_auth_options from integration_tests import IntegrationTests # Hook into testing framework by inheriting unittest.TestCase and reuse # the tests which *each and every* implementation of the interface must # pass, by inheritance from integration_tests.IntegrationTests class OvhProviderTests(TestCase, IntegrationTests): Provider = Provider provider_name = 'ovh' domain = 'elogium.net' def _filter_headers(self): return ['X-Ovh-Application', 'X-Ovh-Consumer', 'X-Ovh-Signature'] # Override _test_options to call env_auth_options and then import auth config from env variables def _test_options(self): cmd_options = env_auth_options(self.provider_name) + cmd_options['auth_entrypoint'] = 'ovh-eu' cmd_options['domain'] = self.domain return cmd_options
Select ovh-eu entrypoint for test integration
## Code Before: from unittest import TestCase from lexicon.providers.ovh import Provider from lexicon.common.options_handler import env_auth_options from integration_tests import IntegrationTests # Hook into testing framework by inheriting unittest.TestCase and reuse # the tests which *each and every* implementation of the interface must # pass, by inheritance from integration_tests.IntegrationTests class OvhProviderTests(TestCase, IntegrationTests): Provider = Provider provider_name = 'ovh' domain = 'elogium.net' def _filter_headers(self): return ['X-Ovh-Application', 'X-Ovh-Consumer', 'X-Ovh-Signature'] # Override _test_options to call env_auth_options and then import auth config from env variables def _test_options(self): cmd_options = env_auth_options(self.provider_name) cmd_options['domain'] = self.domain return cmd_options ## Instruction: Select ovh-eu entrypoint for test integration ## Code After: from unittest import TestCase from lexicon.providers.ovh import Provider from lexicon.common.options_handler import env_auth_options from integration_tests import IntegrationTests # Hook into testing framework by inheriting unittest.TestCase and reuse # the tests which *each and every* implementation of the interface must # pass, by inheritance from integration_tests.IntegrationTests class OvhProviderTests(TestCase, IntegrationTests): Provider = Provider provider_name = 'ovh' domain = 'elogium.net' def _filter_headers(self): return ['X-Ovh-Application', 'X-Ovh-Consumer', 'X-Ovh-Signature'] # Override _test_options to call env_auth_options and then import auth config from env variables def _test_options(self): cmd_options = env_auth_options(self.provider_name) cmd_options['auth_entrypoint'] = 'ovh-eu' cmd_options['domain'] = self.domain return cmd_options
# ... existing code ... cmd_options = env_auth_options(self.provider_name) cmd_options['auth_entrypoint'] = 'ovh-eu' cmd_options['domain'] = self.domain # ... rest of the code ...
9121c8c074a31fd3668f8281c7f093360ed72988
salad/cli.py
salad/cli.py
import sys import argparse from lettuce.bin import main as lettuce_main from lettuce import world from selenium.webdriver.common.desired_capabilities import DesiredCapabilities BROWSER_CHOICES = [browser.lower() for browser in DesiredCapabilities.__dict__.keys() if not browser.startswith('_')] BROWSER_CHOICES.append('zope.testbrowser') BROWSER_CHOICES.sort() DEFAULT_BROWSER = 'firefox' def main(args=sys.argv[1:]): parser = argparse.ArgumentParser(prog="Salad", description='BDD browswer-automation made tasty.') parser.add_argument('--browser', default=DEFAULT_BROWSER, metavar='BROWSER', choices=BROWSER_CHOICES, help=('Browser to use. Options: %s Default is %s.' % (BROWSER_CHOICES, DEFAULT_BROWSER))) parser.add_argument('--remote-url', help='Selenium server url for remote browsers') parser.add_argument('args', nargs=argparse.REMAINDER) parsed_args = parser.parse_args() world.drivers = [parsed_args.browser] world.remote_url = parsed_args.remote_url lettuce_main(args=parsed_args.args) if __name__ == '__main__': main()
import sys import argparse from lettuce.bin import main as lettuce_main from lettuce import world from selenium.webdriver.common.desired_capabilities import DesiredCapabilities BROWSER_CHOICES = [browser.lower() for browser in DesiredCapabilities.__dict__.keys() if not browser.startswith('_')] BROWSER_CHOICES.append('zope.testbrowser') BROWSER_CHOICES.sort() DEFAULT_BROWSER = 'firefox' def main(args=sys.argv[1:]): parser = argparse.ArgumentParser(prog="Salad", description='BDD browswer-automation made tasty.') parser.add_argument('--browser', default=DEFAULT_BROWSER, metavar='BROWSER', choices=BROWSER_CHOICES, help=('Browser to use. Options: %s Default is %s.' % (BROWSER_CHOICES, DEFAULT_BROWSER))) parser.add_argument('--remote-url', help='Selenium server url for remote browsers') (parsed_args, leftovers) = parser.parse_known_args() world.drivers = [parsed_args.browser] world.remote_url = parsed_args.remote_url lettuce_main(args=leftovers) if __name__ == '__main__': main()
Use parse_known_args and pass leftovers to lettuce
Use parse_known_args and pass leftovers to lettuce Seems to be more reliable at handling weird argument ordering than REMAINDER was
Python
bsd-3-clause
salad/salad,salad/salad,beanqueen/salad,beanqueen/salad
import sys import argparse from lettuce.bin import main as lettuce_main from lettuce import world from selenium.webdriver.common.desired_capabilities import DesiredCapabilities BROWSER_CHOICES = [browser.lower() for browser in DesiredCapabilities.__dict__.keys() if not browser.startswith('_')] BROWSER_CHOICES.append('zope.testbrowser') BROWSER_CHOICES.sort() DEFAULT_BROWSER = 'firefox' def main(args=sys.argv[1:]): parser = argparse.ArgumentParser(prog="Salad", description='BDD browswer-automation made tasty.') parser.add_argument('--browser', default=DEFAULT_BROWSER, metavar='BROWSER', choices=BROWSER_CHOICES, help=('Browser to use. Options: %s Default is %s.' % (BROWSER_CHOICES, DEFAULT_BROWSER))) parser.add_argument('--remote-url', help='Selenium server url for remote browsers') - parser.add_argument('args', nargs=argparse.REMAINDER) - parsed_args = parser.parse_args() + (parsed_args, leftovers) = parser.parse_known_args() world.drivers = [parsed_args.browser] world.remote_url = parsed_args.remote_url - lettuce_main(args=parsed_args.args) + lettuce_main(args=leftovers) if __name__ == '__main__': main()
Use parse_known_args and pass leftovers to lettuce
## Code Before: import sys import argparse from lettuce.bin import main as lettuce_main from lettuce import world from selenium.webdriver.common.desired_capabilities import DesiredCapabilities BROWSER_CHOICES = [browser.lower() for browser in DesiredCapabilities.__dict__.keys() if not browser.startswith('_')] BROWSER_CHOICES.append('zope.testbrowser') BROWSER_CHOICES.sort() DEFAULT_BROWSER = 'firefox' def main(args=sys.argv[1:]): parser = argparse.ArgumentParser(prog="Salad", description='BDD browswer-automation made tasty.') parser.add_argument('--browser', default=DEFAULT_BROWSER, metavar='BROWSER', choices=BROWSER_CHOICES, help=('Browser to use. Options: %s Default is %s.' % (BROWSER_CHOICES, DEFAULT_BROWSER))) parser.add_argument('--remote-url', help='Selenium server url for remote browsers') parser.add_argument('args', nargs=argparse.REMAINDER) parsed_args = parser.parse_args() world.drivers = [parsed_args.browser] world.remote_url = parsed_args.remote_url lettuce_main(args=parsed_args.args) if __name__ == '__main__': main() ## Instruction: Use parse_known_args and pass leftovers to lettuce ## Code After: import sys import argparse from lettuce.bin import main as lettuce_main from lettuce import world from selenium.webdriver.common.desired_capabilities import DesiredCapabilities BROWSER_CHOICES = [browser.lower() for browser in DesiredCapabilities.__dict__.keys() if not browser.startswith('_')] BROWSER_CHOICES.append('zope.testbrowser') BROWSER_CHOICES.sort() DEFAULT_BROWSER = 'firefox' def main(args=sys.argv[1:]): parser = argparse.ArgumentParser(prog="Salad", description='BDD browswer-automation made tasty.') parser.add_argument('--browser', default=DEFAULT_BROWSER, metavar='BROWSER', choices=BROWSER_CHOICES, help=('Browser to use. Options: %s Default is %s.' % (BROWSER_CHOICES, DEFAULT_BROWSER))) parser.add_argument('--remote-url', help='Selenium server url for remote browsers') (parsed_args, leftovers) = parser.parse_known_args() world.drivers = [parsed_args.browser] world.remote_url = parsed_args.remote_url lettuce_main(args=leftovers) if __name__ == '__main__': main()
# ... existing code ... help='Selenium server url for remote browsers') (parsed_args, leftovers) = parser.parse_known_args() world.drivers = [parsed_args.browser] # ... modified code ... world.remote_url = parsed_args.remote_url lettuce_main(args=leftovers) # ... rest of the code ...
fc4fa5d06ea0ca557d69112d1c8d0f10c8e594e0
diet_gtfs.py
diet_gtfs.py
import csv import sys # agency.txt done # feed_info.txt nothing to change # calendar_dates.txt depends on service_id. # routes.txt depends on agency.txt # shapes.txt depends on trips.txt # stops.txt depends on stop_times.txt # stop_times.txt depends on trip_id. # transfers.txt depends on stop_id from and to, routes. # trips.txt contains shape_id, also route_id to trip_id. def clean_agency_file(*agencies): with open('agency.txt', 'r') as f: reader = csv.reader(f) next(f) for row in reader: if row[0] in agencies: print(row) def main(): agencies = sys.argv[1:] clean_agency_file(*agencies) if __name__ == '__main__': main()
import csv import sys # agency.txt done # feed_info.txt nothing to change # calendar_dates.txt depends on service_id. # routes.txt depends on agency.txt # shapes.txt depends on trips.txt # stops.txt depends on stop_times.txt # stop_times.txt depends on trip_id. # transfers.txt depends on stop_id from and to, routes. # trips.txt contains shape_id, also route_id to trip_id. def clean_agency_file(*agencies): with open('agency.txt', 'r') as f: reader = csv.reader(f) filtered_rows = [] filtered_rows.append(next(reader)) for row in reader: if row[0] in agencies: filtered_rows.append(row) with open('cleaned/agency.txt', 'w') as f: writer = csv.writer(f) writer.writerows(filtered_rows) def main(): agencies = sys.argv[1:] clean_agency_file(*agencies) if __name__ == '__main__': main()
Create a complete filtered output agency.txt
Create a complete filtered output agency.txt Filter based on arguments passed from shell.
Python
bsd-2-clause
sensiblecodeio/diet-gtfs
import csv import sys # agency.txt done # feed_info.txt nothing to change # calendar_dates.txt depends on service_id. # routes.txt depends on agency.txt # shapes.txt depends on trips.txt # stops.txt depends on stop_times.txt # stop_times.txt depends on trip_id. # transfers.txt depends on stop_id from and to, routes. # trips.txt contains shape_id, also route_id to trip_id. + def clean_agency_file(*agencies): with open('agency.txt', 'r') as f: reader = csv.reader(f) - next(f) + filtered_rows = [] + filtered_rows.append(next(reader)) + for row in reader: if row[0] in agencies: - print(row) + filtered_rows.append(row) + + with open('cleaned/agency.txt', 'w') as f: + writer = csv.writer(f) + writer.writerows(filtered_rows) def main(): agencies = sys.argv[1:] clean_agency_file(*agencies) if __name__ == '__main__': main()
Create a complete filtered output agency.txt
## Code Before: import csv import sys # agency.txt done # feed_info.txt nothing to change # calendar_dates.txt depends on service_id. # routes.txt depends on agency.txt # shapes.txt depends on trips.txt # stops.txt depends on stop_times.txt # stop_times.txt depends on trip_id. # transfers.txt depends on stop_id from and to, routes. # trips.txt contains shape_id, also route_id to trip_id. def clean_agency_file(*agencies): with open('agency.txt', 'r') as f: reader = csv.reader(f) next(f) for row in reader: if row[0] in agencies: print(row) def main(): agencies = sys.argv[1:] clean_agency_file(*agencies) if __name__ == '__main__': main() ## Instruction: Create a complete filtered output agency.txt ## Code After: import csv import sys # agency.txt done # feed_info.txt nothing to change # calendar_dates.txt depends on service_id. # routes.txt depends on agency.txt # shapes.txt depends on trips.txt # stops.txt depends on stop_times.txt # stop_times.txt depends on trip_id. # transfers.txt depends on stop_id from and to, routes. # trips.txt contains shape_id, also route_id to trip_id. def clean_agency_file(*agencies): with open('agency.txt', 'r') as f: reader = csv.reader(f) filtered_rows = [] filtered_rows.append(next(reader)) for row in reader: if row[0] in agencies: filtered_rows.append(row) with open('cleaned/agency.txt', 'w') as f: writer = csv.writer(f) writer.writerows(filtered_rows) def main(): agencies = sys.argv[1:] clean_agency_file(*agencies) if __name__ == '__main__': main()
// ... existing code ... def clean_agency_file(*agencies): // ... modified code ... reader = csv.reader(f) filtered_rows = [] filtered_rows.append(next(reader)) for row in reader: ... if row[0] in agencies: filtered_rows.append(row) with open('cleaned/agency.txt', 'w') as f: writer = csv.writer(f) writer.writerows(filtered_rows) // ... rest of the code ...
ee24b8b57bc73947cd5140aca15389861b33ab00
gui/qt.py
gui/qt.py
from lib.version import AMON_VERSION from lib.keybase import KeybaseUser from lib.gmail import GmailUser from lib.addresses import AddressBook import lib.gpg as gpg import sys import logging import json from PyQt4 import QtGui class Amon(QtGui.QMainWindow): def __init__(self): super(Amon, self).__init__() self.keybase_user = KeybaseUser() self.gmail = GmailUser() self.address_book = AddressBook()
from lib.version import AMON_VERSION from lib.keybase import KeybaseUser from lib.gmail import GmailUser from lib.addresses import AddressBook import lib.gpg as gpg import sys import logging import json from PyQt4 import QtGui class Amon(QtGui.QMainWindow): def __init__(self): super(Amon, self).__init__() self.keybase_user = KeybaseUser() self.gmail = GmailUser() self.address_book = AddressBook() self.initUI() def initUI(self): exitAction = QtGui.QAction(QtGui.QIcon('exit.png'), '&Exit', self) exitAction.setShortcut('Ctrl+Q') exitAction.setStatusTip('Exit application') exitAction.triggered.connect(QtGui.qApp.quit) menubar = self.menuBar() menubar.setNativeMenuBar(False) fileMenu = menubar.addMenu('&File') fileMenu.addAction(exitAction) self.statusBar().showMessage('Ready') self.setGeometry(300, 300, 300, 200) self.setWindowTitle('Amon ' + AMON_VERSION) self.show() def main(): app = QtGui.QApplication(sys.argv) amon = Amon() sys.exit(app.exec_()) if __name__ == '__main__': main()
Update Qt gui to have status bar and menu bar
Update Qt gui to have status bar and menu bar
Python
unlicense
CodingAnarchy/Amon
from lib.version import AMON_VERSION from lib.keybase import KeybaseUser from lib.gmail import GmailUser from lib.addresses import AddressBook import lib.gpg as gpg import sys import logging import json from PyQt4 import QtGui class Amon(QtGui.QMainWindow): def __init__(self): super(Amon, self).__init__() self.keybase_user = KeybaseUser() self.gmail = GmailUser() self.address_book = AddressBook() + self.initUI() + + def initUI(self): + exitAction = QtGui.QAction(QtGui.QIcon('exit.png'), '&Exit', self) + exitAction.setShortcut('Ctrl+Q') + exitAction.setStatusTip('Exit application') + exitAction.triggered.connect(QtGui.qApp.quit) + + menubar = self.menuBar() + menubar.setNativeMenuBar(False) + fileMenu = menubar.addMenu('&File') + fileMenu.addAction(exitAction) + + self.statusBar().showMessage('Ready') + self.setGeometry(300, 300, 300, 200) + self.setWindowTitle('Amon ' + AMON_VERSION) + self.show() + + + def main(): + app = QtGui.QApplication(sys.argv) + amon = Amon() + sys.exit(app.exec_()) + + if __name__ == '__main__': + main() +
Update Qt gui to have status bar and menu bar
## Code Before: from lib.version import AMON_VERSION from lib.keybase import KeybaseUser from lib.gmail import GmailUser from lib.addresses import AddressBook import lib.gpg as gpg import sys import logging import json from PyQt4 import QtGui class Amon(QtGui.QMainWindow): def __init__(self): super(Amon, self).__init__() self.keybase_user = KeybaseUser() self.gmail = GmailUser() self.address_book = AddressBook() ## Instruction: Update Qt gui to have status bar and menu bar ## Code After: from lib.version import AMON_VERSION from lib.keybase import KeybaseUser from lib.gmail import GmailUser from lib.addresses import AddressBook import lib.gpg as gpg import sys import logging import json from PyQt4 import QtGui class Amon(QtGui.QMainWindow): def __init__(self): super(Amon, self).__init__() self.keybase_user = KeybaseUser() self.gmail = GmailUser() self.address_book = AddressBook() self.initUI() def initUI(self): exitAction = QtGui.QAction(QtGui.QIcon('exit.png'), '&Exit', self) exitAction.setShortcut('Ctrl+Q') exitAction.setStatusTip('Exit application') exitAction.triggered.connect(QtGui.qApp.quit) menubar = self.menuBar() menubar.setNativeMenuBar(False) fileMenu = menubar.addMenu('&File') fileMenu.addAction(exitAction) self.statusBar().showMessage('Ready') self.setGeometry(300, 300, 300, 200) self.setWindowTitle('Amon ' + AMON_VERSION) self.show() def main(): app = QtGui.QApplication(sys.argv) amon = Amon() sys.exit(app.exec_()) if __name__ == '__main__': main()
# ... existing code ... self.address_book = AddressBook() self.initUI() def initUI(self): exitAction = QtGui.QAction(QtGui.QIcon('exit.png'), '&Exit', self) exitAction.setShortcut('Ctrl+Q') exitAction.setStatusTip('Exit application') exitAction.triggered.connect(QtGui.qApp.quit) menubar = self.menuBar() menubar.setNativeMenuBar(False) fileMenu = menubar.addMenu('&File') fileMenu.addAction(exitAction) self.statusBar().showMessage('Ready') self.setGeometry(300, 300, 300, 200) self.setWindowTitle('Amon ' + AMON_VERSION) self.show() def main(): app = QtGui.QApplication(sys.argv) amon = Amon() sys.exit(app.exec_()) if __name__ == '__main__': main() # ... rest of the code ...
92d9e9885e241e0bb7df64d3cd696db09cdfc74d
utils.py
utils.py
def fix_str(value): try: return unicode(value) except UnicodeDecodeError: return unicode(value.decode('latin1')) def pandas_to_dict(df): return [{colname: (fix_str(row[i]) if type(row[i]) is str else row[i]) for i, colname in enumerate(df.columns) if colname not in ['pedido_data', 'cliente_data']} for row in df.values]
def fix_str(value): try: return unicode(value) except UnicodeDecodeError: return unicode(value.decode('latin1')) def pandas_to_dict(df): return [{colname: (fix_str(row[i]) if type(row[i]) is str else row[i]) for i, colname in enumerate(df.columns)} for row in df.values]
Remove columns filter on pandas_to_dict
Remove columns filter on pandas_to_dict
Python
mit
mlgruby/mining,mlgruby/mining,mining/mining,chrisdamba/mining,jgabriellima/mining,chrisdamba/mining,AndrzejR/mining,seagoat/mining,jgabriellima/mining,mining/mining,avelino/mining,AndrzejR/mining,seagoat/mining,mlgruby/mining,avelino/mining
def fix_str(value): try: return unicode(value) except UnicodeDecodeError: return unicode(value.decode('latin1')) def pandas_to_dict(df): return [{colname: (fix_str(row[i]) if type(row[i]) is str else row[i]) - for i, colname in enumerate(df.columns) + for i, colname in enumerate(df.columns)} - if colname not in ['pedido_data', 'cliente_data']} for row in df.values]
Remove columns filter on pandas_to_dict
## Code Before: def fix_str(value): try: return unicode(value) except UnicodeDecodeError: return unicode(value.decode('latin1')) def pandas_to_dict(df): return [{colname: (fix_str(row[i]) if type(row[i]) is str else row[i]) for i, colname in enumerate(df.columns) if colname not in ['pedido_data', 'cliente_data']} for row in df.values] ## Instruction: Remove columns filter on pandas_to_dict ## Code After: def fix_str(value): try: return unicode(value) except UnicodeDecodeError: return unicode(value.decode('latin1')) def pandas_to_dict(df): return [{colname: (fix_str(row[i]) if type(row[i]) is str else row[i]) for i, colname in enumerate(df.columns)} for row in df.values]
# ... existing code ... return [{colname: (fix_str(row[i]) if type(row[i]) is str else row[i]) for i, colname in enumerate(df.columns)} for row in df.values] # ... rest of the code ...
22f9fc8a56882f0595d051cb8c5d20fd97091e8c
custom/opm/tests/test_snapshot.py
custom/opm/tests/test_snapshot.py
from datetime import date from unittest import TestCase from couchforms.models import XFormInstance from ..constants import * from ..reports import get_report, BeneficiaryPaymentReport, MetReport from .case_reports import Report, OPMCase, MockCaseRow, MockDataProvider class TestGetReportUtil(TestCase): def get_report_class(self, report_class): obj_dict = { 'get_rows': lambda slf, datespan: [ OPMCase( forms=[], edd=date(2014, 11, 10), ), OPMCase( forms=[], dod=date(2014, 1, 12), ), OPMCase( forms=[], dod=date(2014, 3, 12), ), ], 'data_provider': MockDataProvider(), } return type(report_class.__name__, (Report, report_class), obj_dict) def test_basic_BPR(self): report_class = self.get_report_class(BeneficiaryPaymentReport) report = get_report(report_class, month=6, year=2014, block="Atri") report.rows def test_basic_CMR(self): report_class = self.get_report_class(MetReport) report = get_report(report_class, month=6, year=2014, block="Atri") report.rows
from datetime import date from unittest import TestCase from mock import patch from corehq.apps.users.models import CommCareUser from couchforms.models import XFormInstance from ..constants import * from ..reports import get_report, BeneficiaryPaymentReport, MetReport from .case_reports import Report, OPMCase, MockCaseRow, MockDataProvider class TestGetReportUtil(TestCase): def get_report_class(self, report_class): obj_dict = { 'get_rows': lambda slf, datespan: [ OPMCase( forms=[], edd=date(2014, 11, 10), ), OPMCase( forms=[], dod=date(2014, 1, 12), ), OPMCase( forms=[], dod=date(2014, 3, 12), ), ], 'data_provider': MockDataProvider(), } return type(report_class.__name__, (Report, report_class), obj_dict) def test_basic_BPR(self): report_class = self.get_report_class(BeneficiaryPaymentReport) report = get_report(report_class, month=6, year=2014, block="Atri") report.rows @patch.object(CommCareUser, 'by_domain', return_value=[]) def test_basic_CMR(self, user_mock): report_class = self.get_report_class(MetReport) report = get_report(report_class, month=6, year=2014, block="Atri") report.rows
Fix for test (add mock for CommCareUser)
Fix for test (add mock for CommCareUser)
Python
bsd-3-clause
puttarajubr/commcare-hq,puttarajubr/commcare-hq,qedsoftware/commcare-hq,qedsoftware/commcare-hq,qedsoftware/commcare-hq,qedsoftware/commcare-hq,dimagi/commcare-hq,puttarajubr/commcare-hq,dimagi/commcare-hq,qedsoftware/commcare-hq,dimagi/commcare-hq,puttarajubr/commcare-hq,dimagi/commcare-hq,dimagi/commcare-hq
from datetime import date from unittest import TestCase - + from mock import patch + from corehq.apps.users.models import CommCareUser from couchforms.models import XFormInstance from ..constants import * from ..reports import get_report, BeneficiaryPaymentReport, MetReport from .case_reports import Report, OPMCase, MockCaseRow, MockDataProvider class TestGetReportUtil(TestCase): def get_report_class(self, report_class): obj_dict = { 'get_rows': lambda slf, datespan: [ OPMCase( forms=[], edd=date(2014, 11, 10), ), OPMCase( forms=[], dod=date(2014, 1, 12), ), OPMCase( forms=[], dod=date(2014, 3, 12), ), ], 'data_provider': MockDataProvider(), } return type(report_class.__name__, (Report, report_class), obj_dict) def test_basic_BPR(self): report_class = self.get_report_class(BeneficiaryPaymentReport) report = get_report(report_class, month=6, year=2014, block="Atri") report.rows + @patch.object(CommCareUser, 'by_domain', return_value=[]) - def test_basic_CMR(self): + def test_basic_CMR(self, user_mock): report_class = self.get_report_class(MetReport) report = get_report(report_class, month=6, year=2014, block="Atri") report.rows
Fix for test (add mock for CommCareUser)
## Code Before: from datetime import date from unittest import TestCase from couchforms.models import XFormInstance from ..constants import * from ..reports import get_report, BeneficiaryPaymentReport, MetReport from .case_reports import Report, OPMCase, MockCaseRow, MockDataProvider class TestGetReportUtil(TestCase): def get_report_class(self, report_class): obj_dict = { 'get_rows': lambda slf, datespan: [ OPMCase( forms=[], edd=date(2014, 11, 10), ), OPMCase( forms=[], dod=date(2014, 1, 12), ), OPMCase( forms=[], dod=date(2014, 3, 12), ), ], 'data_provider': MockDataProvider(), } return type(report_class.__name__, (Report, report_class), obj_dict) def test_basic_BPR(self): report_class = self.get_report_class(BeneficiaryPaymentReport) report = get_report(report_class, month=6, year=2014, block="Atri") report.rows def test_basic_CMR(self): report_class = self.get_report_class(MetReport) report = get_report(report_class, month=6, year=2014, block="Atri") report.rows ## Instruction: Fix for test (add mock for CommCareUser) ## Code After: from datetime import date from unittest import TestCase from mock import patch from corehq.apps.users.models import CommCareUser from couchforms.models import XFormInstance from ..constants import * from ..reports import get_report, BeneficiaryPaymentReport, MetReport from .case_reports import Report, OPMCase, MockCaseRow, MockDataProvider class TestGetReportUtil(TestCase): def get_report_class(self, report_class): obj_dict = { 'get_rows': lambda slf, datespan: [ OPMCase( forms=[], edd=date(2014, 11, 10), ), OPMCase( forms=[], dod=date(2014, 1, 12), ), OPMCase( forms=[], dod=date(2014, 3, 12), ), ], 'data_provider': MockDataProvider(), } return type(report_class.__name__, (Report, report_class), obj_dict) def test_basic_BPR(self): report_class = self.get_report_class(BeneficiaryPaymentReport) report = get_report(report_class, month=6, year=2014, block="Atri") report.rows @patch.object(CommCareUser, 'by_domain', return_value=[]) def test_basic_CMR(self, user_mock): report_class = self.get_report_class(MetReport) report = get_report(report_class, month=6, year=2014, block="Atri") report.rows
# ... existing code ... from unittest import TestCase from mock import patch from corehq.apps.users.models import CommCareUser from couchforms.models import XFormInstance # ... modified code ... @patch.object(CommCareUser, 'by_domain', return_value=[]) def test_basic_CMR(self, user_mock): report_class = self.get_report_class(MetReport) # ... rest of the code ...
06f10e09f5b1c5766815b6e7eb219b4e33082709
check_urls.py
check_urls.py
import re, sys, markdown, requests, bs4 as BeautifulSoup reload(sys) sys.setdefaultencoding('utf8') def check_url(url): try: return bool(requests.head(url, allow_redirects=True)) except Exception as e: print 'Error checking URL %s: %s' % (url, e) return False def retrieve_urls(filename): with open(filename) as fd: mdtext = fd.read() html_text = markdown.markdown(mdtext) soup = BeautifulSoup.BeautifulSoup(html_text, "html.parser") return [a['href'] for a in soup.findAll('a')] def check_urls(filename): print 'checking URLs for %s' % (filename,) ok = True for url in retrieve_urls(filename): r = "(?:http[s]?://[^)]+)" u = re.findall(r, url) if not u: continue msg = 'Checking %s => ' % (u[0],) if check_url(u[0]): print msg, 'OK' else: print msg, 'FAILED' ok = False return ok def main(): ok = True for filename in sys.argv[1:]: try: ok &= check_urls(filename) except IOError as e: print e ok = False exit (0 if ok else 1) if __name__ == '__main__': main()
from __future__ import print_function import re, sys, markdown, requests, bs4 as BeautifulSoup try: # Python 2 reload except NameError: # Python 3 from importlib import reload reload(sys) sys.setdefaultencoding('utf8') def check_url(url): try: return bool(requests.head(url, allow_redirects=True)) except Exception as e: print('Error checking URL %s: %s' % (url, e)) return False def retrieve_urls(filename): with open(filename) as fd: mdtext = fd.read() html_text = markdown.markdown(mdtext) soup = BeautifulSoup.BeautifulSoup(html_text, "html.parser") return [a['href'] for a in soup.findAll('a')] def check_urls(filename): print('checking URLs for %s' % filename) ok = True for url in retrieve_urls(filename): r = "(?:http[s]?://[^)]+)" u = re.findall(r, url) if not u: continue msg = 'Checking %s => ' % (u[0],) if check_url(u[0]): print(msg, 'OK') else: print(msg, 'FAILED') ok = False return ok def main(): ok = True for filename in sys.argv[1:]: try: ok &= check_urls(filename) except IOError as e: print(e) ok = False exit(0 if ok else 1) if __name__ == '__main__': main()
Add Python 3 compatibility and flake8 testing
Add Python 3 compatibility and flake8 testing
Python
unlicense
ligurio/free-software-testing-books
+ from __future__ import print_function import re, sys, markdown, requests, bs4 as BeautifulSoup + + try: # Python 2 + reload + except NameError: # Python 3 + from importlib import reload reload(sys) sys.setdefaultencoding('utf8') def check_url(url): try: return bool(requests.head(url, allow_redirects=True)) except Exception as e: - print 'Error checking URL %s: %s' % (url, e) + print('Error checking URL %s: %s' % (url, e)) return False def retrieve_urls(filename): with open(filename) as fd: mdtext = fd.read() html_text = markdown.markdown(mdtext) soup = BeautifulSoup.BeautifulSoup(html_text, "html.parser") return [a['href'] for a in soup.findAll('a')] def check_urls(filename): - print 'checking URLs for %s' % (filename,) + print('checking URLs for %s' % filename) ok = True for url in retrieve_urls(filename): r = "(?:http[s]?://[^)]+)" u = re.findall(r, url) if not u: continue msg = 'Checking %s => ' % (u[0],) if check_url(u[0]): - print msg, 'OK' + print(msg, 'OK') else: - print msg, 'FAILED' + print(msg, 'FAILED') ok = False return ok def main(): ok = True for filename in sys.argv[1:]: try: ok &= check_urls(filename) except IOError as e: - print e + print(e) ok = False - exit (0 if ok else 1) + exit(0 if ok else 1) if __name__ == '__main__': main()
Add Python 3 compatibility and flake8 testing
## Code Before: import re, sys, markdown, requests, bs4 as BeautifulSoup reload(sys) sys.setdefaultencoding('utf8') def check_url(url): try: return bool(requests.head(url, allow_redirects=True)) except Exception as e: print 'Error checking URL %s: %s' % (url, e) return False def retrieve_urls(filename): with open(filename) as fd: mdtext = fd.read() html_text = markdown.markdown(mdtext) soup = BeautifulSoup.BeautifulSoup(html_text, "html.parser") return [a['href'] for a in soup.findAll('a')] def check_urls(filename): print 'checking URLs for %s' % (filename,) ok = True for url in retrieve_urls(filename): r = "(?:http[s]?://[^)]+)" u = re.findall(r, url) if not u: continue msg = 'Checking %s => ' % (u[0],) if check_url(u[0]): print msg, 'OK' else: print msg, 'FAILED' ok = False return ok def main(): ok = True for filename in sys.argv[1:]: try: ok &= check_urls(filename) except IOError as e: print e ok = False exit (0 if ok else 1) if __name__ == '__main__': main() ## Instruction: Add Python 3 compatibility and flake8 testing ## Code After: from __future__ import print_function import re, sys, markdown, requests, bs4 as BeautifulSoup try: # Python 2 reload except NameError: # Python 3 from importlib import reload reload(sys) sys.setdefaultencoding('utf8') def check_url(url): try: return bool(requests.head(url, allow_redirects=True)) except Exception as e: print('Error checking URL %s: %s' % (url, e)) return False def retrieve_urls(filename): with open(filename) as fd: mdtext = fd.read() html_text = markdown.markdown(mdtext) soup = BeautifulSoup.BeautifulSoup(html_text, "html.parser") return [a['href'] for a in soup.findAll('a')] def check_urls(filename): print('checking URLs for %s' % filename) ok = True for url in retrieve_urls(filename): r = "(?:http[s]?://[^)]+)" u = re.findall(r, url) if not u: continue msg = 'Checking %s => ' % (u[0],) if check_url(u[0]): print(msg, 'OK') else: print(msg, 'FAILED') ok = False return ok def main(): ok = True for filename in sys.argv[1:]: try: ok &= check_urls(filename) except IOError as e: print(e) ok = False exit(0 if ok else 1) if __name__ == '__main__': main()
// ... existing code ... from __future__ import print_function import re, sys, markdown, requests, bs4 as BeautifulSoup try: # Python 2 reload except NameError: # Python 3 from importlib import reload // ... modified code ... except Exception as e: print('Error checking URL %s: %s' % (url, e)) return False ... def check_urls(filename): print('checking URLs for %s' % filename) ok = True ... if check_url(u[0]): print(msg, 'OK') else: print(msg, 'FAILED') ok = False ... except IOError as e: print(e) ok = False exit(0 if ok else 1) // ... rest of the code ...
6d8b99b5e4dab49c5a2e90b07f02072c116a7367
robots/models.py
robots/models.py
from django.db import models from django.contrib.sites.models import Site from django.utils.translation import ugettext_lazy as _ class File(models.Model): site = models.OneToOneField(Site, verbose_name=_(u'site')) content = models.TextField(_(u'file content')) objects = models.Manager() class Meta: verbose_name = _(u'robots.txt file') verbose_name_plural = _(u'robots.txt files') def __unicode__(self): return u'/'.join([self.site.domain, u'robots.txt'])
from django.db import models from django.contrib.sites.models import Site from django.utils.translation import ugettext_lazy as _ class File(models.Model): site = models.OneToOneField(Site, verbose_name=_(u'site')) content = models.TextField(_(u'file content')) class Meta: verbose_name = _(u'robots.txt file') verbose_name_plural = _(u'robots.txt files') def __unicode__(self): return u'/'.join([self.site.domain, u'robots.txt'])
Remove unnecessary manager declaration from File model
Remove unnecessary manager declaration from File model
Python
isc
trilan/lemon-robots,trilan/lemon-robots
from django.db import models from django.contrib.sites.models import Site from django.utils.translation import ugettext_lazy as _ class File(models.Model): site = models.OneToOneField(Site, verbose_name=_(u'site')) content = models.TextField(_(u'file content')) - objects = models.Manager() - class Meta: verbose_name = _(u'robots.txt file') verbose_name_plural = _(u'robots.txt files') def __unicode__(self): return u'/'.join([self.site.domain, u'robots.txt'])
Remove unnecessary manager declaration from File model
## Code Before: from django.db import models from django.contrib.sites.models import Site from django.utils.translation import ugettext_lazy as _ class File(models.Model): site = models.OneToOneField(Site, verbose_name=_(u'site')) content = models.TextField(_(u'file content')) objects = models.Manager() class Meta: verbose_name = _(u'robots.txt file') verbose_name_plural = _(u'robots.txt files') def __unicode__(self): return u'/'.join([self.site.domain, u'robots.txt']) ## Instruction: Remove unnecessary manager declaration from File model ## Code After: from django.db import models from django.contrib.sites.models import Site from django.utils.translation import ugettext_lazy as _ class File(models.Model): site = models.OneToOneField(Site, verbose_name=_(u'site')) content = models.TextField(_(u'file content')) class Meta: verbose_name = _(u'robots.txt file') verbose_name_plural = _(u'robots.txt files') def __unicode__(self): return u'/'.join([self.site.domain, u'robots.txt'])
... class Meta: ...
638ea1b12b71f74b357d60b09f1284625db73b2d
migrations/versions/0040_adjust_mmg_provider_rate.py
migrations/versions/0040_adjust_mmg_provider_rate.py
# revision identifiers, used by Alembic. revision = '0040_adjust_mmg_provider_rate' down_revision = '0039_fix_notifications' import uuid from datetime import datetime from alembic import op import sqlalchemy as sa def upgrade(): ### commands auto generated by Alembic - please adjust! ### conn = op.get_bind() conn.execute( sa.sql.text(("INSERT INTO provider_rates (id, valid_from, rate, provider_id) " "VALUES (:id, :valid_from, :rate, (SELECT id FROM provider_details WHERE identifier = 'mmg'))")), id=uuid.uuid4(), valid_from=datetime.utcnow(), rate=1.65 ) ### end Alembic commands ### def downgrade(): ### commands auto generated by Alembic - please adjust! ### conn = op.get_bind() conn.execute(("DELETE FROM provider_rates " "WHERE provider_id = (SELECT id FROM provider_details WHERE identifier = 'mmg') " "AND rate = 1.65")) ### end Alembic commands ###
# revision identifiers, used by Alembic. revision = '0040_adjust_mmg_provider_rate' down_revision = '0039_fix_notifications' import uuid from datetime import datetime from alembic import op import sqlalchemy as sa def upgrade(): ### commands auto generated by Alembic - please adjust! ### conn = op.get_bind() conn.execute( sa.sql.text(("INSERT INTO provider_rates (id, valid_from, rate, provider_id) " "VALUES (:id, :valid_from, :rate, (SELECT id FROM provider_details WHERE identifier = 'mmg'))")), id=uuid.uuid4(), valid_from=datetime(2016, 7, 1), rate=1.65 ) ### end Alembic commands ### def downgrade(): ### commands auto generated by Alembic - please adjust! ### conn = op.get_bind() conn.execute(("DELETE FROM provider_rates " "WHERE provider_id = (SELECT id FROM provider_details WHERE identifier = 'mmg') " "AND rate = 1.65")) ### end Alembic commands ###
Set the start date for the new rate as July 1
Set the start date for the new rate as July 1
Python
mit
alphagov/notifications-api,alphagov/notifications-api
# revision identifiers, used by Alembic. revision = '0040_adjust_mmg_provider_rate' down_revision = '0039_fix_notifications' import uuid from datetime import datetime from alembic import op import sqlalchemy as sa def upgrade(): ### commands auto generated by Alembic - please adjust! ### conn = op.get_bind() conn.execute( sa.sql.text(("INSERT INTO provider_rates (id, valid_from, rate, provider_id) " "VALUES (:id, :valid_from, :rate, (SELECT id FROM provider_details WHERE identifier = 'mmg'))")), id=uuid.uuid4(), - valid_from=datetime.utcnow(), + valid_from=datetime(2016, 7, 1), rate=1.65 ) ### end Alembic commands ### def downgrade(): ### commands auto generated by Alembic - please adjust! ### conn = op.get_bind() conn.execute(("DELETE FROM provider_rates " "WHERE provider_id = (SELECT id FROM provider_details WHERE identifier = 'mmg') " "AND rate = 1.65")) ### end Alembic commands ###
Set the start date for the new rate as July 1
## Code Before: # revision identifiers, used by Alembic. revision = '0040_adjust_mmg_provider_rate' down_revision = '0039_fix_notifications' import uuid from datetime import datetime from alembic import op import sqlalchemy as sa def upgrade(): ### commands auto generated by Alembic - please adjust! ### conn = op.get_bind() conn.execute( sa.sql.text(("INSERT INTO provider_rates (id, valid_from, rate, provider_id) " "VALUES (:id, :valid_from, :rate, (SELECT id FROM provider_details WHERE identifier = 'mmg'))")), id=uuid.uuid4(), valid_from=datetime.utcnow(), rate=1.65 ) ### end Alembic commands ### def downgrade(): ### commands auto generated by Alembic - please adjust! ### conn = op.get_bind() conn.execute(("DELETE FROM provider_rates " "WHERE provider_id = (SELECT id FROM provider_details WHERE identifier = 'mmg') " "AND rate = 1.65")) ### end Alembic commands ### ## Instruction: Set the start date for the new rate as July 1 ## Code After: # revision identifiers, used by Alembic. revision = '0040_adjust_mmg_provider_rate' down_revision = '0039_fix_notifications' import uuid from datetime import datetime from alembic import op import sqlalchemy as sa def upgrade(): ### commands auto generated by Alembic - please adjust! ### conn = op.get_bind() conn.execute( sa.sql.text(("INSERT INTO provider_rates (id, valid_from, rate, provider_id) " "VALUES (:id, :valid_from, :rate, (SELECT id FROM provider_details WHERE identifier = 'mmg'))")), id=uuid.uuid4(), valid_from=datetime(2016, 7, 1), rate=1.65 ) ### end Alembic commands ### def downgrade(): ### commands auto generated by Alembic - please adjust! ### conn = op.get_bind() conn.execute(("DELETE FROM provider_rates " "WHERE provider_id = (SELECT id FROM provider_details WHERE identifier = 'mmg') " "AND rate = 1.65")) ### end Alembic commands ###
# ... existing code ... id=uuid.uuid4(), valid_from=datetime(2016, 7, 1), rate=1.65 # ... rest of the code ...
783766b4f4d65dfb4b41e6386edd8ea2df32d727
tests/test_creation.py
tests/test_creation.py
import generic as g class CreationTest(g.unittest.TestCase): def test_soup(self): count = 100 mesh = g.trimesh.creation.random_soup(face_count=count) self.assertTrue(len(mesh.faces) == count) self.assertTrue(len(mesh.face_adjacency) == 0) self.assertTrue(len(mesh.split(only_watertight=True)) == 0) self.assertTrue(len(mesh.split(only_watertight=False)) == count) def test_uv(self): sphere = g.trimesh.creation.uv_sphere() self.assertTrue(sphere.is_watertight) self.assertTrue(sphere.is_winding_consistent) if __name__ == '__main__': g.trimesh.util.attach_to_log() g.unittest.main()
import generic as g class CreationTest(g.unittest.TestCase): def test_soup(self): count = 100 mesh = g.trimesh.creation.random_soup(face_count=count) self.assertTrue(len(mesh.faces) == count) self.assertTrue(len(mesh.face_adjacency) == 0) self.assertTrue(len(mesh.split(only_watertight=True)) == 0) self.assertTrue(len(mesh.split(only_watertight=False)) == count) def test_uv(self): sphere = g.trimesh.creation.uv_sphere() self.assertTrue(sphere.is_watertight) self.assertTrue(sphere.is_winding_consistent) def test_path_extrude(self): # Create base polygon vec = g.np.array([0,1])*0.2 n_comps = 100 angle = g.np.pi * 2.0 / n_comps rotmat = g.np.array([ [g.np.cos(angle), -g.np.sin(angle)], [g.np.sin(angle), g.np.cos(angle)] ]) perim = [] for i in range(n_comps): perim.append(vec) vec = g.np.dot(rotmat, vec) poly = g.Polygon(perim) # Create 3D path angles = g.np.linspace(0, 8*g.np.pi, 1000) x = angles / 10.0 y = g.np.cos(angles) z = g.np.sin(angles) path = g.np.c_[x,y,z] # Extrude mesh = g.trimesh.creation.extrude_polygon_along_path(poly, path) self.assertTrue(mesh.is_volume) if __name__ == '__main__': g.trimesh.util.attach_to_log() g.unittest.main()
Add integration test for extrusion
Add integration test for extrusion
Python
mit
mikedh/trimesh,mikedh/trimesh,mikedh/trimesh,dajusc/trimesh,mikedh/trimesh,dajusc/trimesh
import generic as g class CreationTest(g.unittest.TestCase): def test_soup(self): count = 100 mesh = g.trimesh.creation.random_soup(face_count=count) self.assertTrue(len(mesh.faces) == count) self.assertTrue(len(mesh.face_adjacency) == 0) self.assertTrue(len(mesh.split(only_watertight=True)) == 0) self.assertTrue(len(mesh.split(only_watertight=False)) == count) def test_uv(self): sphere = g.trimesh.creation.uv_sphere() self.assertTrue(sphere.is_watertight) self.assertTrue(sphere.is_winding_consistent) + def test_path_extrude(self): + # Create base polygon + vec = g.np.array([0,1])*0.2 + n_comps = 100 + angle = g.np.pi * 2.0 / n_comps + rotmat = g.np.array([ + [g.np.cos(angle), -g.np.sin(angle)], + [g.np.sin(angle), g.np.cos(angle)] + ]) + perim = [] + for i in range(n_comps): + perim.append(vec) + vec = g.np.dot(rotmat, vec) + poly = g.Polygon(perim) + + # Create 3D path + angles = g.np.linspace(0, 8*g.np.pi, 1000) + x = angles / 10.0 + y = g.np.cos(angles) + z = g.np.sin(angles) + path = g.np.c_[x,y,z] + + # Extrude + mesh = g.trimesh.creation.extrude_polygon_along_path(poly, path) + self.assertTrue(mesh.is_volume) + if __name__ == '__main__': g.trimesh.util.attach_to_log() g.unittest.main()
Add integration test for extrusion
## Code Before: import generic as g class CreationTest(g.unittest.TestCase): def test_soup(self): count = 100 mesh = g.trimesh.creation.random_soup(face_count=count) self.assertTrue(len(mesh.faces) == count) self.assertTrue(len(mesh.face_adjacency) == 0) self.assertTrue(len(mesh.split(only_watertight=True)) == 0) self.assertTrue(len(mesh.split(only_watertight=False)) == count) def test_uv(self): sphere = g.trimesh.creation.uv_sphere() self.assertTrue(sphere.is_watertight) self.assertTrue(sphere.is_winding_consistent) if __name__ == '__main__': g.trimesh.util.attach_to_log() g.unittest.main() ## Instruction: Add integration test for extrusion ## Code After: import generic as g class CreationTest(g.unittest.TestCase): def test_soup(self): count = 100 mesh = g.trimesh.creation.random_soup(face_count=count) self.assertTrue(len(mesh.faces) == count) self.assertTrue(len(mesh.face_adjacency) == 0) self.assertTrue(len(mesh.split(only_watertight=True)) == 0) self.assertTrue(len(mesh.split(only_watertight=False)) == count) def test_uv(self): sphere = g.trimesh.creation.uv_sphere() self.assertTrue(sphere.is_watertight) self.assertTrue(sphere.is_winding_consistent) def test_path_extrude(self): # Create base polygon vec = g.np.array([0,1])*0.2 n_comps = 100 angle = g.np.pi * 2.0 / n_comps rotmat = g.np.array([ [g.np.cos(angle), -g.np.sin(angle)], [g.np.sin(angle), g.np.cos(angle)] ]) perim = [] for i in range(n_comps): perim.append(vec) vec = g.np.dot(rotmat, vec) poly = g.Polygon(perim) # Create 3D path angles = g.np.linspace(0, 8*g.np.pi, 1000) x = angles / 10.0 y = g.np.cos(angles) z = g.np.sin(angles) path = g.np.c_[x,y,z] # Extrude mesh = g.trimesh.creation.extrude_polygon_along_path(poly, path) self.assertTrue(mesh.is_volume) if __name__ == '__main__': g.trimesh.util.attach_to_log() g.unittest.main()
// ... existing code ... def test_path_extrude(self): # Create base polygon vec = g.np.array([0,1])*0.2 n_comps = 100 angle = g.np.pi * 2.0 / n_comps rotmat = g.np.array([ [g.np.cos(angle), -g.np.sin(angle)], [g.np.sin(angle), g.np.cos(angle)] ]) perim = [] for i in range(n_comps): perim.append(vec) vec = g.np.dot(rotmat, vec) poly = g.Polygon(perim) # Create 3D path angles = g.np.linspace(0, 8*g.np.pi, 1000) x = angles / 10.0 y = g.np.cos(angles) z = g.np.sin(angles) path = g.np.c_[x,y,z] # Extrude mesh = g.trimesh.creation.extrude_polygon_along_path(poly, path) self.assertTrue(mesh.is_volume) // ... rest of the code ...
1f914a04adb4ad7d39ca7104e2ea36acc76b18bd
pvextractor/tests/test_gui.py
pvextractor/tests/test_gui.py
import numpy as np from numpy.testing import assert_allclose import pytest from astropy.io import fits from ..pvextractor import extract_pv_slice from ..geometry.path import Path from ..gui import PVSlicer from .test_slicer import make_test_hdu try: import PyQt5 PYQT5OK = True except ImportError: PYQT5OK = False import matplotlib as mpl if mpl.__version__[0] == '2': MPLOK = False else: MPLOK = True @pytest.mark.skipif('not PYQT5OK or not MPLOK') def test_gui(): hdu = make_test_hdu() pv = PVSlicer(hdu, clim=(-0.02, 2)) pv.show(block=False) x = [100,200,220,330,340] y = [100,200,300,420,430] for i in range(len(x)): pv.fig.canvas.motion_notify_event(x[i],y[i]) pv.fig.canvas.button_press_event(x[i],y[i],1) pv.fig.canvas.key_press_event('enter') pv.fig.canvas.motion_notify_event(310,420) pv.fig.canvas.button_press_event(410,420,1) pv.fig.canvas.draw() assert pv.pv_slice.data.shape == (5,2)
import pytest from distutils.version import LooseVersion import matplotlib as mpl from ..gui import PVSlicer from .test_slicer import make_test_hdu try: import PyQt5 PYQT5OK = True except ImportError: PYQT5OK = False if LooseVersion(mpl.__version__) < LooseVersion('2'): MPLOK = True else: MPLOK = False @pytest.mark.skipif('not PYQT5OK or not MPLOK') def test_gui(): hdu = make_test_hdu() pv = PVSlicer(hdu, clim=(-0.02, 2)) pv.show(block=False) x = [100, 200, 220, 330, 340] y = [100, 200, 300, 420, 430] for i in range(len(x)): pv.fig.canvas.motion_notify_event(x[i], y[i]) pv.fig.canvas.button_press_event(x[i], y[i], 1) pv.fig.canvas.key_press_event('enter') pv.fig.canvas.motion_notify_event(310, 420) pv.fig.canvas.button_press_event(410, 420, 1) pv.fig.canvas.draw() assert pv.pv_slice.data.shape == (5, 2)
Use LooseVersion to compare version numbers
Use LooseVersion to compare version numbers
Python
bsd-3-clause
radio-astro-tools/pvextractor,keflavich/pvextractor
- import numpy as np - from numpy.testing import assert_allclose import pytest + from distutils.version import LooseVersion + import matplotlib as mpl - from astropy.io import fits - from ..pvextractor import extract_pv_slice - from ..geometry.path import Path from ..gui import PVSlicer from .test_slicer import make_test_hdu try: import PyQt5 PYQT5OK = True except ImportError: PYQT5OK = False - import matplotlib as mpl - if mpl.__version__[0] == '2': + + if LooseVersion(mpl.__version__) < LooseVersion('2'): + MPLOK = True + else: MPLOK = False + - else: - MPLOK = True @pytest.mark.skipif('not PYQT5OK or not MPLOK') def test_gui(): hdu = make_test_hdu() pv = PVSlicer(hdu, clim=(-0.02, 2)) pv.show(block=False) - x = [100,200,220,330,340] + x = [100, 200, 220, 330, 340] - y = [100,200,300,420,430] + y = [100, 200, 300, 420, 430] for i in range(len(x)): - pv.fig.canvas.motion_notify_event(x[i],y[i]) + pv.fig.canvas.motion_notify_event(x[i], y[i]) - pv.fig.canvas.button_press_event(x[i],y[i],1) + pv.fig.canvas.button_press_event(x[i], y[i], 1) pv.fig.canvas.key_press_event('enter') - pv.fig.canvas.motion_notify_event(310,420) + pv.fig.canvas.motion_notify_event(310, 420) - pv.fig.canvas.button_press_event(410,420,1) + pv.fig.canvas.button_press_event(410, 420, 1) pv.fig.canvas.draw() - assert pv.pv_slice.data.shape == (5,2) + assert pv.pv_slice.data.shape == (5, 2)
Use LooseVersion to compare version numbers
## Code Before: import numpy as np from numpy.testing import assert_allclose import pytest from astropy.io import fits from ..pvextractor import extract_pv_slice from ..geometry.path import Path from ..gui import PVSlicer from .test_slicer import make_test_hdu try: import PyQt5 PYQT5OK = True except ImportError: PYQT5OK = False import matplotlib as mpl if mpl.__version__[0] == '2': MPLOK = False else: MPLOK = True @pytest.mark.skipif('not PYQT5OK or not MPLOK') def test_gui(): hdu = make_test_hdu() pv = PVSlicer(hdu, clim=(-0.02, 2)) pv.show(block=False) x = [100,200,220,330,340] y = [100,200,300,420,430] for i in range(len(x)): pv.fig.canvas.motion_notify_event(x[i],y[i]) pv.fig.canvas.button_press_event(x[i],y[i],1) pv.fig.canvas.key_press_event('enter') pv.fig.canvas.motion_notify_event(310,420) pv.fig.canvas.button_press_event(410,420,1) pv.fig.canvas.draw() assert pv.pv_slice.data.shape == (5,2) ## Instruction: Use LooseVersion to compare version numbers ## Code After: import pytest from distutils.version import LooseVersion import matplotlib as mpl from ..gui import PVSlicer from .test_slicer import make_test_hdu try: import PyQt5 PYQT5OK = True except ImportError: PYQT5OK = False if LooseVersion(mpl.__version__) < LooseVersion('2'): MPLOK = True else: MPLOK = False @pytest.mark.skipif('not PYQT5OK or not MPLOK') def test_gui(): hdu = make_test_hdu() pv = PVSlicer(hdu, clim=(-0.02, 2)) pv.show(block=False) x = [100, 200, 220, 330, 340] y = [100, 200, 300, 420, 430] for i in range(len(x)): pv.fig.canvas.motion_notify_event(x[i], y[i]) pv.fig.canvas.button_press_event(x[i], y[i], 1) pv.fig.canvas.key_press_event('enter') pv.fig.canvas.motion_notify_event(310, 420) pv.fig.canvas.button_press_event(410, 420, 1) pv.fig.canvas.draw() assert pv.pv_slice.data.shape == (5, 2)
... import pytest from distutils.version import LooseVersion import matplotlib as mpl from ..gui import PVSlicer ... if LooseVersion(mpl.__version__) < LooseVersion('2'): MPLOK = True else: MPLOK = False ... x = [100, 200, 220, 330, 340] y = [100, 200, 300, 420, 430] ... for i in range(len(x)): pv.fig.canvas.motion_notify_event(x[i], y[i]) pv.fig.canvas.button_press_event(x[i], y[i], 1) ... pv.fig.canvas.key_press_event('enter') pv.fig.canvas.motion_notify_event(310, 420) pv.fig.canvas.button_press_event(410, 420, 1) ... assert pv.pv_slice.data.shape == (5, 2) ...
1bd344a3ccda43f4ac1d4b94b1a18fc816c9b6ae
slurmscale/jobs/jobs.py
slurmscale/jobs/jobs.py
"""Get info about jobs running on this cluster.""" import pyslurm from job import Job class Jobs(object): """A service object to inspect jobs.""" @property def _jobs(self): """Fetch fresh data.""" return pyslurm.job().get() def list(self): """List the current jobs on the cluster.""" current_jobs = self._jobs return [Job(current_jobs[j]) for j in current_jobs]
"""Get info about jobs running on this cluster.""" import pyslurm from job import Job class Jobs(object): """A service object to inspect jobs.""" @property def _jobs(self): """Fetch fresh data.""" return pyslurm.job().get() def list(self, states=None): """ List the current jobs on the cluster. :type states: List of ``str`` :param states: Filter jobs in the given states. Available states are ``PENDING``, ``RUNNING``, ``CANCELLED``, ``CONFIGURING``, ``COMPLETING``, ``COMPLETED``, ``FAILED``, ``TIMEOUT``, ``PREEMPTED``, ``NODE_FAIL`` and ``SPECIAL_EXIT``. :rtype: List of ``Job`` :return: A list of current cluster jobs, possibly filtered by supplied states. """ current_jobs = self._jobs jobs = [] if states: for i in current_jobs: if current_jobs[i].get('job_state', '') in states: jobs.append(Job(current_jobs[i])) else: jobs = [Job(current_jobs[j]) for j in current_jobs] return jobs
Add ability to filter job list by job state
Add ability to filter job list by job state
Python
mit
afgane/slurmscale,afgane/slurmscale
"""Get info about jobs running on this cluster.""" import pyslurm from job import Job class Jobs(object): """A service object to inspect jobs.""" @property def _jobs(self): """Fetch fresh data.""" return pyslurm.job().get() - def list(self): + def list(self, states=None): + """ - """List the current jobs on the cluster.""" + List the current jobs on the cluster. + + :type states: List of ``str`` + :param states: Filter jobs in the given states. Available states are + ``PENDING``, ``RUNNING``, ``CANCELLED``, ``CONFIGURING``, + ``COMPLETING``, ``COMPLETED``, ``FAILED``, ``TIMEOUT``, + ``PREEMPTED``, ``NODE_FAIL`` and ``SPECIAL_EXIT``. + + :rtype: List of ``Job`` + :return: A list of current cluster jobs, possibly filtered by supplied + states. + """ current_jobs = self._jobs + jobs = [] + if states: + for i in current_jobs: + if current_jobs[i].get('job_state', '') in states: + jobs.append(Job(current_jobs[i])) + else: - return [Job(current_jobs[j]) for j in current_jobs] + jobs = [Job(current_jobs[j]) for j in current_jobs] + return jobs
Add ability to filter job list by job state
## Code Before: """Get info about jobs running on this cluster.""" import pyslurm from job import Job class Jobs(object): """A service object to inspect jobs.""" @property def _jobs(self): """Fetch fresh data.""" return pyslurm.job().get() def list(self): """List the current jobs on the cluster.""" current_jobs = self._jobs return [Job(current_jobs[j]) for j in current_jobs] ## Instruction: Add ability to filter job list by job state ## Code After: """Get info about jobs running on this cluster.""" import pyslurm from job import Job class Jobs(object): """A service object to inspect jobs.""" @property def _jobs(self): """Fetch fresh data.""" return pyslurm.job().get() def list(self, states=None): """ List the current jobs on the cluster. :type states: List of ``str`` :param states: Filter jobs in the given states. Available states are ``PENDING``, ``RUNNING``, ``CANCELLED``, ``CONFIGURING``, ``COMPLETING``, ``COMPLETED``, ``FAILED``, ``TIMEOUT``, ``PREEMPTED``, ``NODE_FAIL`` and ``SPECIAL_EXIT``. :rtype: List of ``Job`` :return: A list of current cluster jobs, possibly filtered by supplied states. """ current_jobs = self._jobs jobs = [] if states: for i in current_jobs: if current_jobs[i].get('job_state', '') in states: jobs.append(Job(current_jobs[i])) else: jobs = [Job(current_jobs[j]) for j in current_jobs] return jobs
# ... existing code ... def list(self, states=None): """ List the current jobs on the cluster. :type states: List of ``str`` :param states: Filter jobs in the given states. Available states are ``PENDING``, ``RUNNING``, ``CANCELLED``, ``CONFIGURING``, ``COMPLETING``, ``COMPLETED``, ``FAILED``, ``TIMEOUT``, ``PREEMPTED``, ``NODE_FAIL`` and ``SPECIAL_EXIT``. :rtype: List of ``Job`` :return: A list of current cluster jobs, possibly filtered by supplied states. """ current_jobs = self._jobs jobs = [] if states: for i in current_jobs: if current_jobs[i].get('job_state', '') in states: jobs.append(Job(current_jobs[i])) else: jobs = [Job(current_jobs[j]) for j in current_jobs] return jobs # ... rest of the code ...
0ac3ab3278e81aebe4717e0f599f752b4fda06d3
examples/swat-s1/tests.py
examples/swat-s1/tests.py
# from mininet.cli import CLI from mininet.net import Mininet from nose.plugins.skip import SkipTest from utils import STATE, RWT_INIT_LEVEL from utils import TANK_SECTION from topo import SwatTopo from physical_process import RawWaterTank # import subprocess # import sys @SkipTest def test_init(): pass def test_topo(): topo = SwatTopo() net = Mininet(topo=topo) net.start() net.pingAll() net.stop() def test_raw_water_tank(): RawWaterTank( name='test_rwt', state=STATE, protocol=None, section=TANK_SECTION, level=RWT_INIT_LEVEL )
# from mininet.cli import CLI from mininet.net import Mininet from utils import STATE, RWT_INIT_LEVEL from utils import TANK_SECTION from topo import SwatTopo from physical_process import RawWaterTank # import subprocess # import sys def test_init(): pass def test_topo(): topo = SwatTopo() net = Mininet(topo=topo) net.start() net.pingAll() net.stop() def test_raw_water_tank(): RawWaterTank( name='test_rwt', state=STATE, protocol=None, section=TANK_SECTION, level=RWT_INIT_LEVEL )
Remove examples dep from nose
Remove examples dep from nose
Python
mit
remmihsorp/minicps,scy-phy/minicps,remmihsorp/minicps,scy-phy/minicps
# from mininet.cli import CLI from mininet.net import Mininet - - from nose.plugins.skip import SkipTest from utils import STATE, RWT_INIT_LEVEL from utils import TANK_SECTION from topo import SwatTopo from physical_process import RawWaterTank # import subprocess # import sys - @SkipTest def test_init(): pass def test_topo(): topo = SwatTopo() net = Mininet(topo=topo) net.start() net.pingAll() net.stop() def test_raw_water_tank(): RawWaterTank( name='test_rwt', state=STATE, protocol=None, section=TANK_SECTION, level=RWT_INIT_LEVEL )
Remove examples dep from nose
## Code Before: # from mininet.cli import CLI from mininet.net import Mininet from nose.plugins.skip import SkipTest from utils import STATE, RWT_INIT_LEVEL from utils import TANK_SECTION from topo import SwatTopo from physical_process import RawWaterTank # import subprocess # import sys @SkipTest def test_init(): pass def test_topo(): topo = SwatTopo() net = Mininet(topo=topo) net.start() net.pingAll() net.stop() def test_raw_water_tank(): RawWaterTank( name='test_rwt', state=STATE, protocol=None, section=TANK_SECTION, level=RWT_INIT_LEVEL ) ## Instruction: Remove examples dep from nose ## Code After: # from mininet.cli import CLI from mininet.net import Mininet from utils import STATE, RWT_INIT_LEVEL from utils import TANK_SECTION from topo import SwatTopo from physical_process import RawWaterTank # import subprocess # import sys def test_init(): pass def test_topo(): topo = SwatTopo() net = Mininet(topo=topo) net.start() net.pingAll() net.stop() def test_raw_water_tank(): RawWaterTank( name='test_rwt', state=STATE, protocol=None, section=TANK_SECTION, level=RWT_INIT_LEVEL )
... from mininet.net import Mininet ... def test_init(): ...
e3312c773e9e3ac9b939bc3e0ca6a872dae5cdef
pre_commit_hooks/trailing_whitespace_fixer.py
pre_commit_hooks/trailing_whitespace_fixer.py
from __future__ import print_function import argparse import sys from plumbum import local from pre_commit_hooks.util import entry @entry def fix_trailing_whitespace(argv): parser = argparse.ArgumentParser() parser.add_argument('filenames', nargs='*', help='Filenames to fix') args = parser.parse_args(argv) bad_whitespace_files = local['grep'][ ('-l', '[[:space:]]$') + tuple(args.filenames) ](retcode=None).strip().splitlines() if bad_whitespace_files: for bad_whitespace_file in bad_whitespace_files: print('Fixing {0}'.format(bad_whitespace_file)) local['sed']['-i', '-e', 's/[[:space:]]*$//', bad_whitespace_file]() return 1 else: return 0 if __name__ == '__main__': sys.exit(fix_trailing_whitespace())
from __future__ import print_function import argparse import fileinput import sys from plumbum import local from pre_commit_hooks.util import entry def _fix_file(filename): for line in fileinput.input([filename], inplace=True): print(line.rstrip()) @entry def fix_trailing_whitespace(argv): parser = argparse.ArgumentParser() parser.add_argument('filenames', nargs='*', help='Filenames to fix') args = parser.parse_args(argv) bad_whitespace_files = local['grep'][ ('-l', '[[:space:]]$') + tuple(args.filenames) ](retcode=None).strip().splitlines() if bad_whitespace_files: for bad_whitespace_file in bad_whitespace_files: print('Fixing {0}'.format(bad_whitespace_file)) _fix_file(bad_whitespace_file) return 1 else: return 0 if __name__ == '__main__': sys.exit(fix_trailing_whitespace())
Use fileinput instead of sed.
Use fileinput instead of sed.
Python
mit
Coverfox/pre-commit-hooks,Harwood/pre-commit-hooks,bgschiller/pre-commit-hooks,pre-commit/pre-commit-hooks,jordant/pre-commit-hooks,jordant/pre-commit-hooks,chriskuehl/pre-commit-hooks,dupuy/pre-commit-hooks,arahayrabedian/pre-commit-hooks
from __future__ import print_function import argparse + import fileinput import sys from plumbum import local from pre_commit_hooks.util import entry + + + def _fix_file(filename): + for line in fileinput.input([filename], inplace=True): + print(line.rstrip()) @entry def fix_trailing_whitespace(argv): parser = argparse.ArgumentParser() parser.add_argument('filenames', nargs='*', help='Filenames to fix') args = parser.parse_args(argv) bad_whitespace_files = local['grep'][ ('-l', '[[:space:]]$') + tuple(args.filenames) ](retcode=None).strip().splitlines() if bad_whitespace_files: for bad_whitespace_file in bad_whitespace_files: print('Fixing {0}'.format(bad_whitespace_file)) - local['sed']['-i', '-e', 's/[[:space:]]*$//', bad_whitespace_file]() + _fix_file(bad_whitespace_file) return 1 else: return 0 if __name__ == '__main__': sys.exit(fix_trailing_whitespace())
Use fileinput instead of sed.
## Code Before: from __future__ import print_function import argparse import sys from plumbum import local from pre_commit_hooks.util import entry @entry def fix_trailing_whitespace(argv): parser = argparse.ArgumentParser() parser.add_argument('filenames', nargs='*', help='Filenames to fix') args = parser.parse_args(argv) bad_whitespace_files = local['grep'][ ('-l', '[[:space:]]$') + tuple(args.filenames) ](retcode=None).strip().splitlines() if bad_whitespace_files: for bad_whitespace_file in bad_whitespace_files: print('Fixing {0}'.format(bad_whitespace_file)) local['sed']['-i', '-e', 's/[[:space:]]*$//', bad_whitespace_file]() return 1 else: return 0 if __name__ == '__main__': sys.exit(fix_trailing_whitespace()) ## Instruction: Use fileinput instead of sed. ## Code After: from __future__ import print_function import argparse import fileinput import sys from plumbum import local from pre_commit_hooks.util import entry def _fix_file(filename): for line in fileinput.input([filename], inplace=True): print(line.rstrip()) @entry def fix_trailing_whitespace(argv): parser = argparse.ArgumentParser() parser.add_argument('filenames', nargs='*', help='Filenames to fix') args = parser.parse_args(argv) bad_whitespace_files = local['grep'][ ('-l', '[[:space:]]$') + tuple(args.filenames) ](retcode=None).strip().splitlines() if bad_whitespace_files: for bad_whitespace_file in bad_whitespace_files: print('Fixing {0}'.format(bad_whitespace_file)) _fix_file(bad_whitespace_file) return 1 else: return 0 if __name__ == '__main__': sys.exit(fix_trailing_whitespace())
# ... existing code ... import argparse import fileinput import sys # ... modified code ... from pre_commit_hooks.util import entry def _fix_file(filename): for line in fileinput.input([filename], inplace=True): print(line.rstrip()) ... print('Fixing {0}'.format(bad_whitespace_file)) _fix_file(bad_whitespace_file) return 1 # ... rest of the code ...
eb368c344075ce78606d4656ebfb19c7e7ccdf50
src/054.py
src/054.py
from path import dirpath def ans(): lines = open(dirpath() + '054.txt').readlines() cards = [line.strip().split() for line in lines] return None if __name__ == '__main__': print(ans())
from collections import ( defaultdict, namedtuple, ) from path import dirpath def _value(rank): try: return int(rank) except ValueError: return 10 + 'TJQKA'.index(rank) def _sort_by_rank(hand): return list(reversed(sorted( hand, key=lambda card: _value(card[0]), ))) def _of_a_kind(hand, count): counts = defaultdict(list) for card in hand: counts[card[0]].append(card) filtered = { rank: cards for rank, cards in counts.items() if count <= len(cards) } if len(filtered) < 1: return None return max( filtered.values(), key=lambda cards: _value(cards[0][0]) ) def high_card(hand): return _of_a_kind(hand, 1) def two_of_a_kind(hand): return _of_a_kind(hand, 2) def three_of_a_kind(hand): return _of_a_kind(hand, 3) def four_of_a_kind(hand): return _of_a_kind(hand, 4) def full_house(hand): three = three_of_a_kind(hand) if not three: return None pair = two_of_a_kind([card for card in hand if card not in three]) if not pair: return None return three + pair def straight(hand): sorted_ = sorted([_value(card[0]) for card in hand]) if sorted_ == list(range(sorted_[0], sorted_[-1] + 1)): return _sort_by_rank(hand) return None def flush(hand): counts = defaultdict(list) for card in hand: counts[card[1]].append(card) for cards in counts.values(): if len(cards) == 5: return _sort_by_rank(cards) return None def straight_flush(hand): return flush(hand) if straight(hand) else None def ans(): lines = open(dirpath() + '054.txt').readlines() turns = [line.strip().split() for line in lines] num_wins = 0 for cards in turns: one = cards[:5] two = cards[5:] return None if __name__ == '__main__': print(ans())
Write some logic for 54
Write some logic for 54
Python
mit
mackorone/euler
+ from collections import ( + defaultdict, + namedtuple, + ) from path import dirpath + + + def _value(rank): + try: + return int(rank) + except ValueError: + return 10 + 'TJQKA'.index(rank) + + + def _sort_by_rank(hand): + return list(reversed(sorted( + hand, + key=lambda card: _value(card[0]), + ))) + + + def _of_a_kind(hand, count): + counts = defaultdict(list) + for card in hand: + counts[card[0]].append(card) + filtered = { + rank: cards for + rank, cards in counts.items() if + count <= len(cards) + } + if len(filtered) < 1: + return None + return max( + filtered.values(), + key=lambda cards: _value(cards[0][0]) + ) + + + def high_card(hand): + return _of_a_kind(hand, 1) + + + def two_of_a_kind(hand): + return _of_a_kind(hand, 2) + + + def three_of_a_kind(hand): + return _of_a_kind(hand, 3) + + + def four_of_a_kind(hand): + return _of_a_kind(hand, 4) + + + def full_house(hand): + three = three_of_a_kind(hand) + if not three: + return None + pair = two_of_a_kind([card for card in hand if card not in three]) + if not pair: + return None + return three + pair + + + def straight(hand): + sorted_ = sorted([_value(card[0]) for card in hand]) + if sorted_ == list(range(sorted_[0], sorted_[-1] + 1)): + return _sort_by_rank(hand) + return None + + + def flush(hand): + counts = defaultdict(list) + for card in hand: + counts[card[1]].append(card) + for cards in counts.values(): + if len(cards) == 5: + return _sort_by_rank(cards) + return None + + + def straight_flush(hand): + return flush(hand) if straight(hand) else None def ans(): lines = open(dirpath() + '054.txt').readlines() - cards = [line.strip().split() for line in lines] + turns = [line.strip().split() for line in lines] + num_wins = 0 + for cards in turns: + one = cards[:5] + two = cards[5:] return None if __name__ == '__main__': print(ans())
Write some logic for 54
## Code Before: from path import dirpath def ans(): lines = open(dirpath() + '054.txt').readlines() cards = [line.strip().split() for line in lines] return None if __name__ == '__main__': print(ans()) ## Instruction: Write some logic for 54 ## Code After: from collections import ( defaultdict, namedtuple, ) from path import dirpath def _value(rank): try: return int(rank) except ValueError: return 10 + 'TJQKA'.index(rank) def _sort_by_rank(hand): return list(reversed(sorted( hand, key=lambda card: _value(card[0]), ))) def _of_a_kind(hand, count): counts = defaultdict(list) for card in hand: counts[card[0]].append(card) filtered = { rank: cards for rank, cards in counts.items() if count <= len(cards) } if len(filtered) < 1: return None return max( filtered.values(), key=lambda cards: _value(cards[0][0]) ) def high_card(hand): return _of_a_kind(hand, 1) def two_of_a_kind(hand): return _of_a_kind(hand, 2) def three_of_a_kind(hand): return _of_a_kind(hand, 3) def four_of_a_kind(hand): return _of_a_kind(hand, 4) def full_house(hand): three = three_of_a_kind(hand) if not three: return None pair = two_of_a_kind([card for card in hand if card not in three]) if not pair: return None return three + pair def straight(hand): sorted_ = sorted([_value(card[0]) for card in hand]) if sorted_ == list(range(sorted_[0], sorted_[-1] + 1)): return _sort_by_rank(hand) return None def flush(hand): counts = defaultdict(list) for card in hand: counts[card[1]].append(card) for cards in counts.values(): if len(cards) == 5: return _sort_by_rank(cards) return None def straight_flush(hand): return flush(hand) if straight(hand) else None def ans(): lines = open(dirpath() + '054.txt').readlines() turns = [line.strip().split() for line in lines] num_wins = 0 for cards in turns: one = cards[:5] two = cards[5:] return None if __name__ == '__main__': print(ans())
// ... existing code ... from collections import ( defaultdict, namedtuple, ) from path import dirpath def _value(rank): try: return int(rank) except ValueError: return 10 + 'TJQKA'.index(rank) def _sort_by_rank(hand): return list(reversed(sorted( hand, key=lambda card: _value(card[0]), ))) def _of_a_kind(hand, count): counts = defaultdict(list) for card in hand: counts[card[0]].append(card) filtered = { rank: cards for rank, cards in counts.items() if count <= len(cards) } if len(filtered) < 1: return None return max( filtered.values(), key=lambda cards: _value(cards[0][0]) ) def high_card(hand): return _of_a_kind(hand, 1) def two_of_a_kind(hand): return _of_a_kind(hand, 2) def three_of_a_kind(hand): return _of_a_kind(hand, 3) def four_of_a_kind(hand): return _of_a_kind(hand, 4) def full_house(hand): three = three_of_a_kind(hand) if not three: return None pair = two_of_a_kind([card for card in hand if card not in three]) if not pair: return None return three + pair def straight(hand): sorted_ = sorted([_value(card[0]) for card in hand]) if sorted_ == list(range(sorted_[0], sorted_[-1] + 1)): return _sort_by_rank(hand) return None def flush(hand): counts = defaultdict(list) for card in hand: counts[card[1]].append(card) for cards in counts.values(): if len(cards) == 5: return _sort_by_rank(cards) return None def straight_flush(hand): return flush(hand) if straight(hand) else None // ... modified code ... lines = open(dirpath() + '054.txt').readlines() turns = [line.strip().split() for line in lines] num_wins = 0 for cards in turns: one = cards[:5] two = cards[5:] // ... rest of the code ...
4cdf5be2a3c01e1b16a5e49bdf770f9d8573e16e
icekit/utils/testing.py
icekit/utils/testing.py
import glob import os import uuid from django.core.files.base import ContentFile from PIL import Image from StringIO import StringIO def new_test_image(): """ Creates an automatically generated test image. In your testing `tearDown` method make sure to delete the test image with the helper function `delete_test_image`. The recommended way of using this helper function is as follows: object_1.image_property.save(*new_test_image()) :return: Image name and image content file. """ image_name = 'test-{}.png'.format(uuid.uuid4()) image_buf = StringIO() image = Image.new('RGBA', size=(50, 50), color=(256, 0, 0)) image.save(image_buf, 'png') image_buf.seek(0) return image_name, ContentFile(image_buf.read(), image_name) def delete_test_image(image_field): """ Deletes test image generated as well as thumbnails if created. The recommended way of using this helper function is as follows: delete_test_image(object_1.image_property) :param image_field: The image field on an object. :return: None. """ # ensure all thumbs are deleted for filename in glob.glob( os.path.join('public', 'media', 'thumbs', image_field.name) + '*'): os.unlink(filename) # delete the saved file image_field.delete()
import glob import os import uuid from PIL import Image from django.core.files.base import ContentFile from django.utils import six def new_test_image(): """ Creates an automatically generated test image. In your testing `tearDown` method make sure to delete the test image with the helper function `delete_test_image`. The recommended way of using this helper function is as follows: object_1.image_property.save(*new_test_image()) :return: Image name and image content file. """ image_name = 'test-{}.png'.format(uuid.uuid4()) image_buf = six.StringIO() image = Image.new('RGBA', size=(50, 50), color=(256, 0, 0)) image.save(image_buf, 'png') image_buf.seek(0) return image_name, ContentFile(image_buf.read(), image_name) def delete_test_image(image_field): """ Deletes test image generated as well as thumbnails if created. The recommended way of using this helper function is as follows: delete_test_image(object_1.image_property) :param image_field: The image field on an object. :return: None. """ # ensure all thumbs are deleted for filename in glob.glob( os.path.join('public', 'media', 'thumbs', image_field.name) + '*'): os.unlink(filename) # delete the saved file image_field.delete()
Update StringIO import for Python3 compat
Update StringIO import for Python3 compat
Python
mit
ic-labs/django-icekit,ic-labs/django-icekit,ic-labs/django-icekit,ic-labs/django-icekit
import glob import os import uuid + + from PIL import Image + from django.core.files.base import ContentFile + from django.utils import six - from PIL import Image - from StringIO import StringIO def new_test_image(): """ Creates an automatically generated test image. In your testing `tearDown` method make sure to delete the test image with the helper function `delete_test_image`. The recommended way of using this helper function is as follows: object_1.image_property.save(*new_test_image()) :return: Image name and image content file. """ image_name = 'test-{}.png'.format(uuid.uuid4()) - image_buf = StringIO() + image_buf = six.StringIO() image = Image.new('RGBA', size=(50, 50), color=(256, 0, 0)) image.save(image_buf, 'png') image_buf.seek(0) return image_name, ContentFile(image_buf.read(), image_name) def delete_test_image(image_field): """ Deletes test image generated as well as thumbnails if created. The recommended way of using this helper function is as follows: delete_test_image(object_1.image_property) :param image_field: The image field on an object. :return: None. """ # ensure all thumbs are deleted for filename in glob.glob( os.path.join('public', 'media', 'thumbs', image_field.name) + '*'): os.unlink(filename) # delete the saved file image_field.delete()
Update StringIO import for Python3 compat
## Code Before: import glob import os import uuid from django.core.files.base import ContentFile from PIL import Image from StringIO import StringIO def new_test_image(): """ Creates an automatically generated test image. In your testing `tearDown` method make sure to delete the test image with the helper function `delete_test_image`. The recommended way of using this helper function is as follows: object_1.image_property.save(*new_test_image()) :return: Image name and image content file. """ image_name = 'test-{}.png'.format(uuid.uuid4()) image_buf = StringIO() image = Image.new('RGBA', size=(50, 50), color=(256, 0, 0)) image.save(image_buf, 'png') image_buf.seek(0) return image_name, ContentFile(image_buf.read(), image_name) def delete_test_image(image_field): """ Deletes test image generated as well as thumbnails if created. The recommended way of using this helper function is as follows: delete_test_image(object_1.image_property) :param image_field: The image field on an object. :return: None. """ # ensure all thumbs are deleted for filename in glob.glob( os.path.join('public', 'media', 'thumbs', image_field.name) + '*'): os.unlink(filename) # delete the saved file image_field.delete() ## Instruction: Update StringIO import for Python3 compat ## Code After: import glob import os import uuid from PIL import Image from django.core.files.base import ContentFile from django.utils import six def new_test_image(): """ Creates an automatically generated test image. In your testing `tearDown` method make sure to delete the test image with the helper function `delete_test_image`. The recommended way of using this helper function is as follows: object_1.image_property.save(*new_test_image()) :return: Image name and image content file. """ image_name = 'test-{}.png'.format(uuid.uuid4()) image_buf = six.StringIO() image = Image.new('RGBA', size=(50, 50), color=(256, 0, 0)) image.save(image_buf, 'png') image_buf.seek(0) return image_name, ContentFile(image_buf.read(), image_name) def delete_test_image(image_field): """ Deletes test image generated as well as thumbnails if created. The recommended way of using this helper function is as follows: delete_test_image(object_1.image_property) :param image_field: The image field on an object. :return: None. """ # ensure all thumbs are deleted for filename in glob.glob( os.path.join('public', 'media', 'thumbs', image_field.name) + '*'): os.unlink(filename) # delete the saved file image_field.delete()
... import uuid from PIL import Image from django.core.files.base import ContentFile from django.utils import six ... image_name = 'test-{}.png'.format(uuid.uuid4()) image_buf = six.StringIO() image = Image.new('RGBA', size=(50, 50), color=(256, 0, 0)) ...
c224fdecf174077f3b7a15f056e65b10282fed38
tasks.py
tasks.py
from invoke import Collection from invocations import docs, testing # TODO: let from_module specify new name api = Collection.from_module(docs) # TODO: maybe allow rolling configuration into it too heh api.configure({ 'sphinx.source': 'sites/docs', 'sphinx.target': 'sites/docs/_build', }) api.name = 'docs' main = Collection.from_module(docs) main.name = 'main' main.configure({ 'sphinx.source': 'sites/main', 'sphinx.target': 'sites/main/_build', }) ns = Collection(testing.test, docs=api, main=main)
from invoke import Collection from invocations import docs, testing # Usage doc/API site api = Collection.from_module(docs, name='docs', config={ 'sphinx.source': 'sites/docs', 'sphinx.target': 'sites/docs/_build', }) # Main/about/changelog site main = Collection.from_module(docs, name='main', config={ 'sphinx.source': 'sites/main', 'sphinx.target': 'sites/main/_build', }) ns = Collection(testing.test, docs=api, main=main)
Use new behavior from newer Invoke
Use new behavior from newer Invoke
Python
lgpl-2.1
mirrorcoder/paramiko,thisch/paramiko,paramiko/paramiko,remram44/paramiko,zpzgone/paramiko,dlitz/paramiko,Automatic/paramiko,varunarya10/paramiko,ameily/paramiko,mhdaimi/paramiko,redixin/paramiko,rcorrieri/paramiko,thusoy/paramiko,torkil/paramiko,SebastianDeiss/paramiko,zarr12steven/paramiko,dorianpula/paramiko,toby82/paramiko,esc/paramiko,digitalquacks/paramiko,jorik041/paramiko,CptLemming/paramiko,davidbistolas/paramiko,selboo/paramiko,reaperhulk/paramiko,jaraco/paramiko,anadigi/paramiko,fvicente/paramiko
from invoke import Collection from invocations import docs, testing + # Usage doc/API site + api = Collection.from_module(docs, name='docs', config={ - # TODO: let from_module specify new name - api = Collection.from_module(docs) - # TODO: maybe allow rolling configuration into it too heh - api.configure({ 'sphinx.source': 'sites/docs', 'sphinx.target': 'sites/docs/_build', }) + # Main/about/changelog site + main = Collection.from_module(docs, name='main', config={ - api.name = 'docs' - main = Collection.from_module(docs) - main.name = 'main' - main.configure({ 'sphinx.source': 'sites/main', 'sphinx.target': 'sites/main/_build', }) ns = Collection(testing.test, docs=api, main=main)
Use new behavior from newer Invoke
## Code Before: from invoke import Collection from invocations import docs, testing # TODO: let from_module specify new name api = Collection.from_module(docs) # TODO: maybe allow rolling configuration into it too heh api.configure({ 'sphinx.source': 'sites/docs', 'sphinx.target': 'sites/docs/_build', }) api.name = 'docs' main = Collection.from_module(docs) main.name = 'main' main.configure({ 'sphinx.source': 'sites/main', 'sphinx.target': 'sites/main/_build', }) ns = Collection(testing.test, docs=api, main=main) ## Instruction: Use new behavior from newer Invoke ## Code After: from invoke import Collection from invocations import docs, testing # Usage doc/API site api = Collection.from_module(docs, name='docs', config={ 'sphinx.source': 'sites/docs', 'sphinx.target': 'sites/docs/_build', }) # Main/about/changelog site main = Collection.from_module(docs, name='main', config={ 'sphinx.source': 'sites/main', 'sphinx.target': 'sites/main/_build', }) ns = Collection(testing.test, docs=api, main=main)
... # Usage doc/API site api = Collection.from_module(docs, name='docs', config={ 'sphinx.source': 'sites/docs', ... }) # Main/about/changelog site main = Collection.from_module(docs, name='main', config={ 'sphinx.source': 'sites/main', ...
96b06d80f6108997fab44ac1e6042fcae93cc82a
server.py
server.py
import json import tornado.ioloop import tornado.web import Adafruit_BMP.BMP085 as BMP085 class SensorAccess(tornado.web.RequestHandler): @tornado.web.asynchronous def get(self): self.write(json.dumps(self.read_sensor())) self.finish() def read_sensor(self): pass class TempSensorAccess(SensorAccess): def read_sensor(self): sensor = BMP085.BMP085(mode=BMP085.BMP085_ULTRAHIGHRES) return { 'temperature': sensor.read_temperature(), 'pressure': sensor.read_pressure(), } class IndexHandler(tornado.web.RequestHandler): @tornado.web.asynchronous def get(self): self.write(json.dumps({ 'inde': 'pitools service' })) def start_server(): application = tornado.web.Application([ (r"/", IndexHandler), (r"/sensors/env", TempSensorAccess), ]) application.listen(9876) tornado.ioloop.IOLoop.instance().start() if __name__ == "__main__": start_server()
import json import tornado.ioloop import tornado.web import Adafruit_BMP.BMP085 as BMP085 class SensorAccess(tornado.web.RequestHandler): @tornado.web.asynchronous def get(self): self.write(json.dumps(self.read_sensor())) self.finish() def read_sensor(self): pass class TempSensorAccess(SensorAccess): def read_sensor(self): sensor = BMP085.BMP085(mode=BMP085.BMP085_ULTRAHIGHRES) return { 'temperature': sensor.read_temperature(), 'pressure': sensor.read_pressure(), } class IndexHandler(tornado.web.RequestHandler): @tornado.web.asynchronous def get(self): self.write(json.dumps({ 'index': 'pitools service' })) self.finish() def start_server(): application = tornado.web.Application([ (r"/", IndexHandler), (r"/sensors/env", TempSensorAccess), ]) application.listen(9876) tornado.ioloop.IOLoop.instance().start() if __name__ == "__main__": start_server()
Fix typo; Fix request never finish
Fix typo; Fix request never finish
Python
bsd-2-clause
JokerQyou/pitools
import json import tornado.ioloop import tornado.web import Adafruit_BMP.BMP085 as BMP085 class SensorAccess(tornado.web.RequestHandler): @tornado.web.asynchronous def get(self): self.write(json.dumps(self.read_sensor())) self.finish() def read_sensor(self): pass class TempSensorAccess(SensorAccess): def read_sensor(self): sensor = BMP085.BMP085(mode=BMP085.BMP085_ULTRAHIGHRES) return { 'temperature': sensor.read_temperature(), 'pressure': sensor.read_pressure(), } class IndexHandler(tornado.web.RequestHandler): @tornado.web.asynchronous def get(self): self.write(json.dumps({ - 'inde': 'pitools service' + 'index': 'pitools service' })) + self.finish() def start_server(): application = tornado.web.Application([ (r"/", IndexHandler), (r"/sensors/env", TempSensorAccess), ]) application.listen(9876) tornado.ioloop.IOLoop.instance().start() if __name__ == "__main__": start_server()
Fix typo; Fix request never finish
## Code Before: import json import tornado.ioloop import tornado.web import Adafruit_BMP.BMP085 as BMP085 class SensorAccess(tornado.web.RequestHandler): @tornado.web.asynchronous def get(self): self.write(json.dumps(self.read_sensor())) self.finish() def read_sensor(self): pass class TempSensorAccess(SensorAccess): def read_sensor(self): sensor = BMP085.BMP085(mode=BMP085.BMP085_ULTRAHIGHRES) return { 'temperature': sensor.read_temperature(), 'pressure': sensor.read_pressure(), } class IndexHandler(tornado.web.RequestHandler): @tornado.web.asynchronous def get(self): self.write(json.dumps({ 'inde': 'pitools service' })) def start_server(): application = tornado.web.Application([ (r"/", IndexHandler), (r"/sensors/env", TempSensorAccess), ]) application.listen(9876) tornado.ioloop.IOLoop.instance().start() if __name__ == "__main__": start_server() ## Instruction: Fix typo; Fix request never finish ## Code After: import json import tornado.ioloop import tornado.web import Adafruit_BMP.BMP085 as BMP085 class SensorAccess(tornado.web.RequestHandler): @tornado.web.asynchronous def get(self): self.write(json.dumps(self.read_sensor())) self.finish() def read_sensor(self): pass class TempSensorAccess(SensorAccess): def read_sensor(self): sensor = BMP085.BMP085(mode=BMP085.BMP085_ULTRAHIGHRES) return { 'temperature': sensor.read_temperature(), 'pressure': sensor.read_pressure(), } class IndexHandler(tornado.web.RequestHandler): @tornado.web.asynchronous def get(self): self.write(json.dumps({ 'index': 'pitools service' })) self.finish() def start_server(): application = tornado.web.Application([ (r"/", IndexHandler), (r"/sensors/env", TempSensorAccess), ]) application.listen(9876) tornado.ioloop.IOLoop.instance().start() if __name__ == "__main__": start_server()
# ... existing code ... self.write(json.dumps({ 'index': 'pitools service' })) self.finish() # ... rest of the code ...