commit
stringlengths
40
40
old_file
stringlengths
4
118
new_file
stringlengths
4
118
old_contents
stringlengths
10
3.52k
new_contents
stringlengths
21
3.18k
subject
stringlengths
16
444
message
stringlengths
17
2.63k
lang
stringclasses
1 value
license
stringclasses
13 values
repos
stringlengths
7
43k
ndiff
stringlengths
52
3.32k
instruction
stringlengths
16
444
content
stringlengths
133
4.32k
fuzzy_diff
stringlengths
17
3.24k
52d9ed9c08ef0686a891e3428349b70d74a7ecf8
scripts/munge_fah_data.py
scripts/munge_fah_data.py
import numpy as np import os import glob import mdtraj as md import fahmunge import pandas as pd projects = pd.read_csv("./projects.csv", index_col=0) output_path = "/data/choderalab/fah/munged/" for (project, location, pdb) in projects.itertuples(): print(project, location, pdb) allatom_output_path = os.path.join(output_path, str(project), "allatoms/") protein_output_path = os.path.join(output_path, str(project), "protein/") fahmunge.automation.make_path(allatom_output_path) fahmunge.automation.make_path(protein_output_path) fahmunge.automation.merge_fah_trajectories(location, allatom_output_path, pdb) trj0 = md.load(pdb) # Hacky temporary solution. top, bonds = trj0.top.to_dataframe() protein_atom_indices = top.index[top.chainID == 0].values fahmunge.automation.strip_water(allatom_output_path, protein_output_path, protein_atom_indices)
import numpy as np import os import glob import mdtraj as md import fahmunge import pandas as pd projects = pd.read_csv("./projects.csv", index_col=0) output_path = "/data/choderalab/fah/munged/" for (project, location, pdb) in projects.itertuples(): print(project, location, pdb) allatom_output_path = os.path.join(output_path, "allatoms/", "%s/" % project) protein_output_path = os.path.join(output_path, "protein/", "%s/" % project) fahmunge.automation.make_path(allatom_output_path) fahmunge.automation.make_path(protein_output_path) fahmunge.automation.merge_fah_trajectories(location, allatom_output_path, pdb) trj0 = md.load(pdb) # Hacky temporary solution. top, bonds = trj0.top.to_dataframe() protein_atom_indices = top.index[top.chainID == 0].values fahmunge.automation.strip_water(allatom_output_path, protein_output_path, protein_atom_indices)
Change output data structure to support faster rsync
Change output data structure to support faster rsync
Python
lgpl-2.1
steven-albanese/FAHMunge,kyleabeauchamp/FAHMunge,choderalab/FAHMunge
import numpy as np import os import glob import mdtraj as md import fahmunge import pandas as pd projects = pd.read_csv("./projects.csv", index_col=0) output_path = "/data/choderalab/fah/munged/" for (project, location, pdb) in projects.itertuples(): print(project, location, pdb) - allatom_output_path = os.path.join(output_path, str(project), "allatoms/") + allatom_output_path = os.path.join(output_path, "allatoms/", "%s/" % project) - protein_output_path = os.path.join(output_path, str(project), "protein/") + protein_output_path = os.path.join(output_path, "protein/", "%s/" % project) fahmunge.automation.make_path(allatom_output_path) fahmunge.automation.make_path(protein_output_path) fahmunge.automation.merge_fah_trajectories(location, allatom_output_path, pdb) trj0 = md.load(pdb) # Hacky temporary solution. top, bonds = trj0.top.to_dataframe() protein_atom_indices = top.index[top.chainID == 0].values fahmunge.automation.strip_water(allatom_output_path, protein_output_path, protein_atom_indices)
Change output data structure to support faster rsync
## Code Before: import numpy as np import os import glob import mdtraj as md import fahmunge import pandas as pd projects = pd.read_csv("./projects.csv", index_col=0) output_path = "/data/choderalab/fah/munged/" for (project, location, pdb) in projects.itertuples(): print(project, location, pdb) allatom_output_path = os.path.join(output_path, str(project), "allatoms/") protein_output_path = os.path.join(output_path, str(project), "protein/") fahmunge.automation.make_path(allatom_output_path) fahmunge.automation.make_path(protein_output_path) fahmunge.automation.merge_fah_trajectories(location, allatom_output_path, pdb) trj0 = md.load(pdb) # Hacky temporary solution. top, bonds = trj0.top.to_dataframe() protein_atom_indices = top.index[top.chainID == 0].values fahmunge.automation.strip_water(allatom_output_path, protein_output_path, protein_atom_indices) ## Instruction: Change output data structure to support faster rsync ## Code After: import numpy as np import os import glob import mdtraj as md import fahmunge import pandas as pd projects = pd.read_csv("./projects.csv", index_col=0) output_path = "/data/choderalab/fah/munged/" for (project, location, pdb) in projects.itertuples(): print(project, location, pdb) allatom_output_path = os.path.join(output_path, "allatoms/", "%s/" % project) protein_output_path = os.path.join(output_path, "protein/", "%s/" % project) fahmunge.automation.make_path(allatom_output_path) fahmunge.automation.make_path(protein_output_path) fahmunge.automation.merge_fah_trajectories(location, allatom_output_path, pdb) trj0 = md.load(pdb) # Hacky temporary solution. top, bonds = trj0.top.to_dataframe() protein_atom_indices = top.index[top.chainID == 0].values fahmunge.automation.strip_water(allatom_output_path, protein_output_path, protein_atom_indices)
// ... existing code ... for (project, location, pdb) in projects.itertuples(): print(project, location, pdb) allatom_output_path = os.path.join(output_path, "allatoms/", "%s/" % project) protein_output_path = os.path.join(output_path, "protein/", "%s/" % project) fahmunge.automation.make_path(allatom_output_path) fahmunge.automation.make_path(protein_output_path) // ... rest of the code ...
096f9e86755a6967d732986c51ae00855551cf4d
project_name/urls.py
project_name/urls.py
from django.conf import settings from django.conf.urls import include, url # noqa from django.contrib import admin from django.views.generic import TemplateView import django_js_reverse.views urlpatterns = [ url(r'^admin/', admin.site.urls), url(r'^jsreverse/$', django_js_reverse.views.urls_js, name='js_reverse'), url(r'^$', TemplateView.as_view(template_name='exampleapp/itworks.html'), name='home'), ] if settings.DEBUG: import debug_toolbar urlpatterns = [ url(r'^__debug__/', include(debug_toolbar.urls)), ] + urlpatterns
from django.conf.urls import include, url # noqa from django.contrib import admin from django.views.generic import TemplateView import django_js_reverse.views urlpatterns = [ url(r'^admin/', admin.site.urls), url(r'^jsreverse/$', django_js_reverse.views.urls_js, name='js_reverse'), url(r'^$', TemplateView.as_view(template_name='exampleapp/itworks.html'), name='home'), ]
Remove usage from debug toolbar
Remove usage from debug toolbar
Python
mit
vintasoftware/django-react-boilerplate,vintasoftware/django-react-boilerplate,vintasoftware/django-react-boilerplate,vintasoftware/django-react-boilerplate
- from django.conf import settings from django.conf.urls import include, url # noqa from django.contrib import admin from django.views.generic import TemplateView import django_js_reverse.views urlpatterns = [ url(r'^admin/', admin.site.urls), url(r'^jsreverse/$', django_js_reverse.views.urls_js, name='js_reverse'), url(r'^$', TemplateView.as_view(template_name='exampleapp/itworks.html'), name='home'), ] - if settings.DEBUG: - import debug_toolbar - urlpatterns = [ - url(r'^__debug__/', include(debug_toolbar.urls)), - ] + urlpatterns -
Remove usage from debug toolbar
## Code Before: from django.conf import settings from django.conf.urls import include, url # noqa from django.contrib import admin from django.views.generic import TemplateView import django_js_reverse.views urlpatterns = [ url(r'^admin/', admin.site.urls), url(r'^jsreverse/$', django_js_reverse.views.urls_js, name='js_reverse'), url(r'^$', TemplateView.as_view(template_name='exampleapp/itworks.html'), name='home'), ] if settings.DEBUG: import debug_toolbar urlpatterns = [ url(r'^__debug__/', include(debug_toolbar.urls)), ] + urlpatterns ## Instruction: Remove usage from debug toolbar ## Code After: from django.conf.urls import include, url # noqa from django.contrib import admin from django.views.generic import TemplateView import django_js_reverse.views urlpatterns = [ url(r'^admin/', admin.site.urls), url(r'^jsreverse/$', django_js_reverse.views.urls_js, name='js_reverse'), url(r'^$', TemplateView.as_view(template_name='exampleapp/itworks.html'), name='home'), ]
... from django.conf.urls import include, url # noqa from django.contrib import admin ... url(r'^$', TemplateView.as_view(template_name='exampleapp/itworks.html'), name='home'), ] ...
f112e7754e4f4368f0a82c3aae3a58f5300176f0
spacy/language_data/tag_map.py
spacy/language_data/tag_map.py
from __future__ import unicode_literals from ..symbols import * TAG_MAP = { "ADV": {POS: ADV}, "NOUN": {POS: NOUN}, "ADP": {POS: ADP}, "PRON": {POS: PRON}, "SCONJ": {POS: SCONJ}, "PROPN": {POS: PROPN}, "DET": {POS: DET}, "SYM": {POS: SYM}, "INTJ": {POS: INTJ}, "PUNCT": {POS: PUNCT}, "NUM": {POS: NUM}, "AUX": {POS: AUX}, "X": {POS: X}, "CONJ": {POS: CONJ}, "ADJ": {POS: ADJ}, "VERB": {POS: VERB} }
from __future__ import unicode_literals from ..symbols import * TAG_MAP = { "ADV": {POS: ADV}, "NOUN": {POS: NOUN}, "ADP": {POS: ADP}, "PRON": {POS: PRON}, "SCONJ": {POS: SCONJ}, "PROPN": {POS: PROPN}, "DET": {POS: DET}, "SYM": {POS: SYM}, "INTJ": {POS: INTJ}, "PUNCT": {POS: PUNCT}, "NUM": {POS: NUM}, "AUX": {POS: AUX}, "X": {POS: X}, "CONJ": {POS: CONJ}, "ADJ": {POS: ADJ}, "VERB": {POS: VERB}, "PART": {POS: PART} }
Add PART to tag map
Add PART to tag map 16 of the 17 PoS tags in the UD tag set is added; PART is missing.
Python
mit
banglakit/spaCy,raphael0202/spaCy,recognai/spaCy,spacy-io/spaCy,explosion/spaCy,spacy-io/spaCy,spacy-io/spaCy,honnibal/spaCy,recognai/spaCy,Gregory-Howard/spaCy,aikramer2/spaCy,Gregory-Howard/spaCy,oroszgy/spaCy.hu,oroszgy/spaCy.hu,raphael0202/spaCy,recognai/spaCy,Gregory-Howard/spaCy,aikramer2/spaCy,explosion/spaCy,honnibal/spaCy,explosion/spaCy,Gregory-Howard/spaCy,banglakit/spaCy,banglakit/spaCy,Gregory-Howard/spaCy,banglakit/spaCy,spacy-io/spaCy,oroszgy/spaCy.hu,raphael0202/spaCy,aikramer2/spaCy,honnibal/spaCy,banglakit/spaCy,recognai/spaCy,explosion/spaCy,aikramer2/spaCy,recognai/spaCy,raphael0202/spaCy,oroszgy/spaCy.hu,raphael0202/spaCy,recognai/spaCy,aikramer2/spaCy,spacy-io/spaCy,oroszgy/spaCy.hu,explosion/spaCy,oroszgy/spaCy.hu,raphael0202/spaCy,honnibal/spaCy,banglakit/spaCy,aikramer2/spaCy,Gregory-Howard/spaCy,spacy-io/spaCy,explosion/spaCy
from __future__ import unicode_literals from ..symbols import * TAG_MAP = { "ADV": {POS: ADV}, "NOUN": {POS: NOUN}, "ADP": {POS: ADP}, "PRON": {POS: PRON}, "SCONJ": {POS: SCONJ}, "PROPN": {POS: PROPN}, "DET": {POS: DET}, "SYM": {POS: SYM}, "INTJ": {POS: INTJ}, "PUNCT": {POS: PUNCT}, "NUM": {POS: NUM}, "AUX": {POS: AUX}, "X": {POS: X}, "CONJ": {POS: CONJ}, "ADJ": {POS: ADJ}, - "VERB": {POS: VERB} + "VERB": {POS: VERB}, + "PART": {POS: PART} }
Add PART to tag map
## Code Before: from __future__ import unicode_literals from ..symbols import * TAG_MAP = { "ADV": {POS: ADV}, "NOUN": {POS: NOUN}, "ADP": {POS: ADP}, "PRON": {POS: PRON}, "SCONJ": {POS: SCONJ}, "PROPN": {POS: PROPN}, "DET": {POS: DET}, "SYM": {POS: SYM}, "INTJ": {POS: INTJ}, "PUNCT": {POS: PUNCT}, "NUM": {POS: NUM}, "AUX": {POS: AUX}, "X": {POS: X}, "CONJ": {POS: CONJ}, "ADJ": {POS: ADJ}, "VERB": {POS: VERB} } ## Instruction: Add PART to tag map ## Code After: from __future__ import unicode_literals from ..symbols import * TAG_MAP = { "ADV": {POS: ADV}, "NOUN": {POS: NOUN}, "ADP": {POS: ADP}, "PRON": {POS: PRON}, "SCONJ": {POS: SCONJ}, "PROPN": {POS: PROPN}, "DET": {POS: DET}, "SYM": {POS: SYM}, "INTJ": {POS: INTJ}, "PUNCT": {POS: PUNCT}, "NUM": {POS: NUM}, "AUX": {POS: AUX}, "X": {POS: X}, "CONJ": {POS: CONJ}, "ADJ": {POS: ADJ}, "VERB": {POS: VERB}, "PART": {POS: PART} }
// ... existing code ... "CONJ": {POS: CONJ}, "ADJ": {POS: ADJ}, "VERB": {POS: VERB}, "PART": {POS: PART} } // ... rest of the code ...
1b9c4935b2edf6601c2d75d8a2d318266de2d456
circuits/tools/__init__.py
circuits/tools/__init__.py
try: from cStringIO import StringIO except ImportError: from StringIO import StringIO def graph(x): s = StringIO() d = 0 i = 0 done = False stack = [] visited = set() children = list(x.components) while not done: if x not in visited: if d: s.write("%s%s\n" % (" " * d, "|")) s.write("%s%s%s\n" % (" " * d, "|-", x)) else: s.write(" .%s\n" % x) if x.components: d += 1 visited.add(x) if i < len(children): x = children[i] i += 1 if x.components: stack.append((i, children)) children = list(x.components) i = 0 else: if stack: i, children = stack.pop() d -= 1 else: done = True return s.getvalue()
try: from cStringIO import StringIO except ImportError: from StringIO import StringIO def graph(x): s = StringIO() d = 0 i = 0 done = False stack = [] visited = set() children = list(x.components) while not done: if x not in visited: if d: s.write("%s%s\n" % (" " * d, "|")) s.write("%s%s%s\n" % (" " * d, "|-", x)) else: s.write(" .%s\n" % x) if x.components: d += 1 visited.add(x) if i < len(children): x = children[i] i += 1 if x.components: stack.append((i, d, children)) children = list(x.components) i = 0 else: if stack: i, d, children = stack.pop() else: done = True return s.getvalue()
Store the depth (d) on the stack and restore when backtracking
tools: Store the depth (d) on the stack and restore when backtracking
Python
mit
treemo/circuits,treemo/circuits,eriol/circuits,treemo/circuits,eriol/circuits,nizox/circuits,eriol/circuits
try: from cStringIO import StringIO except ImportError: from StringIO import StringIO def graph(x): s = StringIO() d = 0 i = 0 done = False stack = [] visited = set() children = list(x.components) while not done: if x not in visited: if d: s.write("%s%s\n" % (" " * d, "|")) s.write("%s%s%s\n" % (" " * d, "|-", x)) else: s.write(" .%s\n" % x) if x.components: d += 1 visited.add(x) if i < len(children): x = children[i] i += 1 if x.components: - stack.append((i, children)) + stack.append((i, d, children)) children = list(x.components) i = 0 else: if stack: - i, children = stack.pop() + i, d, children = stack.pop() - d -= 1 else: done = True return s.getvalue()
Store the depth (d) on the stack and restore when backtracking
## Code Before: try: from cStringIO import StringIO except ImportError: from StringIO import StringIO def graph(x): s = StringIO() d = 0 i = 0 done = False stack = [] visited = set() children = list(x.components) while not done: if x not in visited: if d: s.write("%s%s\n" % (" " * d, "|")) s.write("%s%s%s\n" % (" " * d, "|-", x)) else: s.write(" .%s\n" % x) if x.components: d += 1 visited.add(x) if i < len(children): x = children[i] i += 1 if x.components: stack.append((i, children)) children = list(x.components) i = 0 else: if stack: i, children = stack.pop() d -= 1 else: done = True return s.getvalue() ## Instruction: Store the depth (d) on the stack and restore when backtracking ## Code After: try: from cStringIO import StringIO except ImportError: from StringIO import StringIO def graph(x): s = StringIO() d = 0 i = 0 done = False stack = [] visited = set() children = list(x.components) while not done: if x not in visited: if d: s.write("%s%s\n" % (" " * d, "|")) s.write("%s%s%s\n" % (" " * d, "|-", x)) else: s.write(" .%s\n" % x) if x.components: d += 1 visited.add(x) if i < len(children): x = children[i] i += 1 if x.components: stack.append((i, d, children)) children = list(x.components) i = 0 else: if stack: i, d, children = stack.pop() else: done = True return s.getvalue()
# ... existing code ... i += 1 if x.components: stack.append((i, d, children)) children = list(x.components) i = 0 # ... modified code ... else: if stack: i, d, children = stack.pop() else: done = True # ... rest of the code ...
07b7efdc848deeb7634f23cdb878774a99c9c535
ambassador/tests/t_grpc_bridge.py
ambassador/tests/t_grpc_bridge.py
import json from kat.harness import Query from abstract_tests import AmbassadorTest, ServiceType, EGRPC class AcceptanceGrpcBridgeTest(AmbassadorTest): target: ServiceType def init(self): self.target = EGRPC() def config(self): yield self, self.format(""" --- apiVersion: ambassador/v0 kind: Module name: ambassador config: enable_grpc_http11_bridge: True """) yield self, self.format(""" --- apiVersion: ambassador/v0 kind: Mapping grpc: True prefix: /echo.EchoService/ rewrite: /echo.EchoService/ name: {self.target.path.k8s} service: {self.target.path.k8s} """) def queries(self): # [0] yield Query(self.url("echo.EchoService/Echo"), headers={ "content-type": "application/grpc", "requested-status": "0" }, expected=200) # [1] yield Query(self.url("echo.EchoService/Echo"), headers={ "content-type": "application/grpc", "requested-status": "7" }, expected=200) def check(self): # [0] assert self.results[0].status == 200 assert self.results[0].headers["Grpc-Status"] == ["0"] # [0] assert self.results[1].status == 200 assert self.results[1].headers["Grpc-Status"] == ["7"]
from kat.harness import Query from abstract_tests import AmbassadorTest, ServiceType, EGRPC class AcceptanceGrpcBridgeTest(AmbassadorTest): target: ServiceType def init(self): self.target = EGRPC() def config(self): yield self, self.format(""" --- apiVersion: ambassador/v0 kind: Module name: ambassador config: enable_grpc_http11_bridge: True """) yield self, self.format(""" --- apiVersion: ambassador/v0 kind: Mapping grpc: True prefix: /echo.EchoService/ rewrite: /echo.EchoService/ name: {self.target.path.k8s} service: {self.target.path.k8s} """) def queries(self): # [0] yield Query(self.url("echo.EchoService/Echo"), headers={ "content-type": "application/grpc", "requested-status": "0" }, expected=200) # [1] yield Query(self.url("echo.EchoService/Echo"), headers={ "content-type": "application/grpc", "requested-status": "7" }, expected=200) def check(self): # [0] assert self.results[0].headers["Grpc-Status"] == ["0"] # [1] assert self.results[1].headers["Grpc-Status"] == ["7"]
Clean up existing gRPC bridge test (make it like t_tcpmapping.py)
Clean up existing gRPC bridge test (make it like t_tcpmapping.py) - Remove redundant HTTP status assertions - Adjust formatting - Remove unused import
Python
apache-2.0
datawire/ambassador,datawire/ambassador,datawire/ambassador,datawire/ambassador,datawire/ambassador
- import json - from kat.harness import Query from abstract_tests import AmbassadorTest, ServiceType, EGRPC class AcceptanceGrpcBridgeTest(AmbassadorTest): target: ServiceType def init(self): self.target = EGRPC() def config(self): yield self, self.format(""" --- apiVersion: ambassador/v0 kind: Module name: ambassador config: enable_grpc_http11_bridge: True """) yield self, self.format(""" --- apiVersion: ambassador/v0 kind: Mapping grpc: True prefix: /echo.EchoService/ rewrite: /echo.EchoService/ name: {self.target.path.k8s} service: {self.target.path.k8s} """) def queries(self): # [0] - yield Query(self.url("echo.EchoService/Echo"), headers={ "content-type": "application/grpc", - "requested-status": "0" }, expected=200) + yield Query(self.url("echo.EchoService/Echo"), + headers={ "content-type": "application/grpc", "requested-status": "0" }, + expected=200) - # [1] + # [1] - yield Query(self.url("echo.EchoService/Echo"), headers={ "content-type": "application/grpc", - "requested-status": "7" }, expected=200) + yield Query(self.url("echo.EchoService/Echo"), + headers={ "content-type": "application/grpc", "requested-status": "7" }, + expected=200) def check(self): # [0] - assert self.results[0].status == 200 assert self.results[0].headers["Grpc-Status"] == ["0"] - # [0] + # [1] - assert self.results[1].status == 200 assert self.results[1].headers["Grpc-Status"] == ["7"]
Clean up existing gRPC bridge test (make it like t_tcpmapping.py)
## Code Before: import json from kat.harness import Query from abstract_tests import AmbassadorTest, ServiceType, EGRPC class AcceptanceGrpcBridgeTest(AmbassadorTest): target: ServiceType def init(self): self.target = EGRPC() def config(self): yield self, self.format(""" --- apiVersion: ambassador/v0 kind: Module name: ambassador config: enable_grpc_http11_bridge: True """) yield self, self.format(""" --- apiVersion: ambassador/v0 kind: Mapping grpc: True prefix: /echo.EchoService/ rewrite: /echo.EchoService/ name: {self.target.path.k8s} service: {self.target.path.k8s} """) def queries(self): # [0] yield Query(self.url("echo.EchoService/Echo"), headers={ "content-type": "application/grpc", "requested-status": "0" }, expected=200) # [1] yield Query(self.url("echo.EchoService/Echo"), headers={ "content-type": "application/grpc", "requested-status": "7" }, expected=200) def check(self): # [0] assert self.results[0].status == 200 assert self.results[0].headers["Grpc-Status"] == ["0"] # [0] assert self.results[1].status == 200 assert self.results[1].headers["Grpc-Status"] == ["7"] ## Instruction: Clean up existing gRPC bridge test (make it like t_tcpmapping.py) ## Code After: from kat.harness import Query from abstract_tests import AmbassadorTest, ServiceType, EGRPC class AcceptanceGrpcBridgeTest(AmbassadorTest): target: ServiceType def init(self): self.target = EGRPC() def config(self): yield self, self.format(""" --- apiVersion: ambassador/v0 kind: Module name: ambassador config: enable_grpc_http11_bridge: True """) yield self, self.format(""" --- apiVersion: ambassador/v0 kind: Mapping grpc: True prefix: /echo.EchoService/ rewrite: /echo.EchoService/ name: {self.target.path.k8s} service: {self.target.path.k8s} """) def queries(self): # [0] yield Query(self.url("echo.EchoService/Echo"), headers={ "content-type": "application/grpc", "requested-status": "0" }, expected=200) # [1] yield Query(self.url("echo.EchoService/Echo"), headers={ "content-type": "application/grpc", "requested-status": "7" }, expected=200) def check(self): # [0] assert self.results[0].headers["Grpc-Status"] == ["0"] # [1] assert self.results[1].headers["Grpc-Status"] == ["7"]
# ... existing code ... from kat.harness import Query # ... modified code ... def queries(self): # [0] yield Query(self.url("echo.EchoService/Echo"), headers={ "content-type": "application/grpc", "requested-status": "0" }, expected=200) # [1] yield Query(self.url("echo.EchoService/Echo"), headers={ "content-type": "application/grpc", "requested-status": "7" }, expected=200) def check(self): # [0] assert self.results[0].headers["Grpc-Status"] == ["0"] # [1] assert self.results[1].headers["Grpc-Status"] == ["7"] # ... rest of the code ...
e87490ea157f4882f644329e4b447f51c0a2acb3
benchmarks/bench_vectorize.py
benchmarks/bench_vectorize.py
import numpy as np from numba import vectorize @vectorize(["float32(float32, float32)", "float64(float64, float64)", "complex64(complex64, complex64)", "complex128(complex128, complex128)"]) def mul(x, y): return x * y class TimeSuite: n = 10000 dtypes = ('float32', 'float64', 'complex64', 'complex128') def setup(self): self.samples = {} self.out = {} for dtype in self.dtypes: self.samples[dtype] = np.linspace(0, 1, self.n, dtype=dtype) self.out[dtype] = np.zeros(self.n, dtype=dtype) def _binary_func(func, dtype): def f(self): func(self.samples[dtype], self.samples[dtype], self.out[dtype]) return f for dtype in dtypes: locals()['time_mul_%s' % dtype] = _binary_func(mul, dtype) del _binary_func
import numpy as np from numba import vectorize @vectorize(["float32(float32, float32)", "float64(float64, float64)", "complex64(complex64, complex64)", "complex128(complex128, complex128)"]) def mul(x, y): return x * y @vectorize(["float32(float32, float32)", "float64(float64, float64)"]) def rel_diff(x, y): # XXX for float32 performance, we should write `np.float32(2)`, but # that's not the natural way to write this code... return 2 * (x - y) / (x + y) class TimeSuite: n = 10000 dtypes = ('float32', 'float64', 'complex64', 'complex128') def setup(self): self.samples = {} self.out = {} for dtype in self.dtypes: self.samples[dtype] = np.linspace(0.1, 1, self.n, dtype=dtype) self.out[dtype] = np.zeros(self.n, dtype=dtype) def _binary_func(func, dtype): def f(self): func(self.samples[dtype], self.samples[dtype], self.out[dtype]) return f for dtype in dtypes: locals()['time_mul_%s' % dtype] = _binary_func(mul, dtype) time_rel_diff_float32 = _binary_func(rel_diff, 'float32') time_rel_diff_float64 = _binary_func(rel_diff, 'float64') del _binary_func
Add a relative difference vectorization benchmark
Add a relative difference vectorization benchmark
Python
bsd-2-clause
gmarkall/numba-benchmark,numba/numba-benchmark
import numpy as np from numba import vectorize @vectorize(["float32(float32, float32)", "float64(float64, float64)", "complex64(complex64, complex64)", "complex128(complex128, complex128)"]) def mul(x, y): return x * y + @vectorize(["float32(float32, float32)", + "float64(float64, float64)"]) + def rel_diff(x, y): + # XXX for float32 performance, we should write `np.float32(2)`, but + # that's not the natural way to write this code... + return 2 * (x - y) / (x + y) + + class TimeSuite: n = 10000 dtypes = ('float32', 'float64', 'complex64', 'complex128') def setup(self): self.samples = {} self.out = {} for dtype in self.dtypes: - self.samples[dtype] = np.linspace(0, 1, self.n, dtype=dtype) + self.samples[dtype] = np.linspace(0.1, 1, self.n, dtype=dtype) self.out[dtype] = np.zeros(self.n, dtype=dtype) def _binary_func(func, dtype): def f(self): func(self.samples[dtype], self.samples[dtype], self.out[dtype]) return f for dtype in dtypes: locals()['time_mul_%s' % dtype] = _binary_func(mul, dtype) + time_rel_diff_float32 = _binary_func(rel_diff, 'float32') + time_rel_diff_float64 = _binary_func(rel_diff, 'float64') + del _binary_func
Add a relative difference vectorization benchmark
## Code Before: import numpy as np from numba import vectorize @vectorize(["float32(float32, float32)", "float64(float64, float64)", "complex64(complex64, complex64)", "complex128(complex128, complex128)"]) def mul(x, y): return x * y class TimeSuite: n = 10000 dtypes = ('float32', 'float64', 'complex64', 'complex128') def setup(self): self.samples = {} self.out = {} for dtype in self.dtypes: self.samples[dtype] = np.linspace(0, 1, self.n, dtype=dtype) self.out[dtype] = np.zeros(self.n, dtype=dtype) def _binary_func(func, dtype): def f(self): func(self.samples[dtype], self.samples[dtype], self.out[dtype]) return f for dtype in dtypes: locals()['time_mul_%s' % dtype] = _binary_func(mul, dtype) del _binary_func ## Instruction: Add a relative difference vectorization benchmark ## Code After: import numpy as np from numba import vectorize @vectorize(["float32(float32, float32)", "float64(float64, float64)", "complex64(complex64, complex64)", "complex128(complex128, complex128)"]) def mul(x, y): return x * y @vectorize(["float32(float32, float32)", "float64(float64, float64)"]) def rel_diff(x, y): # XXX for float32 performance, we should write `np.float32(2)`, but # that's not the natural way to write this code... return 2 * (x - y) / (x + y) class TimeSuite: n = 10000 dtypes = ('float32', 'float64', 'complex64', 'complex128') def setup(self): self.samples = {} self.out = {} for dtype in self.dtypes: self.samples[dtype] = np.linspace(0.1, 1, self.n, dtype=dtype) self.out[dtype] = np.zeros(self.n, dtype=dtype) def _binary_func(func, dtype): def f(self): func(self.samples[dtype], self.samples[dtype], self.out[dtype]) return f for dtype in dtypes: locals()['time_mul_%s' % dtype] = _binary_func(mul, dtype) time_rel_diff_float32 = _binary_func(rel_diff, 'float32') time_rel_diff_float64 = _binary_func(rel_diff, 'float64') del _binary_func
// ... existing code ... @vectorize(["float32(float32, float32)", "float64(float64, float64)"]) def rel_diff(x, y): # XXX for float32 performance, we should write `np.float32(2)`, but # that's not the natural way to write this code... return 2 * (x - y) / (x + y) class TimeSuite: // ... modified code ... self.out = {} for dtype in self.dtypes: self.samples[dtype] = np.linspace(0.1, 1, self.n, dtype=dtype) self.out[dtype] = np.zeros(self.n, dtype=dtype) ... locals()['time_mul_%s' % dtype] = _binary_func(mul, dtype) time_rel_diff_float32 = _binary_func(rel_diff, 'float32') time_rel_diff_float64 = _binary_func(rel_diff, 'float64') del _binary_func // ... rest of the code ...
72045f86b25b396160e1a4c9237e977ed575afb2
apps/catalogue/constants.py
apps/catalogue/constants.py
from django.utils.translation import ugettext_lazy as _ LICENSES = { 'http://creativecommons.org/licenses/by-sa/3.0/': { 'icon': 'cc-by-sa', 'description': _('Creative Commons Attribution-ShareAlike 3.0 Unported'), }, } # Those will be generated only for books with own HTML. EBOOK_FORMATS_WITHOUT_CHILDREN = ['txt', 'fb2'] # Those will be generated for all books. EBOOK_FORMATS_WITH_CHILDREN = ['pdf', 'epub', 'mobi'] # Those will be generated when inherited cover changes. EBOOK_FORMATS_WITH_COVERS = ['pdf', 'epub', 'mobi'] EBOOK_FORMATS = EBOOK_FORMATS_WITHOUT_CHILDREN + EBOOK_FORMATS_WITH_CHILDREN
from django.utils.translation import ugettext_lazy as _ LICENSES = { 'http://creativecommons.org/licenses/by-sa/3.0/': { 'icon': 'cc-by-sa', 'description': _('Creative Commons Attribution-ShareAlike 3.0 Unported'), }, } LICENSES['http://creativecommons.org/licenses/by-sa/3.0/deed.pl'] = \ LICENSES['http://creativecommons.org/licenses/by-sa/3.0/'] # Those will be generated only for books with own HTML. EBOOK_FORMATS_WITHOUT_CHILDREN = ['txt', 'fb2'] # Those will be generated for all books. EBOOK_FORMATS_WITH_CHILDREN = ['pdf', 'epub', 'mobi'] # Those will be generated when inherited cover changes. EBOOK_FORMATS_WITH_COVERS = ['pdf', 'epub', 'mobi'] EBOOK_FORMATS = EBOOK_FORMATS_WITHOUT_CHILDREN + EBOOK_FORMATS_WITH_CHILDREN
Support for 'deed.pl' license URL.
Support for 'deed.pl' license URL.
Python
agpl-3.0
fnp/wolnelektury,fnp/wolnelektury,fnp/wolnelektury,fnp/wolnelektury
from django.utils.translation import ugettext_lazy as _ LICENSES = { 'http://creativecommons.org/licenses/by-sa/3.0/': { 'icon': 'cc-by-sa', 'description': _('Creative Commons Attribution-ShareAlike 3.0 Unported'), }, } + LICENSES['http://creativecommons.org/licenses/by-sa/3.0/deed.pl'] = \ + LICENSES['http://creativecommons.org/licenses/by-sa/3.0/'] # Those will be generated only for books with own HTML. EBOOK_FORMATS_WITHOUT_CHILDREN = ['txt', 'fb2'] # Those will be generated for all books. EBOOK_FORMATS_WITH_CHILDREN = ['pdf', 'epub', 'mobi'] # Those will be generated when inherited cover changes. EBOOK_FORMATS_WITH_COVERS = ['pdf', 'epub', 'mobi'] EBOOK_FORMATS = EBOOK_FORMATS_WITHOUT_CHILDREN + EBOOK_FORMATS_WITH_CHILDREN
Support for 'deed.pl' license URL.
## Code Before: from django.utils.translation import ugettext_lazy as _ LICENSES = { 'http://creativecommons.org/licenses/by-sa/3.0/': { 'icon': 'cc-by-sa', 'description': _('Creative Commons Attribution-ShareAlike 3.0 Unported'), }, } # Those will be generated only for books with own HTML. EBOOK_FORMATS_WITHOUT_CHILDREN = ['txt', 'fb2'] # Those will be generated for all books. EBOOK_FORMATS_WITH_CHILDREN = ['pdf', 'epub', 'mobi'] # Those will be generated when inherited cover changes. EBOOK_FORMATS_WITH_COVERS = ['pdf', 'epub', 'mobi'] EBOOK_FORMATS = EBOOK_FORMATS_WITHOUT_CHILDREN + EBOOK_FORMATS_WITH_CHILDREN ## Instruction: Support for 'deed.pl' license URL. ## Code After: from django.utils.translation import ugettext_lazy as _ LICENSES = { 'http://creativecommons.org/licenses/by-sa/3.0/': { 'icon': 'cc-by-sa', 'description': _('Creative Commons Attribution-ShareAlike 3.0 Unported'), }, } LICENSES['http://creativecommons.org/licenses/by-sa/3.0/deed.pl'] = \ LICENSES['http://creativecommons.org/licenses/by-sa/3.0/'] # Those will be generated only for books with own HTML. EBOOK_FORMATS_WITHOUT_CHILDREN = ['txt', 'fb2'] # Those will be generated for all books. EBOOK_FORMATS_WITH_CHILDREN = ['pdf', 'epub', 'mobi'] # Those will be generated when inherited cover changes. EBOOK_FORMATS_WITH_COVERS = ['pdf', 'epub', 'mobi'] EBOOK_FORMATS = EBOOK_FORMATS_WITHOUT_CHILDREN + EBOOK_FORMATS_WITH_CHILDREN
... }, } LICENSES['http://creativecommons.org/licenses/by-sa/3.0/deed.pl'] = \ LICENSES['http://creativecommons.org/licenses/by-sa/3.0/'] # Those will be generated only for books with own HTML. ...
8869eba1f74e677d1802aad0cc2592344ab81000
podium/talks/models.py
podium/talks/models.py
from django.db import models from django.urls import reverse TALK_STATUS_CHOICES = ( ('S', 'Submitted'), ('A', 'Approved'), ('R', 'Rejected'), ('C', 'Confirmed'), ) class Talk(models.Model): speaker_name = models.CharField(max_length=1000) speaker_email = models.CharField(max_length=1000) title = models.CharField(max_length=1000) description = models.TextField() sessions_available = models.ManyToManyField( 'Session', related_name='talks_available') status = models.CharField( max_length=1, choices=TALK_STATUS_CHOICES, default='S') def get_absolute_url(self): return reverse('talks-talks-id', args=[self.id]) def __str__(self): return self.speaker_name class Session(models.Model): date = models.DateField() description = models.TextField( blank=True, help_text='Any special theme or info about the session.') def __str__(self): return '{} - {} '.format(self.date, self.description) def approved_talks(self): sets = [ self.talks_available.filter(status=status) for status in ('A', 'C') ] return sets[0].union(sets[1]) def get_absolute_url(self): return reverse('talks-sessions-id', args=[self.id])
from django.db import models from django.urls import reverse TALK_STATUS_CHOICES = ( ('S', 'Submitted'), ('A', 'Approved'), ('R', 'Rejected'), ('C', 'Confirmed'), ) class Talk(models.Model): speaker_name = models.CharField(max_length=1000) speaker_email = models.CharField(max_length=1000) title = models.CharField(max_length=1000) description = models.TextField() sessions_available = models.ManyToManyField( 'Session', related_name='talks_available') status = models.CharField( max_length=1, choices=TALK_STATUS_CHOICES, default='S') def get_absolute_url(self): return reverse('talks-talks-id', args=[self.id]) def __str__(self): return self.speaker_name class Session(models.Model): date = models.DateField() description = models.TextField( blank=True, help_text='Any special theme or info about the session.') def __str__(self): return '{} - {} '.format(self.date, self.description) def approved_talks(self): return self.talks_available.filter(status__in=('A', 'C')) def get_absolute_url(self): return reverse('talks-sessions-id', args=[self.id])
Use a filter field lookup
Use a filter field lookup Looks like I forgot to do this when JR suggested it.
Python
mit
pyatl/podium-django,pyatl/podium-django,pyatl/podium-django
from django.db import models from django.urls import reverse TALK_STATUS_CHOICES = ( ('S', 'Submitted'), ('A', 'Approved'), ('R', 'Rejected'), ('C', 'Confirmed'), ) class Talk(models.Model): speaker_name = models.CharField(max_length=1000) speaker_email = models.CharField(max_length=1000) title = models.CharField(max_length=1000) description = models.TextField() sessions_available = models.ManyToManyField( 'Session', related_name='talks_available') status = models.CharField( max_length=1, choices=TALK_STATUS_CHOICES, default='S') def get_absolute_url(self): return reverse('talks-talks-id', args=[self.id]) def __str__(self): return self.speaker_name class Session(models.Model): date = models.DateField() description = models.TextField( blank=True, help_text='Any special theme or info about the session.') def __str__(self): return '{} - {} '.format(self.date, self.description) def approved_talks(self): - sets = [ - self.talks_available.filter(status=status) for status in ('A', 'C') + return self.talks_available.filter(status__in=('A', 'C')) - ] - return sets[0].union(sets[1]) def get_absolute_url(self): return reverse('talks-sessions-id', args=[self.id])
Use a filter field lookup
## Code Before: from django.db import models from django.urls import reverse TALK_STATUS_CHOICES = ( ('S', 'Submitted'), ('A', 'Approved'), ('R', 'Rejected'), ('C', 'Confirmed'), ) class Talk(models.Model): speaker_name = models.CharField(max_length=1000) speaker_email = models.CharField(max_length=1000) title = models.CharField(max_length=1000) description = models.TextField() sessions_available = models.ManyToManyField( 'Session', related_name='talks_available') status = models.CharField( max_length=1, choices=TALK_STATUS_CHOICES, default='S') def get_absolute_url(self): return reverse('talks-talks-id', args=[self.id]) def __str__(self): return self.speaker_name class Session(models.Model): date = models.DateField() description = models.TextField( blank=True, help_text='Any special theme or info about the session.') def __str__(self): return '{} - {} '.format(self.date, self.description) def approved_talks(self): sets = [ self.talks_available.filter(status=status) for status in ('A', 'C') ] return sets[0].union(sets[1]) def get_absolute_url(self): return reverse('talks-sessions-id', args=[self.id]) ## Instruction: Use a filter field lookup ## Code After: from django.db import models from django.urls import reverse TALK_STATUS_CHOICES = ( ('S', 'Submitted'), ('A', 'Approved'), ('R', 'Rejected'), ('C', 'Confirmed'), ) class Talk(models.Model): speaker_name = models.CharField(max_length=1000) speaker_email = models.CharField(max_length=1000) title = models.CharField(max_length=1000) description = models.TextField() sessions_available = models.ManyToManyField( 'Session', related_name='talks_available') status = models.CharField( max_length=1, choices=TALK_STATUS_CHOICES, default='S') def get_absolute_url(self): return reverse('talks-talks-id', args=[self.id]) def __str__(self): return self.speaker_name class Session(models.Model): date = models.DateField() description = models.TextField( blank=True, help_text='Any special theme or info about the session.') def __str__(self): return '{} - {} '.format(self.date, self.description) def approved_talks(self): return self.talks_available.filter(status__in=('A', 'C')) def get_absolute_url(self): return reverse('talks-sessions-id', args=[self.id])
# ... existing code ... def approved_talks(self): return self.talks_available.filter(status__in=('A', 'C')) def get_absolute_url(self): # ... rest of the code ...
929909513e71282de388cf4e93476ba614e6c0c5
Malcom/feeds/malwaredomains.py
Malcom/feeds/malwaredomains.py
import urllib2 import re from Malcom.model.datatypes import Hostname, Evil from feed import Feed import Malcom.auxiliary.toolbox as toolbox class MalwareDomains(Feed): def __init__(self, name): super(MalwareDomains, self).__init__(name) self.source = "http://mirror1.malwaredomains.com/files/domains.txt" self.description = "Malware domains blocklist" self.confidence = 50 self.name = "MalwareDomains" def update(self): self.update_lines() def analyze(self, line): if line.startswith('#') or line.startswith('\n'): return splitted_mdl = line.split('\t') # 20151201 agasi-story.info malicious blog.dynamoo.com 20131130 20121201 20120521 20110217 # Create the new hostname and store it in the DB hostname = Hostname(hostname=splitted_mdl[2]) if hostname['value'] == None: return # hostname not found evil = Evil() evil['value'] = "Malware domain blocklist (%s)" % hostname['value'] evil['tags'] = ['malwaredomains', splitted_mdl[3]] evil['reference'] = splitted_mdl[4] return hostname, evil
import urllib2 import re from Malcom.model.datatypes import Hostname, Evil from feed import Feed import Malcom.auxiliary.toolbox as toolbox class MalwareDomains(Feed): def __init__(self, name): super(MalwareDomains, self).__init__(name) self.source = "http://mirror1.malwaredomains.com/files/domains.txt" self.description = "Malware domains blocklist" self.confidence = 50 self.name = "MalwareDomains" def update(self): self.update_lines() def analyze(self, line): if line.startswith('#') or line.startswith('\n'): return splitted_mdl = line.split('\t') # 20151201 agasi-story.info malicious blog.dynamoo.com 20131130 20121201 20120521 20110217 # Create the new hostname and store it in the DB hostname = Hostname(hostname=splitted_mdl[2]) if hostname['value'] == None: return # hostname not found evil = Evil() evil['value'] = "Malware domain blocklist (%s)" % hostname['value'] evil['tags'] = ['malwaredomains', re.sub(r'[^\w]', '', splitted_mdl[3])] evil['reference'] = splitted_mdl[4] return hostname, evil
Deal with MalwareDomains non-ASCII characters
Deal with MalwareDomains non-ASCII characters
Python
apache-2.0
yeti-platform/yeti,yeti-platform/yeti,yeti-platform/yeti,yeti-platform/yeti
import urllib2 import re from Malcom.model.datatypes import Hostname, Evil from feed import Feed import Malcom.auxiliary.toolbox as toolbox class MalwareDomains(Feed): def __init__(self, name): super(MalwareDomains, self).__init__(name) self.source = "http://mirror1.malwaredomains.com/files/domains.txt" self.description = "Malware domains blocklist" self.confidence = 50 self.name = "MalwareDomains" def update(self): self.update_lines() def analyze(self, line): if line.startswith('#') or line.startswith('\n'): return splitted_mdl = line.split('\t') # 20151201 agasi-story.info malicious blog.dynamoo.com 20131130 20121201 20120521 20110217 # Create the new hostname and store it in the DB hostname = Hostname(hostname=splitted_mdl[2]) if hostname['value'] == None: return # hostname not found evil = Evil() evil['value'] = "Malware domain blocklist (%s)" % hostname['value'] - evil['tags'] = ['malwaredomains', splitted_mdl[3]] + evil['tags'] = ['malwaredomains', re.sub(r'[^\w]', '', splitted_mdl[3])] evil['reference'] = splitted_mdl[4] return hostname, evil
Deal with MalwareDomains non-ASCII characters
## Code Before: import urllib2 import re from Malcom.model.datatypes import Hostname, Evil from feed import Feed import Malcom.auxiliary.toolbox as toolbox class MalwareDomains(Feed): def __init__(self, name): super(MalwareDomains, self).__init__(name) self.source = "http://mirror1.malwaredomains.com/files/domains.txt" self.description = "Malware domains blocklist" self.confidence = 50 self.name = "MalwareDomains" def update(self): self.update_lines() def analyze(self, line): if line.startswith('#') or line.startswith('\n'): return splitted_mdl = line.split('\t') # 20151201 agasi-story.info malicious blog.dynamoo.com 20131130 20121201 20120521 20110217 # Create the new hostname and store it in the DB hostname = Hostname(hostname=splitted_mdl[2]) if hostname['value'] == None: return # hostname not found evil = Evil() evil['value'] = "Malware domain blocklist (%s)" % hostname['value'] evil['tags'] = ['malwaredomains', splitted_mdl[3]] evil['reference'] = splitted_mdl[4] return hostname, evil ## Instruction: Deal with MalwareDomains non-ASCII characters ## Code After: import urllib2 import re from Malcom.model.datatypes import Hostname, Evil from feed import Feed import Malcom.auxiliary.toolbox as toolbox class MalwareDomains(Feed): def __init__(self, name): super(MalwareDomains, self).__init__(name) self.source = "http://mirror1.malwaredomains.com/files/domains.txt" self.description = "Malware domains blocklist" self.confidence = 50 self.name = "MalwareDomains" def update(self): self.update_lines() def analyze(self, line): if line.startswith('#') or line.startswith('\n'): return splitted_mdl = line.split('\t') # 20151201 agasi-story.info malicious blog.dynamoo.com 20131130 20121201 20120521 20110217 # Create the new hostname and store it in the DB hostname = Hostname(hostname=splitted_mdl[2]) if hostname['value'] == None: return # hostname not found evil = Evil() evil['value'] = "Malware domain blocklist (%s)" % hostname['value'] evil['tags'] = ['malwaredomains', re.sub(r'[^\w]', '', splitted_mdl[3])] evil['reference'] = splitted_mdl[4] return hostname, evil
# ... existing code ... evil = Evil() evil['value'] = "Malware domain blocklist (%s)" % hostname['value'] evil['tags'] = ['malwaredomains', re.sub(r'[^\w]', '', splitted_mdl[3])] evil['reference'] = splitted_mdl[4] # ... rest of the code ...
c3745e7017c1788f4633d09ef4d29a37018b53d3
populus/cli/main.py
populus/cli/main.py
import click @click.group() def main(): """ Populus """ pass
import click CONTEXT_SETTINGS = dict( # Support -h as a shortcut for --help help_option_names=['-h', '--help'], ) @click.group(context_settings=CONTEXT_SETTINGS) def main(): """ Populus """ pass
Support -h as a shortcut for --help
CLI: Support -h as a shortcut for --help
Python
mit
pipermerriam/populus,euri10/populus,euri10/populus,pipermerriam/populus,euri10/populus
import click - @click.group() + CONTEXT_SETTINGS = dict( + # Support -h as a shortcut for --help + help_option_names=['-h', '--help'], + ) + + + @click.group(context_settings=CONTEXT_SETTINGS) def main(): """ Populus """ pass
Support -h as a shortcut for --help
## Code Before: import click @click.group() def main(): """ Populus """ pass ## Instruction: Support -h as a shortcut for --help ## Code After: import click CONTEXT_SETTINGS = dict( # Support -h as a shortcut for --help help_option_names=['-h', '--help'], ) @click.group(context_settings=CONTEXT_SETTINGS) def main(): """ Populus """ pass
// ... existing code ... CONTEXT_SETTINGS = dict( # Support -h as a shortcut for --help help_option_names=['-h', '--help'], ) @click.group(context_settings=CONTEXT_SETTINGS) def main(): """ // ... rest of the code ...
ff725b4ae24c58cb126c1d49ce58a69d9b32d3b0
app/soc/models/timeline.py
app/soc/models/timeline.py
from google.appengine.ext import db from django.utils.translation import ugettext from soc.models import linkable class Timeline(linkable.Linkable): """The Timeline Model, representing the timeline for a Program. """ program_start = db.DateTimeProperty( verbose_name=ugettext('Program Start date')) program_end = db.DateTimeProperty( verbose_name=ugettext('Program End date')) accepted_organization_announced_deadline = db.DateTimeProperty( verbose_name=ugettext('Accepted Organizations Announced Deadline')) student_signup_start = db.DateTimeProperty( verbose_name=ugettext('Student Signup Start date')) student_signup_end = db.DateTimeProperty( verbose_name=ugettext('Student Signup End date'))
from google.appengine.ext import db from django.utils.translation import ugettext from soc.models import linkable class Timeline(linkable.Linkable): """The Timeline Model, representing the timeline for a Program. """ program_start = db.DateTimeProperty( verbose_name=ugettext('Program Start date')) program_end = db.DateTimeProperty( verbose_name=ugettext('Program End date')) program_end.help_text = ugettext( 'After this date no data (such as profiles and forms) can be changed.') accepted_organization_announced_deadline = db.DateTimeProperty( verbose_name=ugettext('Accepted Organizations Announced Deadline')) student_signup_start = db.DateTimeProperty( verbose_name=ugettext('Student Signup Start date')) student_signup_end = db.DateTimeProperty( verbose_name=ugettext('Student Signup End date'))
Add help text for program_end date.
Add help text for program_end date. Fixes 1411.
Python
apache-2.0
rhyolight/nupic.son,rhyolight/nupic.son,rhyolight/nupic.son
from google.appengine.ext import db from django.utils.translation import ugettext from soc.models import linkable class Timeline(linkable.Linkable): """The Timeline Model, representing the timeline for a Program. """ program_start = db.DateTimeProperty( verbose_name=ugettext('Program Start date')) program_end = db.DateTimeProperty( verbose_name=ugettext('Program End date')) + program_end.help_text = ugettext( + 'After this date no data (such as profiles and forms) can be changed.') accepted_organization_announced_deadline = db.DateTimeProperty( verbose_name=ugettext('Accepted Organizations Announced Deadline')) student_signup_start = db.DateTimeProperty( verbose_name=ugettext('Student Signup Start date')) student_signup_end = db.DateTimeProperty( verbose_name=ugettext('Student Signup End date'))
Add help text for program_end date.
## Code Before: from google.appengine.ext import db from django.utils.translation import ugettext from soc.models import linkable class Timeline(linkable.Linkable): """The Timeline Model, representing the timeline for a Program. """ program_start = db.DateTimeProperty( verbose_name=ugettext('Program Start date')) program_end = db.DateTimeProperty( verbose_name=ugettext('Program End date')) accepted_organization_announced_deadline = db.DateTimeProperty( verbose_name=ugettext('Accepted Organizations Announced Deadline')) student_signup_start = db.DateTimeProperty( verbose_name=ugettext('Student Signup Start date')) student_signup_end = db.DateTimeProperty( verbose_name=ugettext('Student Signup End date')) ## Instruction: Add help text for program_end date. ## Code After: from google.appengine.ext import db from django.utils.translation import ugettext from soc.models import linkable class Timeline(linkable.Linkable): """The Timeline Model, representing the timeline for a Program. """ program_start = db.DateTimeProperty( verbose_name=ugettext('Program Start date')) program_end = db.DateTimeProperty( verbose_name=ugettext('Program End date')) program_end.help_text = ugettext( 'After this date no data (such as profiles and forms) can be changed.') accepted_organization_announced_deadline = db.DateTimeProperty( verbose_name=ugettext('Accepted Organizations Announced Deadline')) student_signup_start = db.DateTimeProperty( verbose_name=ugettext('Student Signup Start date')) student_signup_end = db.DateTimeProperty( verbose_name=ugettext('Student Signup End date'))
// ... existing code ... program_end = db.DateTimeProperty( verbose_name=ugettext('Program End date')) program_end.help_text = ugettext( 'After this date no data (such as profiles and forms) can be changed.') accepted_organization_announced_deadline = db.DateTimeProperty( // ... rest of the code ...
fc94ac89d2f602c381f4c882ec963995f3ce3043
cla_frontend/apps/core/context_processors.py
cla_frontend/apps/core/context_processors.py
from django.conf import settings def globals(request): context = { 'app_title': 'Civil Legal Advice', 'proposition_title': 'Civil Legal Advice', 'phase': 'alpha', 'product_type': 'service', 'feedback_url': '#', 'ga_id': '', 'raven_config_site': settings.RAVEN_CONFIG['site'] or '' } if hasattr(request, 'zone') and request.zone: context['app_base_template'] = '%s/base.html' % request.zone['name'] context['zone'] = request.zone return context
from django.conf import settings def globals(request): context = { 'app_title': 'Civil Legal Advice', 'proposition_title': 'Civil Legal Advice', 'phase': 'alpha', 'product_type': 'service', 'feedback_url': '#', 'ga_id': '', 'raven_config_site': settings.RAVEN_CONFIG['site'] or '', 'socketio_server_url': settings.SOCKETIO_SERVER_URL } if hasattr(request, 'zone') and request.zone: context['app_base_template'] = '%s/base.html' % request.zone['name'] context['zone'] = request.zone return context
Make socketio server url a global context variable in Django
Make socketio server url a global context variable in Django
Python
mit
ministryofjustice/cla_frontend,ministryofjustice/cla_frontend,ministryofjustice/cla_frontend,ministryofjustice/cla_frontend
from django.conf import settings def globals(request): context = { 'app_title': 'Civil Legal Advice', 'proposition_title': 'Civil Legal Advice', 'phase': 'alpha', 'product_type': 'service', 'feedback_url': '#', 'ga_id': '', - 'raven_config_site': settings.RAVEN_CONFIG['site'] or '' + 'raven_config_site': settings.RAVEN_CONFIG['site'] or '', + 'socketio_server_url': settings.SOCKETIO_SERVER_URL } if hasattr(request, 'zone') and request.zone: context['app_base_template'] = '%s/base.html' % request.zone['name'] context['zone'] = request.zone return context
Make socketio server url a global context variable in Django
## Code Before: from django.conf import settings def globals(request): context = { 'app_title': 'Civil Legal Advice', 'proposition_title': 'Civil Legal Advice', 'phase': 'alpha', 'product_type': 'service', 'feedback_url': '#', 'ga_id': '', 'raven_config_site': settings.RAVEN_CONFIG['site'] or '' } if hasattr(request, 'zone') and request.zone: context['app_base_template'] = '%s/base.html' % request.zone['name'] context['zone'] = request.zone return context ## Instruction: Make socketio server url a global context variable in Django ## Code After: from django.conf import settings def globals(request): context = { 'app_title': 'Civil Legal Advice', 'proposition_title': 'Civil Legal Advice', 'phase': 'alpha', 'product_type': 'service', 'feedback_url': '#', 'ga_id': '', 'raven_config_site': settings.RAVEN_CONFIG['site'] or '', 'socketio_server_url': settings.SOCKETIO_SERVER_URL } if hasattr(request, 'zone') and request.zone: context['app_base_template'] = '%s/base.html' % request.zone['name'] context['zone'] = request.zone return context
# ... existing code ... 'feedback_url': '#', 'ga_id': '', 'raven_config_site': settings.RAVEN_CONFIG['site'] or '', 'socketio_server_url': settings.SOCKETIO_SERVER_URL } # ... rest of the code ...
0bf00b40e84a5c5fbcdbeb7b81911998e3f1081a
src/idea/tests/smoke_tests.py
src/idea/tests/smoke_tests.py
import os from django.utils import timezone from django_webtest import WebTest from exam.decorators import fixture from exam.cases import Exam from django.core.urlresolvers import reverse class SmokeTest(Exam, WebTest): csrf_checks = False fixtures = ['state'] @fixture def user(self): try: from collab.django_factories import UserF return UserF(username="[email protected]", person__title='') except ImportError: from django.contrib.auth.models import User user = User() user.username = "[email protected]" user.first_name = 'first' user.last_name = 'last' user.email = '"[email protected]"' user.password = 'pbkdf2_sha256$10000$ggAKkiHobFL8$xQzwPeHNX1vWr9uNmZ/gKbd17uLGZVM8QNcgmaIEAUs=' user.is_staff = False user.is_active = True user.is_superuser = False user.last_login = timezone.now() user.date_joined = timezone.now() user.save() return user def get(self, url): return self.app.get(url, user=self.user) def test_idea_home(self): page = self.get(reverse('idea:idea_list')) self.assertEquals(200, page.status_code)
import os from django.utils import timezone from django_webtest import WebTest from exam.decorators import fixture from exam.cases import Exam from django.core.urlresolvers import reverse from django.contrib.auth.models import User class SmokeTest(Exam, WebTest): csrf_checks = False fixtures = ['state', 'core-test-fixtures'] @fixture def user(self): user = User.objects.get(username="[email protected]") return user def get(self, url): return self.app.get(url, user=self.user) def test_idea_home(self): page = self.get(reverse('idea:idea_list')) self.assertEquals(200, page.status_code)
Use fixtures for smoke tests
Use fixtures for smoke tests
Python
cc0-1.0
cmc333333/idea-box,m3brown/idea-box,18F/idea-box,cmc333333/idea-box,CapeSepias/idea-box,geomapdev/idea-box,CapeSepias/idea-box,cmc333333/idea-box,18F/idea-box,geomapdev/idea-box,geomapdev/idea-box,18F/idea-box,CapeSepias/idea-box,m3brown/idea-box
import os from django.utils import timezone from django_webtest import WebTest from exam.decorators import fixture from exam.cases import Exam from django.core.urlresolvers import reverse + from django.contrib.auth.models import User class SmokeTest(Exam, WebTest): csrf_checks = False - fixtures = ['state'] + fixtures = ['state', 'core-test-fixtures'] @fixture def user(self): - try: - from collab.django_factories import UserF - return UserF(username="[email protected]", person__title='') - except ImportError: - from django.contrib.auth.models import User - user = User() - user.username = "[email protected]" + user = User.objects.get(username="[email protected]") - user.first_name = 'first' - user.last_name = 'last' - user.email = '"[email protected]"' - user.password = 'pbkdf2_sha256$10000$ggAKkiHobFL8$xQzwPeHNX1vWr9uNmZ/gKbd17uLGZVM8QNcgmaIEAUs=' - user.is_staff = False - user.is_active = True - user.is_superuser = False - user.last_login = timezone.now() - user.date_joined = timezone.now() - user.save() - return user + return user def get(self, url): return self.app.get(url, user=self.user) def test_idea_home(self): page = self.get(reverse('idea:idea_list')) self.assertEquals(200, page.status_code)
Use fixtures for smoke tests
## Code Before: import os from django.utils import timezone from django_webtest import WebTest from exam.decorators import fixture from exam.cases import Exam from django.core.urlresolvers import reverse class SmokeTest(Exam, WebTest): csrf_checks = False fixtures = ['state'] @fixture def user(self): try: from collab.django_factories import UserF return UserF(username="[email protected]", person__title='') except ImportError: from django.contrib.auth.models import User user = User() user.username = "[email protected]" user.first_name = 'first' user.last_name = 'last' user.email = '"[email protected]"' user.password = 'pbkdf2_sha256$10000$ggAKkiHobFL8$xQzwPeHNX1vWr9uNmZ/gKbd17uLGZVM8QNcgmaIEAUs=' user.is_staff = False user.is_active = True user.is_superuser = False user.last_login = timezone.now() user.date_joined = timezone.now() user.save() return user def get(self, url): return self.app.get(url, user=self.user) def test_idea_home(self): page = self.get(reverse('idea:idea_list')) self.assertEquals(200, page.status_code) ## Instruction: Use fixtures for smoke tests ## Code After: import os from django.utils import timezone from django_webtest import WebTest from exam.decorators import fixture from exam.cases import Exam from django.core.urlresolvers import reverse from django.contrib.auth.models import User class SmokeTest(Exam, WebTest): csrf_checks = False fixtures = ['state', 'core-test-fixtures'] @fixture def user(self): user = User.objects.get(username="[email protected]") return user def get(self, url): return self.app.get(url, user=self.user) def test_idea_home(self): page = self.get(reverse('idea:idea_list')) self.assertEquals(200, page.status_code)
... from exam.cases import Exam from django.core.urlresolvers import reverse from django.contrib.auth.models import User ... class SmokeTest(Exam, WebTest): csrf_checks = False fixtures = ['state', 'core-test-fixtures'] @fixture def user(self): user = User.objects.get(username="[email protected]") return user def get(self, url): ...
bf2ace8bd6cb0c492ff4347f9c2fe10a003abaff
sqlalchemy_redshift/__init__.py
sqlalchemy_redshift/__init__.py
from pkg_resources import get_distribution, parse_version try: import psycopg2 # noqa: F401 if get_distribution('psycopg2').parsed_version < parse_version('2.5'): raise ImportError('Minimum required version for psycopg2 is 2.5') except ImportError: raise ImportError( 'No module named psycopg2. Please install either ' 'psycopg2 or psycopg2-binary package for CPython ' 'or psycopg2cffi for Pypy.' ) __version__ = get_distribution('sqlalchemy-redshift').version from sqlalchemy.dialects import registry registry.register("redshift", "sqlalchemy_redshift.dialect", "RedshiftDialect") registry.register( "redshift.psycopg2", "sqlalchemy_redshift.dialect", "RedshiftDialect" )
from pkg_resources import DistributionNotFound, get_distribution, parse_version try: import psycopg2 # noqa: F401 except ImportError: raise ImportError( 'No module named psycopg2. Please install either ' 'psycopg2 or psycopg2-binary package for CPython ' 'or psycopg2cffi for Pypy.' ) from None for package in ['psycopg2', 'psycopg2-binary', 'psycopg2cffi']: try: if get_distribution(package).parsed_version < parse_version('2.5'): raise ImportError('Minimum required version for psycopg2 is 2.5') break except DistributionNotFound: pass else: raise ImportError( 'A module was found named psycopg2, ' 'but the version of it could not be checked ' 'as it was neither the Python package psycopg2, ' 'psycopg2-binary or psycopg2cffi.' ) __version__ = get_distribution('sqlalchemy-redshift').version from sqlalchemy.dialects import registry registry.register("redshift", "sqlalchemy_redshift.dialect", "RedshiftDialect") registry.register( "redshift.psycopg2", "sqlalchemy_redshift.dialect", "RedshiftDialect" )
Check the version of any of the supported Psycopg2 packages
Check the version of any of the supported Psycopg2 packages A check was introduced in commit 8e0c4857a1c08f257b95d3b1ee5f6eb795d55cdc which would check what version of the 'psycopg2' Python (pip) package was installed as the dependency was removed from setup.py. The check would however only check the 'psycopg2' package and not the other two supported providers of the psycopg2 module, which meant importing the sqlalchemy_redshift module would throw an exception, even though they were installed. This changes the check to check for either of the three supported psycopg2 packages and throws an exception if any of them fail to validate.
Python
mit
sqlalchemy-redshift/sqlalchemy-redshift,graingert/redshift_sqlalchemy,sqlalchemy-redshift/sqlalchemy-redshift
- from pkg_resources import get_distribution, parse_version + from pkg_resources import DistributionNotFound, get_distribution, parse_version try: import psycopg2 # noqa: F401 - if get_distribution('psycopg2').parsed_version < parse_version('2.5'): - raise ImportError('Minimum required version for psycopg2 is 2.5') except ImportError: raise ImportError( 'No module named psycopg2. Please install either ' 'psycopg2 or psycopg2-binary package for CPython ' 'or psycopg2cffi for Pypy.' + ) from None + + for package in ['psycopg2', 'psycopg2-binary', 'psycopg2cffi']: + try: + if get_distribution(package).parsed_version < parse_version('2.5'): + raise ImportError('Minimum required version for psycopg2 is 2.5') + break + except DistributionNotFound: + pass + else: + raise ImportError( + 'A module was found named psycopg2, ' + 'but the version of it could not be checked ' + 'as it was neither the Python package psycopg2, ' + 'psycopg2-binary or psycopg2cffi.' ) __version__ = get_distribution('sqlalchemy-redshift').version from sqlalchemy.dialects import registry registry.register("redshift", "sqlalchemy_redshift.dialect", "RedshiftDialect") registry.register( "redshift.psycopg2", "sqlalchemy_redshift.dialect", "RedshiftDialect" )
Check the version of any of the supported Psycopg2 packages
## Code Before: from pkg_resources import get_distribution, parse_version try: import psycopg2 # noqa: F401 if get_distribution('psycopg2').parsed_version < parse_version('2.5'): raise ImportError('Minimum required version for psycopg2 is 2.5') except ImportError: raise ImportError( 'No module named psycopg2. Please install either ' 'psycopg2 or psycopg2-binary package for CPython ' 'or psycopg2cffi for Pypy.' ) __version__ = get_distribution('sqlalchemy-redshift').version from sqlalchemy.dialects import registry registry.register("redshift", "sqlalchemy_redshift.dialect", "RedshiftDialect") registry.register( "redshift.psycopg2", "sqlalchemy_redshift.dialect", "RedshiftDialect" ) ## Instruction: Check the version of any of the supported Psycopg2 packages ## Code After: from pkg_resources import DistributionNotFound, get_distribution, parse_version try: import psycopg2 # noqa: F401 except ImportError: raise ImportError( 'No module named psycopg2. Please install either ' 'psycopg2 or psycopg2-binary package for CPython ' 'or psycopg2cffi for Pypy.' ) from None for package in ['psycopg2', 'psycopg2-binary', 'psycopg2cffi']: try: if get_distribution(package).parsed_version < parse_version('2.5'): raise ImportError('Minimum required version for psycopg2 is 2.5') break except DistributionNotFound: pass else: raise ImportError( 'A module was found named psycopg2, ' 'but the version of it could not be checked ' 'as it was neither the Python package psycopg2, ' 'psycopg2-binary or psycopg2cffi.' ) __version__ = get_distribution('sqlalchemy-redshift').version from sqlalchemy.dialects import registry registry.register("redshift", "sqlalchemy_redshift.dialect", "RedshiftDialect") registry.register( "redshift.psycopg2", "sqlalchemy_redshift.dialect", "RedshiftDialect" )
// ... existing code ... from pkg_resources import DistributionNotFound, get_distribution, parse_version try: import psycopg2 # noqa: F401 except ImportError: raise ImportError( // ... modified code ... 'psycopg2 or psycopg2-binary package for CPython ' 'or psycopg2cffi for Pypy.' ) from None for package in ['psycopg2', 'psycopg2-binary', 'psycopg2cffi']: try: if get_distribution(package).parsed_version < parse_version('2.5'): raise ImportError('Minimum required version for psycopg2 is 2.5') break except DistributionNotFound: pass else: raise ImportError( 'A module was found named psycopg2, ' 'but the version of it could not be checked ' 'as it was neither the Python package psycopg2, ' 'psycopg2-binary or psycopg2cffi.' ) // ... rest of the code ...
156d62f15963bc95f52db7eb1493fad6890e2fc7
dadi/__init__.py
dadi/__init__.py
import numpy # This gives a nicer printout for masked arrays. numpy.ma.default_real_fill_value = numpy.nan import IO import Integration import PhiManip import SFS import ms try: import os __DIRECTORY__ = os.path.dirname(IO.__file__) __svn_file__ = os.path.join(__DIRECTORY__, 'svnversion') __SVNVERSION__ = file(__svn_file__).read().strip() except: __SVNVERSION__ = 'Unknown'
import numpy # This gives a nicer printout for masked arrays. numpy.ma.default_real_fill_value = numpy.nan import IO import Integration import PhiManip import SFS import ms import Plotting try: import os __DIRECTORY__ = os.path.dirname(IO.__file__) __svn_file__ = os.path.join(__DIRECTORY__, 'svnversion') __SVNVERSION__ = file(__svn_file__).read().strip() except: __SVNVERSION__ = 'Unknown'
Add Plotting to default imports.
Add Plotting to default imports. git-svn-id: 4c7b13231a96299fde701bb5dec4bd2aaf383fc6@43 979d6bd5-6d4d-0410-bece-f567c23bd345
Python
bsd-3-clause
beni55/dadi,beni55/dadi,ChenHsiang/dadi,RyanGutenkunst/dadi,paulirish/dadi,yangjl/dadi,ChenHsiang/dadi,cheese1213/dadi,yangjl/dadi,cheese1213/dadi,paulirish/dadi,niuhuifei/dadi,niuhuifei/dadi,RyanGutenkunst/dadi
import numpy # This gives a nicer printout for masked arrays. numpy.ma.default_real_fill_value = numpy.nan import IO import Integration import PhiManip import SFS import ms + import Plotting try: import os __DIRECTORY__ = os.path.dirname(IO.__file__) __svn_file__ = os.path.join(__DIRECTORY__, 'svnversion') __SVNVERSION__ = file(__svn_file__).read().strip() except: __SVNVERSION__ = 'Unknown'
Add Plotting to default imports.
## Code Before: import numpy # This gives a nicer printout for masked arrays. numpy.ma.default_real_fill_value = numpy.nan import IO import Integration import PhiManip import SFS import ms try: import os __DIRECTORY__ = os.path.dirname(IO.__file__) __svn_file__ = os.path.join(__DIRECTORY__, 'svnversion') __SVNVERSION__ = file(__svn_file__).read().strip() except: __SVNVERSION__ = 'Unknown' ## Instruction: Add Plotting to default imports. ## Code After: import numpy # This gives a nicer printout for masked arrays. numpy.ma.default_real_fill_value = numpy.nan import IO import Integration import PhiManip import SFS import ms import Plotting try: import os __DIRECTORY__ = os.path.dirname(IO.__file__) __svn_file__ = os.path.join(__DIRECTORY__, 'svnversion') __SVNVERSION__ = file(__svn_file__).read().strip() except: __SVNVERSION__ = 'Unknown'
// ... existing code ... import SFS import ms import Plotting try: // ... rest of the code ...
7ed3ba20aae568d0c12ec361210d1189ecd534cf
lazysignup/backends.py
lazysignup/backends.py
from django.contrib.auth.backends import ModelBackend from django.contrib.auth.models import User class LazySignupBackend(ModelBackend): def authenticate(self, username=None): users = [u for u in User.objects.filter(username=username) if not u.has_usable_password()] if len(users) != 1: return None return users[0] def get_user(self, user_id): # Annotate the user with our backend so it's always available, # not just when authenticate() has been called. This will be # used by the is_lazy_user filter. user = super(LazySignupBackend, self).get_user(user_id) if user: user.backend = 'lazysignup.backends.LazySignupBackend' return user
from django.contrib.auth.backends import ModelBackend from lazysignup.models import LazyUser class LazySignupBackend(ModelBackend): def authenticate(self, username=None): lazy_users = LazyUser.objects.filter( user__username=username ).select_related('user') try: return lazy_users[0].user except IndexError: return None def get_user(self, user_id): # Annotate the user with our backend so it's always available, # not just when authenticate() has been called. This will be # used by the is_lazy_user filter. user = super(LazySignupBackend, self).get_user(user_id) if user: user.backend = 'lazysignup.backends.LazySignupBackend' return user
Remove the lazy signup backend's hard dependency on django.contrib.auth.user (and remove the inconsistency in checking for whether a user is lazy or not).
Remove the lazy signup backend's hard dependency on django.contrib.auth.user (and remove the inconsistency in checking for whether a user is lazy or not).
Python
bsd-3-clause
stefanklug/django-lazysignup,rwillmer/django-lazysignup,rwillmer/django-lazysignup,danfairs/django-lazysignup,stefanklug/django-lazysignup,danfairs/django-lazysignup
from django.contrib.auth.backends import ModelBackend - from django.contrib.auth.models import User + from lazysignup.models import LazyUser class LazySignupBackend(ModelBackend): def authenticate(self, username=None): - users = [u for u in User.objects.filter(username=username) - if not u.has_usable_password()] - if len(users) != 1: + lazy_users = LazyUser.objects.filter( + user__username=username + ).select_related('user') + try: + return lazy_users[0].user + except IndexError: return None - return users[0] def get_user(self, user_id): # Annotate the user with our backend so it's always available, # not just when authenticate() has been called. This will be # used by the is_lazy_user filter. user = super(LazySignupBackend, self).get_user(user_id) if user: user.backend = 'lazysignup.backends.LazySignupBackend' return user +
Remove the lazy signup backend's hard dependency on django.contrib.auth.user (and remove the inconsistency in checking for whether a user is lazy or not).
## Code Before: from django.contrib.auth.backends import ModelBackend from django.contrib.auth.models import User class LazySignupBackend(ModelBackend): def authenticate(self, username=None): users = [u for u in User.objects.filter(username=username) if not u.has_usable_password()] if len(users) != 1: return None return users[0] def get_user(self, user_id): # Annotate the user with our backend so it's always available, # not just when authenticate() has been called. This will be # used by the is_lazy_user filter. user = super(LazySignupBackend, self).get_user(user_id) if user: user.backend = 'lazysignup.backends.LazySignupBackend' return user ## Instruction: Remove the lazy signup backend's hard dependency on django.contrib.auth.user (and remove the inconsistency in checking for whether a user is lazy or not). ## Code After: from django.contrib.auth.backends import ModelBackend from lazysignup.models import LazyUser class LazySignupBackend(ModelBackend): def authenticate(self, username=None): lazy_users = LazyUser.objects.filter( user__username=username ).select_related('user') try: return lazy_users[0].user except IndexError: return None def get_user(self, user_id): # Annotate the user with our backend so it's always available, # not just when authenticate() has been called. This will be # used by the is_lazy_user filter. user = super(LazySignupBackend, self).get_user(user_id) if user: user.backend = 'lazysignup.backends.LazySignupBackend' return user
# ... existing code ... from django.contrib.auth.backends import ModelBackend from lazysignup.models import LazyUser class LazySignupBackend(ModelBackend): # ... modified code ... def authenticate(self, username=None): lazy_users = LazyUser.objects.filter( user__username=username ).select_related('user') try: return lazy_users[0].user except IndexError: return None def get_user(self, user_id): ... user.backend = 'lazysignup.backends.LazySignupBackend' return user # ... rest of the code ...
a17933c7806634391137244e2c17327898187146
djstripe/__init__.py
djstripe/__init__.py
from __future__ import absolute_import, division, print_function, unicode_literals import pkg_resources from . import checks # noqa: Register the checks __version__ = pkg_resources.require("dj-stripe")[0].version
from __future__ import absolute_import, division, print_function, unicode_literals import pkg_resources import stripe from . import checks # noqa: Register the checks __version__ = pkg_resources.require("dj-stripe")[0].version # Set app info # https://stripe.com/docs/building-plugins#setappinfo stripe.set_app_info( "dj-stripe", version=__version__, url="https://github.com/dj-stripe/dj-stripe" )
Set dj-stripe as stripe app info
Set dj-stripe as stripe app info https://stripe.com/docs/building-plugins#setappinfo
Python
mit
pydanny/dj-stripe,pydanny/dj-stripe,jleclanche/dj-stripe,dj-stripe/dj-stripe,kavdev/dj-stripe,jleclanche/dj-stripe,dj-stripe/dj-stripe,kavdev/dj-stripe
from __future__ import absolute_import, division, print_function, unicode_literals import pkg_resources + + import stripe + from . import checks # noqa: Register the checks __version__ = pkg_resources.require("dj-stripe")[0].version + + # Set app info + # https://stripe.com/docs/building-plugins#setappinfo + stripe.set_app_info( + "dj-stripe", + version=__version__, + url="https://github.com/dj-stripe/dj-stripe" + ) +
Set dj-stripe as stripe app info
## Code Before: from __future__ import absolute_import, division, print_function, unicode_literals import pkg_resources from . import checks # noqa: Register the checks __version__ = pkg_resources.require("dj-stripe")[0].version ## Instruction: Set dj-stripe as stripe app info ## Code After: from __future__ import absolute_import, division, print_function, unicode_literals import pkg_resources import stripe from . import checks # noqa: Register the checks __version__ = pkg_resources.require("dj-stripe")[0].version # Set app info # https://stripe.com/docs/building-plugins#setappinfo stripe.set_app_info( "dj-stripe", version=__version__, url="https://github.com/dj-stripe/dj-stripe" )
# ... existing code ... from __future__ import absolute_import, division, print_function, unicode_literals import pkg_resources import stripe from . import checks # noqa: Register the checks # ... modified code ... __version__ = pkg_resources.require("dj-stripe")[0].version # Set app info # https://stripe.com/docs/building-plugins#setappinfo stripe.set_app_info( "dj-stripe", version=__version__, url="https://github.com/dj-stripe/dj-stripe" ) # ... rest of the code ...
2dfe6e78088f974310c1e7fc309f008310be0080
dask_ndmeasure/_utils.py
dask_ndmeasure/_utils.py
import operator import numpy import dask.array from . import _compat def _norm_input_labels_index(input, labels=None, index=None): """ Normalize arguments to a standard form. """ input = _compat._asarray(input) if labels is None: labels = (input != 0).astype(numpy.int64) index = None if index is None: labels = (labels > 0).astype(numpy.int64) index = dask.array.ones(tuple(), dtype=numpy.int64, chunks=tuple()) labels = _compat._asarray(labels) index = _compat._asarray(index) # SciPy transposes these for some reason. # So we do the same thing here. # This only matters if index is some array. index = index.T if input.shape != labels.shape: raise ValueError("The input and labels arrays must be the same shape.") return (input, labels, index) def _get_label_matches(input, labels, index): input_i = _compat._indices( input.shape, dtype=numpy.int64, chunks=input.chunks ) lbl_mtch = operator.eq( index[(Ellipsis,) + labels.ndim * (None,)], labels[index.ndim * (None,)] ) input_i_mtch = dask.array.where( lbl_mtch[index.ndim * (slice(None),) + (None,)], input_i[index.ndim * (None,)], input_i.dtype.type(0) ) input_mtch = dask.array.where( lbl_mtch, input[index.ndim * (None,)], input.dtype.type(0) ) return (lbl_mtch, input_i_mtch, input_mtch)
import operator import numpy import dask.array from . import _compat def _norm_input_labels_index(input, labels=None, index=None): """ Normalize arguments to a standard form. """ input = _compat._asarray(input) if labels is None: labels = (input != 0).astype(numpy.int64) index = None if index is None: labels = (labels > 0).astype(numpy.int64) index = dask.array.ones(tuple(), dtype=numpy.int64, chunks=tuple()) labels = _compat._asarray(labels) index = _compat._asarray(index) # SciPy transposes these for some reason. # So we do the same thing here. # This only matters if index is some array. index = index.T if input.shape != labels.shape: raise ValueError("The input and labels arrays must be the same shape.") return (input, labels, index) def _get_label_matches(labels, index): lbl_mtch = operator.eq( index[(Ellipsis,) + labels.ndim * (None,)], labels[index.ndim * (None,)] ) return lbl_mtch
Simplify the label matches function
Simplify the label matches function Focus only creating a mask of selected labels and none of the other products that this function was creating before.
Python
bsd-3-clause
dask-image/dask-ndmeasure
import operator import numpy import dask.array from . import _compat def _norm_input_labels_index(input, labels=None, index=None): """ Normalize arguments to a standard form. """ input = _compat._asarray(input) if labels is None: labels = (input != 0).astype(numpy.int64) index = None if index is None: labels = (labels > 0).astype(numpy.int64) index = dask.array.ones(tuple(), dtype=numpy.int64, chunks=tuple()) labels = _compat._asarray(labels) index = _compat._asarray(index) # SciPy transposes these for some reason. # So we do the same thing here. # This only matters if index is some array. index = index.T if input.shape != labels.shape: raise ValueError("The input and labels arrays must be the same shape.") return (input, labels, index) - def _get_label_matches(input, labels, index): + def _get_label_matches(labels, index): - input_i = _compat._indices( - input.shape, dtype=numpy.int64, chunks=input.chunks - ) - lbl_mtch = operator.eq( index[(Ellipsis,) + labels.ndim * (None,)], labels[index.ndim * (None,)] ) + return lbl_mtch - input_i_mtch = dask.array.where( - lbl_mtch[index.ndim * (slice(None),) + (None,)], - input_i[index.ndim * (None,)], - input_i.dtype.type(0) - ) - input_mtch = dask.array.where( - lbl_mtch, input[index.ndim * (None,)], input.dtype.type(0) - ) - - return (lbl_mtch, input_i_mtch, input_mtch) -
Simplify the label matches function
## Code Before: import operator import numpy import dask.array from . import _compat def _norm_input_labels_index(input, labels=None, index=None): """ Normalize arguments to a standard form. """ input = _compat._asarray(input) if labels is None: labels = (input != 0).astype(numpy.int64) index = None if index is None: labels = (labels > 0).astype(numpy.int64) index = dask.array.ones(tuple(), dtype=numpy.int64, chunks=tuple()) labels = _compat._asarray(labels) index = _compat._asarray(index) # SciPy transposes these for some reason. # So we do the same thing here. # This only matters if index is some array. index = index.T if input.shape != labels.shape: raise ValueError("The input and labels arrays must be the same shape.") return (input, labels, index) def _get_label_matches(input, labels, index): input_i = _compat._indices( input.shape, dtype=numpy.int64, chunks=input.chunks ) lbl_mtch = operator.eq( index[(Ellipsis,) + labels.ndim * (None,)], labels[index.ndim * (None,)] ) input_i_mtch = dask.array.where( lbl_mtch[index.ndim * (slice(None),) + (None,)], input_i[index.ndim * (None,)], input_i.dtype.type(0) ) input_mtch = dask.array.where( lbl_mtch, input[index.ndim * (None,)], input.dtype.type(0) ) return (lbl_mtch, input_i_mtch, input_mtch) ## Instruction: Simplify the label matches function ## Code After: import operator import numpy import dask.array from . import _compat def _norm_input_labels_index(input, labels=None, index=None): """ Normalize arguments to a standard form. """ input = _compat._asarray(input) if labels is None: labels = (input != 0).astype(numpy.int64) index = None if index is None: labels = (labels > 0).astype(numpy.int64) index = dask.array.ones(tuple(), dtype=numpy.int64, chunks=tuple()) labels = _compat._asarray(labels) index = _compat._asarray(index) # SciPy transposes these for some reason. # So we do the same thing here. # This only matters if index is some array. index = index.T if input.shape != labels.shape: raise ValueError("The input and labels arrays must be the same shape.") return (input, labels, index) def _get_label_matches(labels, index): lbl_mtch = operator.eq( index[(Ellipsis,) + labels.ndim * (None,)], labels[index.ndim * (None,)] ) return lbl_mtch
# ... existing code ... def _get_label_matches(labels, index): lbl_mtch = operator.eq( index[(Ellipsis,) + labels.ndim * (None,)], # ... modified code ... ) return lbl_mtch # ... rest of the code ...
56aa0448fb3cd1df1a0fd43abc9a0e37e8ddf55b
trans_sync/management/commands/save_trans.py
trans_sync/management/commands/save_trans.py
from __future__ import unicode_literals from optparse import make_option from django.core.management.base import NoArgsCommand class Command(NoArgsCommand): option_list = NoArgsCommand.option_list + ( make_option( '--dry-run', action='store_true', dest='dry_run', default=False, help='Do not actually send signals (and all connected stuff).' ), ) def handle(self, *args, **options): if not options['dry_run']: pass
from __future__ import unicode_literals import os from os.path import join, isdir from optparse import make_option from django.core.management.base import NoArgsCommand from django.conf import settings from modeltranslation.translator import translator from babel.messages.catalog import Catalog from babel.messages.pofile import write_po class Command(NoArgsCommand): option_list = NoArgsCommand.option_list + ( make_option( '--dry-run', action='store_true', dest='dry_run', default=False, help='Do not actually save files.' ), ) def handle(self, *args, **options): if not options['dry_run']: pass locale_path = settings.LOCALE_MODEL_TRANS if not isdir(locale_path): os.mkdir(locale_path) for lang in [l[0] for l in list(settings.LANGUAGES)]: catalog = Catalog(locale=lang) for model in translator.get_registered_models(): opts = translator.get_options_for_model(model) for field in opts.get_field_names(): tr_field = "%s_%s" % (field, lang) for item in model.objects.all(): msgid = "%s.%s.%s" % (item._meta, item.pk, field) msgstr = "%s" % getattr(item, tr_field) catalog.add(id=msgid, string=msgstr) # write catalog to file lang_path = os.path.join(locale_path, lang) if not isdir(lang_path): os.mkdir(lang_path) f = open(join(lang_path, "LC_MESSAGES", "modeltranslation.po"), "w") write_po(f, catalog) f.close()
Save trans to .po files
Save trans to .po files
Python
mit
djentlemen/django-modeltranslation-sync
from __future__ import unicode_literals - + import os + from os.path import join, isdir from optparse import make_option from django.core.management.base import NoArgsCommand + from django.conf import settings + from modeltranslation.translator import translator + + from babel.messages.catalog import Catalog + from babel.messages.pofile import write_po class Command(NoArgsCommand): option_list = NoArgsCommand.option_list + ( make_option( '--dry-run', action='store_true', dest='dry_run', default=False, - help='Do not actually send signals (and all connected stuff).' + help='Do not actually save files.' ), ) def handle(self, *args, **options): if not options['dry_run']: pass + + locale_path = settings.LOCALE_MODEL_TRANS + if not isdir(locale_path): + os.mkdir(locale_path) + + for lang in [l[0] for l in list(settings.LANGUAGES)]: + + catalog = Catalog(locale=lang) + + for model in translator.get_registered_models(): + opts = translator.get_options_for_model(model) + + for field in opts.get_field_names(): + tr_field = "%s_%s" % (field, lang) + for item in model.objects.all(): + msgid = "%s.%s.%s" % (item._meta, item.pk, field) + msgstr = "%s" % getattr(item, tr_field) + catalog.add(id=msgid, string=msgstr) + + # write catalog to file + lang_path = os.path.join(locale_path, lang) + if not isdir(lang_path): + os.mkdir(lang_path) + f = open(join(lang_path, "LC_MESSAGES", "modeltranslation.po"), "w") + write_po(f, catalog) + f.close()
Save trans to .po files
## Code Before: from __future__ import unicode_literals from optparse import make_option from django.core.management.base import NoArgsCommand class Command(NoArgsCommand): option_list = NoArgsCommand.option_list + ( make_option( '--dry-run', action='store_true', dest='dry_run', default=False, help='Do not actually send signals (and all connected stuff).' ), ) def handle(self, *args, **options): if not options['dry_run']: pass ## Instruction: Save trans to .po files ## Code After: from __future__ import unicode_literals import os from os.path import join, isdir from optparse import make_option from django.core.management.base import NoArgsCommand from django.conf import settings from modeltranslation.translator import translator from babel.messages.catalog import Catalog from babel.messages.pofile import write_po class Command(NoArgsCommand): option_list = NoArgsCommand.option_list + ( make_option( '--dry-run', action='store_true', dest='dry_run', default=False, help='Do not actually save files.' ), ) def handle(self, *args, **options): if not options['dry_run']: pass locale_path = settings.LOCALE_MODEL_TRANS if not isdir(locale_path): os.mkdir(locale_path) for lang in [l[0] for l in list(settings.LANGUAGES)]: catalog = Catalog(locale=lang) for model in translator.get_registered_models(): opts = translator.get_options_for_model(model) for field in opts.get_field_names(): tr_field = "%s_%s" % (field, lang) for item in model.objects.all(): msgid = "%s.%s.%s" % (item._meta, item.pk, field) msgstr = "%s" % getattr(item, tr_field) catalog.add(id=msgid, string=msgstr) # write catalog to file lang_path = os.path.join(locale_path, lang) if not isdir(lang_path): os.mkdir(lang_path) f = open(join(lang_path, "LC_MESSAGES", "modeltranslation.po"), "w") write_po(f, catalog) f.close()
# ... existing code ... from __future__ import unicode_literals import os from os.path import join, isdir from optparse import make_option from django.core.management.base import NoArgsCommand from django.conf import settings from modeltranslation.translator import translator from babel.messages.catalog import Catalog from babel.messages.pofile import write_po # ... modified code ... dest='dry_run', default=False, help='Do not actually save files.' ), ) ... if not options['dry_run']: pass locale_path = settings.LOCALE_MODEL_TRANS if not isdir(locale_path): os.mkdir(locale_path) for lang in [l[0] for l in list(settings.LANGUAGES)]: catalog = Catalog(locale=lang) for model in translator.get_registered_models(): opts = translator.get_options_for_model(model) for field in opts.get_field_names(): tr_field = "%s_%s" % (field, lang) for item in model.objects.all(): msgid = "%s.%s.%s" % (item._meta, item.pk, field) msgstr = "%s" % getattr(item, tr_field) catalog.add(id=msgid, string=msgstr) # write catalog to file lang_path = os.path.join(locale_path, lang) if not isdir(lang_path): os.mkdir(lang_path) f = open(join(lang_path, "LC_MESSAGES", "modeltranslation.po"), "w") write_po(f, catalog) f.close() # ... rest of the code ...
35b45fd793ac695f6ec6a792534fdde77a3023aa
napalm_yang/supported_models.py
napalm_yang/supported_models.py
SUPPORTED_MODELS = ( # module_name, models ("openconfig-interfaces", ["interfaces"]), ("openconfig-network-instance", ["network_instances"]), ("openconfig-platform", ["components"]), ("openconfig-vlan", ["vlans"]), )
SUPPORTED_MODELS = ( # module_name, models ("openconfig-interfaces", ["interfaces"]), ("openconfig-network-instance", ["network_instances"]), ("openconfig-platform", ["components"]), ("openconfig-vlan", ["vlans"]), ('openconfig-system', ['system']) )
Add system as supported models
Add system as supported models
Python
apache-2.0
napalm-automation/napalm-yang,napalm-automation/napalm-yang
SUPPORTED_MODELS = ( # module_name, models ("openconfig-interfaces", ["interfaces"]), ("openconfig-network-instance", ["network_instances"]), ("openconfig-platform", ["components"]), ("openconfig-vlan", ["vlans"]), + ('openconfig-system', ['system']) )
Add system as supported models
## Code Before: SUPPORTED_MODELS = ( # module_name, models ("openconfig-interfaces", ["interfaces"]), ("openconfig-network-instance", ["network_instances"]), ("openconfig-platform", ["components"]), ("openconfig-vlan", ["vlans"]), ) ## Instruction: Add system as supported models ## Code After: SUPPORTED_MODELS = ( # module_name, models ("openconfig-interfaces", ["interfaces"]), ("openconfig-network-instance", ["network_instances"]), ("openconfig-platform", ["components"]), ("openconfig-vlan", ["vlans"]), ('openconfig-system', ['system']) )
... ("openconfig-platform", ["components"]), ("openconfig-vlan", ["vlans"]), ('openconfig-system', ['system']) ) ...
53b9eff3ffc1768d3503021e7248351e24d59af7
tests/httpd.py
tests/httpd.py
import SimpleHTTPServer import BaseHTTPServer class Handler(SimpleHTTPServer.SimpleHTTPRequestHandler): def do_POST(s): s.send_response(200) s.end_headers() if __name__ == '__main__': server_class = BaseHTTPServer.HTTPServer httpd = server_class(('0.0.0.0', 8328), Handler) try: httpd.serve_forever() except KeyboardInterrupt: httpd.server_close()
import BaseHTTPServer class Handler(BaseHTTPServer.BaseHTTPRequestHandler): def do_POST(self): content_type = self.headers.getheader('content-type') content_length = int(self.headers.getheader('content-length')) self.send_response(200) self.send_header('Content-Type', content_type) self.send_header('Content-Length', str(content_length)) self.end_headers() self.wfile.write(self.rfile.read(content_length)) if __name__ == '__main__': server_class = BaseHTTPServer.HTTPServer httpd = server_class(('0.0.0.0', 8328), Handler) try: httpd.serve_forever() except KeyboardInterrupt: httpd.server_close()
Fix test http server, change to echo back request body
Fix test http server, change to echo back request body
Python
bsd-2-clause
chop-dbhi/django-webhooks,pombredanne/django-webhooks,pombredanne/django-webhooks,chop-dbhi/django-webhooks
- import SimpleHTTPServer import BaseHTTPServer - class Handler(SimpleHTTPServer.SimpleHTTPRequestHandler): + class Handler(BaseHTTPServer.BaseHTTPRequestHandler): - def do_POST(s): + def do_POST(self): + content_type = self.headers.getheader('content-type') + content_length = int(self.headers.getheader('content-length')) - s.send_response(200) + self.send_response(200) + self.send_header('Content-Type', content_type) + self.send_header('Content-Length', str(content_length)) - s.end_headers() + self.end_headers() + self.wfile.write(self.rfile.read(content_length)) if __name__ == '__main__': server_class = BaseHTTPServer.HTTPServer httpd = server_class(('0.0.0.0', 8328), Handler) try: httpd.serve_forever() except KeyboardInterrupt: httpd.server_close()
Fix test http server, change to echo back request body
## Code Before: import SimpleHTTPServer import BaseHTTPServer class Handler(SimpleHTTPServer.SimpleHTTPRequestHandler): def do_POST(s): s.send_response(200) s.end_headers() if __name__ == '__main__': server_class = BaseHTTPServer.HTTPServer httpd = server_class(('0.0.0.0', 8328), Handler) try: httpd.serve_forever() except KeyboardInterrupt: httpd.server_close() ## Instruction: Fix test http server, change to echo back request body ## Code After: import BaseHTTPServer class Handler(BaseHTTPServer.BaseHTTPRequestHandler): def do_POST(self): content_type = self.headers.getheader('content-type') content_length = int(self.headers.getheader('content-length')) self.send_response(200) self.send_header('Content-Type', content_type) self.send_header('Content-Length', str(content_length)) self.end_headers() self.wfile.write(self.rfile.read(content_length)) if __name__ == '__main__': server_class = BaseHTTPServer.HTTPServer httpd = server_class(('0.0.0.0', 8328), Handler) try: httpd.serve_forever() except KeyboardInterrupt: httpd.server_close()
# ... existing code ... import BaseHTTPServer class Handler(BaseHTTPServer.BaseHTTPRequestHandler): def do_POST(self): content_type = self.headers.getheader('content-type') content_length = int(self.headers.getheader('content-length')) self.send_response(200) self.send_header('Content-Type', content_type) self.send_header('Content-Length', str(content_length)) self.end_headers() self.wfile.write(self.rfile.read(content_length)) if __name__ == '__main__': # ... rest of the code ...
b1feed0ced6d1328cc39bc9bba36331ec6da7803
pre_commit_hooks/detect_private_key.py
pre_commit_hooks/detect_private_key.py
from __future__ import print_function import argparse import sys BLACKLIST = [ b'BEGIN RSA PRIVATE KEY', b'BEGIN DSA PRIVATE KEY', b'BEGIN EC PRIVATE KEY', b'BEGIN OPENSSH PRIVATE KEY', b'BEGIN PRIVATE KEY', b'PuTTY-User-Key-File-2', b'BEGIN SSH2 ENCRYPTED PRIVATE KEY', ] def detect_private_key(argv=None): parser = argparse.ArgumentParser() parser.add_argument('filenames', nargs='*', help='Filenames to check') args = parser.parse_args(argv) private_key_files = [] for filename in args.filenames: with open(filename, 'rb') as f: content = f.read() if any(line in content for line in BLACKLIST): private_key_files.append(filename) if private_key_files: for private_key_file in private_key_files: print('Private key found: {}'.format(private_key_file)) return 1 else: return 0 if __name__ == '__main__': sys.exit(detect_private_key())
from __future__ import print_function import argparse import sys BLACKLIST = [ b'BEGIN RSA PRIVATE KEY', b'BEGIN DSA PRIVATE KEY', b'BEGIN EC PRIVATE KEY', b'BEGIN OPENSSH PRIVATE KEY', b'BEGIN PRIVATE KEY', b'PuTTY-User-Key-File-2', b'BEGIN SSH2 ENCRYPTED PRIVATE KEY', b'BEGIN PGP PRIVATE KEY BLOCK', ] def detect_private_key(argv=None): parser = argparse.ArgumentParser() parser.add_argument('filenames', nargs='*', help='Filenames to check') args = parser.parse_args(argv) private_key_files = [] for filename in args.filenames: with open(filename, 'rb') as f: content = f.read() if any(line in content for line in BLACKLIST): private_key_files.append(filename) if private_key_files: for private_key_file in private_key_files: print('Private key found: {}'.format(private_key_file)) return 1 else: return 0 if __name__ == '__main__': sys.exit(detect_private_key())
Add ban for pgp/gpg private key blocks
Add ban for pgp/gpg private key blocks
Python
mit
pre-commit/pre-commit-hooks,Harwood/pre-commit-hooks
from __future__ import print_function import argparse import sys BLACKLIST = [ b'BEGIN RSA PRIVATE KEY', b'BEGIN DSA PRIVATE KEY', b'BEGIN EC PRIVATE KEY', b'BEGIN OPENSSH PRIVATE KEY', b'BEGIN PRIVATE KEY', b'PuTTY-User-Key-File-2', b'BEGIN SSH2 ENCRYPTED PRIVATE KEY', + b'BEGIN PGP PRIVATE KEY BLOCK', ] def detect_private_key(argv=None): parser = argparse.ArgumentParser() parser.add_argument('filenames', nargs='*', help='Filenames to check') args = parser.parse_args(argv) private_key_files = [] for filename in args.filenames: with open(filename, 'rb') as f: content = f.read() if any(line in content for line in BLACKLIST): private_key_files.append(filename) if private_key_files: for private_key_file in private_key_files: print('Private key found: {}'.format(private_key_file)) return 1 else: return 0 if __name__ == '__main__': sys.exit(detect_private_key())
Add ban for pgp/gpg private key blocks
## Code Before: from __future__ import print_function import argparse import sys BLACKLIST = [ b'BEGIN RSA PRIVATE KEY', b'BEGIN DSA PRIVATE KEY', b'BEGIN EC PRIVATE KEY', b'BEGIN OPENSSH PRIVATE KEY', b'BEGIN PRIVATE KEY', b'PuTTY-User-Key-File-2', b'BEGIN SSH2 ENCRYPTED PRIVATE KEY', ] def detect_private_key(argv=None): parser = argparse.ArgumentParser() parser.add_argument('filenames', nargs='*', help='Filenames to check') args = parser.parse_args(argv) private_key_files = [] for filename in args.filenames: with open(filename, 'rb') as f: content = f.read() if any(line in content for line in BLACKLIST): private_key_files.append(filename) if private_key_files: for private_key_file in private_key_files: print('Private key found: {}'.format(private_key_file)) return 1 else: return 0 if __name__ == '__main__': sys.exit(detect_private_key()) ## Instruction: Add ban for pgp/gpg private key blocks ## Code After: from __future__ import print_function import argparse import sys BLACKLIST = [ b'BEGIN RSA PRIVATE KEY', b'BEGIN DSA PRIVATE KEY', b'BEGIN EC PRIVATE KEY', b'BEGIN OPENSSH PRIVATE KEY', b'BEGIN PRIVATE KEY', b'PuTTY-User-Key-File-2', b'BEGIN SSH2 ENCRYPTED PRIVATE KEY', b'BEGIN PGP PRIVATE KEY BLOCK', ] def detect_private_key(argv=None): parser = argparse.ArgumentParser() parser.add_argument('filenames', nargs='*', help='Filenames to check') args = parser.parse_args(argv) private_key_files = [] for filename in args.filenames: with open(filename, 'rb') as f: content = f.read() if any(line in content for line in BLACKLIST): private_key_files.append(filename) if private_key_files: for private_key_file in private_key_files: print('Private key found: {}'.format(private_key_file)) return 1 else: return 0 if __name__ == '__main__': sys.exit(detect_private_key())
... b'PuTTY-User-Key-File-2', b'BEGIN SSH2 ENCRYPTED PRIVATE KEY', b'BEGIN PGP PRIVATE KEY BLOCK', ] ...
1914bd9a9fc2f1ce7a557dd2134d98796866283d
rootpy/userdata.py
rootpy/userdata.py
import os import tempfile import atexit DATA_ROOT = None if os.getenv('ROOTPY_GRIDMODE') not in ('1', 'true'): DATA_ROOT = os.getenv('ROOTPY_DATA') if DATA_ROOT is None: DATA_ROOT = os.path.expanduser('~/.rootpy') else: DATA_ROOT = os.path.expandvars(os.path.expanduser(DATA_ROOT)) # check if expanduser failed: if DATA_ROOT.startswith('~'): DATA_ROOT = None elif not os.path.exists(DATA_ROOT): os.mkdir(DATA_ROOT) elif not os.path.isdir(DATA_ROOT): # A file at DATA_ROOT already exists DATA_ROOT = None __is_tmp = False if DATA_ROOT is None: print "Warning: placing user data in /tmp.\n" \ "Make sure ~/.rootpy or $ROOTPY_DATA\n" \ "is a writable directory so that I don't need to\n" \ "recreate all user data each time" DATA_ROOT = tempfile.mkdtemp() __is_tmp = True @atexit.register def __cleanup(): if __is_tmp: import shutil shutil.rmtree(DATA_ROOT)
import os import tempfile import atexit from os.path import expanduser, expandvars, exists, isdir if "XDG_CONFIG_HOME" not in os.environ: os.environ["XDG_CONFIG_HOME"] = expanduser('~/.config') if "XDG_CACHE_HOME" not in os.environ: os.environ["XDG_CACHE_HOME"] = expanduser('~/.cache') def ensure_directory(variable, default): path = os.getenv(variable) if path is None: path = expandvars(default) else: path = expandvars(expanduser(path)) # check if expanduser failed: if path.startswith('~'): path = None elif not exists(path): os.makedirs(path) elif not isdir(path): # A file at path already exists path = None return path DATA_ROOT = CONFIG_ROOT = None if os.getenv('ROOTPY_GRIDMODE') not in ('1', 'true'): DATA_ROOT = ensure_directory('ROOTPY_DATA', '${XDG_CACHE_HOME}/rootpy') CONFIG_ROOT = ensure_directory('ROOTPY_CONFIG', '${XDG_CONFIG_HOME}/rootpy') if DATA_ROOT is None: log.info("Placing user data in /tmp.") log.warning("Make sure '~/.cache/rootpy' or $ROOTPY_DATA is a writable " "directory so that it isn't necessary to recreate all user data" " each time") DATA_ROOT = tempfile.mkdtemp() @atexit.register def __cleanup(): import shutil shutil.rmtree(DATA_ROOT)
Use XDG base directory specificiation
Use XDG base directory specificiation
Python
bsd-3-clause
rootpy/rootpy,rootpy/rootpy,ndawe/rootpy,kreczko/rootpy,ndawe/rootpy,kreczko/rootpy,ndawe/rootpy,kreczko/rootpy,rootpy/rootpy
import os import tempfile import atexit + from os.path import expanduser, expandvars, exists, isdir - - DATA_ROOT = None - if os.getenv('ROOTPY_GRIDMODE') not in ('1', 'true'): - DATA_ROOT = os.getenv('ROOTPY_DATA') - if DATA_ROOT is None: - DATA_ROOT = os.path.expanduser('~/.rootpy') - else: - DATA_ROOT = os.path.expandvars(os.path.expanduser(DATA_ROOT)) - # check if expanduser failed: - if DATA_ROOT.startswith('~'): - DATA_ROOT = None - elif not os.path.exists(DATA_ROOT): - os.mkdir(DATA_ROOT) - elif not os.path.isdir(DATA_ROOT): - # A file at DATA_ROOT already exists - DATA_ROOT = None - - __is_tmp = False - if DATA_ROOT is None: - print "Warning: placing user data in /tmp.\n" \ - "Make sure ~/.rootpy or $ROOTPY_DATA\n" \ - "is a writable directory so that I don't need to\n" \ - "recreate all user data each time" - DATA_ROOT = tempfile.mkdtemp() - __is_tmp = True + if "XDG_CONFIG_HOME" not in os.environ: + os.environ["XDG_CONFIG_HOME"] = expanduser('~/.config') + if "XDG_CACHE_HOME" not in os.environ: + os.environ["XDG_CACHE_HOME"] = expanduser('~/.cache') + + def ensure_directory(variable, default): + path = os.getenv(variable) + if path is None: + path = expandvars(default) + else: + path = expandvars(expanduser(path)) + + # check if expanduser failed: + if path.startswith('~'): + path = None + elif not exists(path): + os.makedirs(path) + elif not isdir(path): + # A file at path already exists + path = None + return path + + DATA_ROOT = CONFIG_ROOT = None + if os.getenv('ROOTPY_GRIDMODE') not in ('1', 'true'): + DATA_ROOT = ensure_directory('ROOTPY_DATA', '${XDG_CACHE_HOME}/rootpy') + CONFIG_ROOT = ensure_directory('ROOTPY_CONFIG', '${XDG_CONFIG_HOME}/rootpy') + + + if DATA_ROOT is None: + log.info("Placing user data in /tmp.") + log.warning("Make sure '~/.cache/rootpy' or $ROOTPY_DATA is a writable " + "directory so that it isn't necessary to recreate all user data" + " each time") + + DATA_ROOT = tempfile.mkdtemp() + - @atexit.register + @atexit.register - def __cleanup(): + def __cleanup(): - if __is_tmp: import shutil shutil.rmtree(DATA_ROOT) +
Use XDG base directory specificiation
## Code Before: import os import tempfile import atexit DATA_ROOT = None if os.getenv('ROOTPY_GRIDMODE') not in ('1', 'true'): DATA_ROOT = os.getenv('ROOTPY_DATA') if DATA_ROOT is None: DATA_ROOT = os.path.expanduser('~/.rootpy') else: DATA_ROOT = os.path.expandvars(os.path.expanduser(DATA_ROOT)) # check if expanduser failed: if DATA_ROOT.startswith('~'): DATA_ROOT = None elif not os.path.exists(DATA_ROOT): os.mkdir(DATA_ROOT) elif not os.path.isdir(DATA_ROOT): # A file at DATA_ROOT already exists DATA_ROOT = None __is_tmp = False if DATA_ROOT is None: print "Warning: placing user data in /tmp.\n" \ "Make sure ~/.rootpy or $ROOTPY_DATA\n" \ "is a writable directory so that I don't need to\n" \ "recreate all user data each time" DATA_ROOT = tempfile.mkdtemp() __is_tmp = True @atexit.register def __cleanup(): if __is_tmp: import shutil shutil.rmtree(DATA_ROOT) ## Instruction: Use XDG base directory specificiation ## Code After: import os import tempfile import atexit from os.path import expanduser, expandvars, exists, isdir if "XDG_CONFIG_HOME" not in os.environ: os.environ["XDG_CONFIG_HOME"] = expanduser('~/.config') if "XDG_CACHE_HOME" not in os.environ: os.environ["XDG_CACHE_HOME"] = expanduser('~/.cache') def ensure_directory(variable, default): path = os.getenv(variable) if path is None: path = expandvars(default) else: path = expandvars(expanduser(path)) # check if expanduser failed: if path.startswith('~'): path = None elif not exists(path): os.makedirs(path) elif not isdir(path): # A file at path already exists path = None return path DATA_ROOT = CONFIG_ROOT = None if os.getenv('ROOTPY_GRIDMODE') not in ('1', 'true'): DATA_ROOT = ensure_directory('ROOTPY_DATA', '${XDG_CACHE_HOME}/rootpy') CONFIG_ROOT = ensure_directory('ROOTPY_CONFIG', '${XDG_CONFIG_HOME}/rootpy') if DATA_ROOT is None: log.info("Placing user data in /tmp.") log.warning("Make sure '~/.cache/rootpy' or $ROOTPY_DATA is a writable " "directory so that it isn't necessary to recreate all user data" " each time") DATA_ROOT = tempfile.mkdtemp() @atexit.register def __cleanup(): import shutil shutil.rmtree(DATA_ROOT)
# ... existing code ... import atexit from os.path import expanduser, expandvars, exists, isdir if "XDG_CONFIG_HOME" not in os.environ: os.environ["XDG_CONFIG_HOME"] = expanduser('~/.config') if "XDG_CACHE_HOME" not in os.environ: os.environ["XDG_CACHE_HOME"] = expanduser('~/.cache') def ensure_directory(variable, default): path = os.getenv(variable) if path is None: path = expandvars(default) else: path = expandvars(expanduser(path)) # check if expanduser failed: if path.startswith('~'): path = None elif not exists(path): os.makedirs(path) elif not isdir(path): # A file at path already exists path = None return path DATA_ROOT = CONFIG_ROOT = None if os.getenv('ROOTPY_GRIDMODE') not in ('1', 'true'): DATA_ROOT = ensure_directory('ROOTPY_DATA', '${XDG_CACHE_HOME}/rootpy') CONFIG_ROOT = ensure_directory('ROOTPY_CONFIG', '${XDG_CONFIG_HOME}/rootpy') if DATA_ROOT is None: log.info("Placing user data in /tmp.") log.warning("Make sure '~/.cache/rootpy' or $ROOTPY_DATA is a writable " "directory so that it isn't necessary to recreate all user data" " each time") DATA_ROOT = tempfile.mkdtemp() @atexit.register def __cleanup(): import shutil shutil.rmtree(DATA_ROOT) # ... rest of the code ...
cd9048f64c6a2184e148daf0baa7bb3be51b3268
vol/__init__.py
vol/__init__.py
from __future__ import unicode_literals, print_function from sys import platform if platform == 'darwin': from .osx import OSXVolumeController as VolumeController else: raise NotImplementedError( 'VolumeController for {} platform has not been implemented yet'.format(platform) )
''' A cross platform implementation of volume control ''' from __future__ import unicode_literals, print_function from sys import platform if platform == 'darwin': from .osx import OSXVolumeController as VolumeController else: raise NotImplementedError( 'VolumeController for {} platform has not been implemented yet'.format(platform) )
Update doc for vol pkg
Update doc for vol pkg
Python
bsd-3-clause
Microcore/AGT,Microcore/YAS
+ ''' + A cross platform implementation of volume control + ''' from __future__ import unicode_literals, print_function from sys import platform if platform == 'darwin': from .osx import OSXVolumeController as VolumeController else: raise NotImplementedError( 'VolumeController for {} platform has not been implemented yet'.format(platform) )
Update doc for vol pkg
## Code Before: from __future__ import unicode_literals, print_function from sys import platform if platform == 'darwin': from .osx import OSXVolumeController as VolumeController else: raise NotImplementedError( 'VolumeController for {} platform has not been implemented yet'.format(platform) ) ## Instruction: Update doc for vol pkg ## Code After: ''' A cross platform implementation of volume control ''' from __future__ import unicode_literals, print_function from sys import platform if platform == 'darwin': from .osx import OSXVolumeController as VolumeController else: raise NotImplementedError( 'VolumeController for {} platform has not been implemented yet'.format(platform) )
// ... existing code ... ''' A cross platform implementation of volume control ''' from __future__ import unicode_literals, print_function from sys import platform // ... rest of the code ...
0ea1153438c1d98232a921c8d14d401a541e95fd
examples/regex/regex_parser.py
examples/regex/regex_parser.py
from __future__ import print_function, division, absolute_import, unicode_literals from parser_base import RegexParser import model class RegexSemantics(object): def __init__(self): super(RegexSemantics, self).__init__() self._count = 0 def START(self, ast): return model.Regex(ast) def CHOICE(self, ast): return model.Choice(ast.opts) def SEQUENCE(self, ast): if not ast.terms: return model.Empty() elif len(ast.terms) < 2: return ast.terms[0] else: return model.Sequence(ast.terms) def CLOSURE(self, ast): return model.Closure(ast) def SUBEXP(self, ast): return ast def LITERAL(self, ast): return model.Literal(ast) def translate(regex, trace=False): parser = RegexParser(trace=trace, semantics=RegexSemantics()) model = parser.parse(regex, 'START') model.set_rule_numbers() return model.render()
from __future__ import print_function, division, absolute_import, unicode_literals from parser_base import RegexParser import model class RegexSemantics(object): def __init__(self): super(RegexSemantics, self).__init__() self._count = 0 def START(self, ast): return model.Regex(ast) def CHOICE(self, ast): return model.Choice(ast.opts) def SEQUENCE(self, ast): if not ast.terms: return model.Empty() elif len(ast.terms) < 2: return ast.terms[0] else: return model.Sequence(ast.terms) def CLOSURE(self, ast): return model.Closure(ast) def SUBEXP(self, ast): return ast def LITERAL(self, ast): return model.Literal(ast) def translate(regex, trace=False): parser = RegexParser(trace=trace, semantics=RegexSemantics()) model = parser.parse(regex, 'START') model.set_rule_numbers() return model.render().encode("ascii")
Fix regex example, the model must not be a unicode string.
Fix regex example, the model must not be a unicode string.
Python
bsd-2-clause
vmuriart/grako,frnknglrt/grako
from __future__ import print_function, division, absolute_import, unicode_literals from parser_base import RegexParser import model class RegexSemantics(object): def __init__(self): super(RegexSemantics, self).__init__() self._count = 0 def START(self, ast): return model.Regex(ast) def CHOICE(self, ast): return model.Choice(ast.opts) def SEQUENCE(self, ast): if not ast.terms: return model.Empty() elif len(ast.terms) < 2: return ast.terms[0] else: return model.Sequence(ast.terms) def CLOSURE(self, ast): return model.Closure(ast) def SUBEXP(self, ast): return ast def LITERAL(self, ast): return model.Literal(ast) def translate(regex, trace=False): parser = RegexParser(trace=trace, semantics=RegexSemantics()) model = parser.parse(regex, 'START') model.set_rule_numbers() - return model.render() + return model.render().encode("ascii")
Fix regex example, the model must not be a unicode string.
## Code Before: from __future__ import print_function, division, absolute_import, unicode_literals from parser_base import RegexParser import model class RegexSemantics(object): def __init__(self): super(RegexSemantics, self).__init__() self._count = 0 def START(self, ast): return model.Regex(ast) def CHOICE(self, ast): return model.Choice(ast.opts) def SEQUENCE(self, ast): if not ast.terms: return model.Empty() elif len(ast.terms) < 2: return ast.terms[0] else: return model.Sequence(ast.terms) def CLOSURE(self, ast): return model.Closure(ast) def SUBEXP(self, ast): return ast def LITERAL(self, ast): return model.Literal(ast) def translate(regex, trace=False): parser = RegexParser(trace=trace, semantics=RegexSemantics()) model = parser.parse(regex, 'START') model.set_rule_numbers() return model.render() ## Instruction: Fix regex example, the model must not be a unicode string. ## Code After: from __future__ import print_function, division, absolute_import, unicode_literals from parser_base import RegexParser import model class RegexSemantics(object): def __init__(self): super(RegexSemantics, self).__init__() self._count = 0 def START(self, ast): return model.Regex(ast) def CHOICE(self, ast): return model.Choice(ast.opts) def SEQUENCE(self, ast): if not ast.terms: return model.Empty() elif len(ast.terms) < 2: return ast.terms[0] else: return model.Sequence(ast.terms) def CLOSURE(self, ast): return model.Closure(ast) def SUBEXP(self, ast): return ast def LITERAL(self, ast): return model.Literal(ast) def translate(regex, trace=False): parser = RegexParser(trace=trace, semantics=RegexSemantics()) model = parser.parse(regex, 'START') model.set_rule_numbers() return model.render().encode("ascii")
# ... existing code ... model = parser.parse(regex, 'START') model.set_rule_numbers() return model.render().encode("ascii") # ... rest of the code ...
03d628abc4711bb0de4a7a0ef13cc4c0ecb92032
opps/articles/tests/models.py
opps/articles/tests/models.py
from django.test import TestCase from opps.articles.models import Post class PostModelTest(TestCase): fixtures = ['tests/initial_data.json'] def test_basic_post_exist(self): post = Post.objects.all() self.assertTrue(post) self.assertEqual(len(post), 1) self.assertEqual(post[0].slug, u'test-post-application') self.assertEqual(post[0].title, u'test post application') self.assertTrue(post[0].short_url)
from django.test import TestCase from opps.articles.models import Post class PostModelTest(TestCase): fixtures = ['tests/initial_data.json'] def test_basic_post_exist(self): post = Post.objects.all() self.assertTrue(post) self.assertEqual(len(post), 1) self.assertEqual(post[0].slug, u'test-post-application') self.assertEqual(post[0].title, u'test post application') self.assertTrue(post[0].short_url) def test_child_class(self): post = Post.objects.get(id=1) self.assertTrue(post.child_class) self.assertEqual(post.child_class, 'Post')
Add test articles (post), check child_class
Add test articles (post), check child_class
Python
mit
YACOWS/opps,jeanmask/opps,williamroot/opps,opps/opps,YACOWS/opps,opps/opps,jeanmask/opps,opps/opps,williamroot/opps,YACOWS/opps,jeanmask/opps,williamroot/opps,YACOWS/opps,jeanmask/opps,opps/opps,williamroot/opps
from django.test import TestCase from opps.articles.models import Post class PostModelTest(TestCase): fixtures = ['tests/initial_data.json'] def test_basic_post_exist(self): post = Post.objects.all() self.assertTrue(post) self.assertEqual(len(post), 1) self.assertEqual(post[0].slug, u'test-post-application') self.assertEqual(post[0].title, u'test post application') self.assertTrue(post[0].short_url) + def test_child_class(self): + post = Post.objects.get(id=1) + + self.assertTrue(post.child_class) + self.assertEqual(post.child_class, 'Post') +
Add test articles (post), check child_class
## Code Before: from django.test import TestCase from opps.articles.models import Post class PostModelTest(TestCase): fixtures = ['tests/initial_data.json'] def test_basic_post_exist(self): post = Post.objects.all() self.assertTrue(post) self.assertEqual(len(post), 1) self.assertEqual(post[0].slug, u'test-post-application') self.assertEqual(post[0].title, u'test post application') self.assertTrue(post[0].short_url) ## Instruction: Add test articles (post), check child_class ## Code After: from django.test import TestCase from opps.articles.models import Post class PostModelTest(TestCase): fixtures = ['tests/initial_data.json'] def test_basic_post_exist(self): post = Post.objects.all() self.assertTrue(post) self.assertEqual(len(post), 1) self.assertEqual(post[0].slug, u'test-post-application') self.assertEqual(post[0].title, u'test post application') self.assertTrue(post[0].short_url) def test_child_class(self): post = Post.objects.get(id=1) self.assertTrue(post.child_class) self.assertEqual(post.child_class, 'Post')
... self.assertEqual(post[0].title, u'test post application') self.assertTrue(post[0].short_url) def test_child_class(self): post = Post.objects.get(id=1) self.assertTrue(post.child_class) self.assertEqual(post.child_class, 'Post') ...
6af05b8af7bb284388af4960bbf240122f7f3dae
plugins/PerObjectSettingsTool/__init__.py
plugins/PerObjectSettingsTool/__init__.py
from . import PerObjectSettingsTool from UM.i18n import i18nCatalog i18n_catalog = i18nCatalog("uranium") def getMetaData(): return { "plugin": { "name": i18n_catalog.i18nc("@label", "Per Object Settings Tool"), "author": "Ultimaker", "version": "1.0", "description": i18n_catalog.i18nc("@info:whatsthis", "Provides the Per Object Settings."), "api": 2 }, "tool": { "name": i18n_catalog.i18nc("@label", "Per Object Settings"), "description": i18n_catalog.i18nc("@info:tooltip", "Configure Per Object Settings"), "icon": "setting_per_object", "tool_panel": "PerObjectSettingsPanel.qml" }, } def register(app): return { "tool": PerObjectSettingsTool.PerObjectSettingsTool() }
from . import PerObjectSettingsTool from UM.i18n import i18nCatalog i18n_catalog = i18nCatalog("uranium") def getMetaData(): return { "plugin": { "name": i18n_catalog.i18nc("@label", "Per Object Settings Tool"), "author": "Ultimaker", "version": "1.0", "description": i18n_catalog.i18nc("@info:whatsthis", "Provides the Per Object Settings."), "api": 2 }, "tool": { "name": i18n_catalog.i18nc("@label", "Per Object Settings"), "description": i18n_catalog.i18nc("@info:tooltip", "Configure Per Object Settings"), "icon": "setting_per_object", "tool_panel": "PerObjectSettingsPanel.qml", "weight": 3 }, } def register(app): return { "tool": PerObjectSettingsTool.PerObjectSettingsTool() }
Add order to PerObjectSettings tool
Add order to PerObjectSettings tool
Python
agpl-3.0
ynotstartups/Wanhao,hmflash/Cura,fieldOfView/Cura,senttech/Cura,hmflash/Cura,fieldOfView/Cura,Curahelper/Cura,senttech/Cura,Curahelper/Cura,totalretribution/Cura,totalretribution/Cura,ynotstartups/Wanhao
from . import PerObjectSettingsTool from UM.i18n import i18nCatalog i18n_catalog = i18nCatalog("uranium") def getMetaData(): return { "plugin": { "name": i18n_catalog.i18nc("@label", "Per Object Settings Tool"), "author": "Ultimaker", "version": "1.0", "description": i18n_catalog.i18nc("@info:whatsthis", "Provides the Per Object Settings."), "api": 2 }, "tool": { "name": i18n_catalog.i18nc("@label", "Per Object Settings"), "description": i18n_catalog.i18nc("@info:tooltip", "Configure Per Object Settings"), "icon": "setting_per_object", - "tool_panel": "PerObjectSettingsPanel.qml" + "tool_panel": "PerObjectSettingsPanel.qml", + "weight": 3 }, } def register(app): return { "tool": PerObjectSettingsTool.PerObjectSettingsTool() }
Add order to PerObjectSettings tool
## Code Before: from . import PerObjectSettingsTool from UM.i18n import i18nCatalog i18n_catalog = i18nCatalog("uranium") def getMetaData(): return { "plugin": { "name": i18n_catalog.i18nc("@label", "Per Object Settings Tool"), "author": "Ultimaker", "version": "1.0", "description": i18n_catalog.i18nc("@info:whatsthis", "Provides the Per Object Settings."), "api": 2 }, "tool": { "name": i18n_catalog.i18nc("@label", "Per Object Settings"), "description": i18n_catalog.i18nc("@info:tooltip", "Configure Per Object Settings"), "icon": "setting_per_object", "tool_panel": "PerObjectSettingsPanel.qml" }, } def register(app): return { "tool": PerObjectSettingsTool.PerObjectSettingsTool() } ## Instruction: Add order to PerObjectSettings tool ## Code After: from . import PerObjectSettingsTool from UM.i18n import i18nCatalog i18n_catalog = i18nCatalog("uranium") def getMetaData(): return { "plugin": { "name": i18n_catalog.i18nc("@label", "Per Object Settings Tool"), "author": "Ultimaker", "version": "1.0", "description": i18n_catalog.i18nc("@info:whatsthis", "Provides the Per Object Settings."), "api": 2 }, "tool": { "name": i18n_catalog.i18nc("@label", "Per Object Settings"), "description": i18n_catalog.i18nc("@info:tooltip", "Configure Per Object Settings"), "icon": "setting_per_object", "tool_panel": "PerObjectSettingsPanel.qml", "weight": 3 }, } def register(app): return { "tool": PerObjectSettingsTool.PerObjectSettingsTool() }
# ... existing code ... "description": i18n_catalog.i18nc("@info:tooltip", "Configure Per Object Settings"), "icon": "setting_per_object", "tool_panel": "PerObjectSettingsPanel.qml", "weight": 3 }, } # ... rest of the code ...
52e9390d88062e9442b18a7793e6696a36f5b9c3
testinfra/functional/test_tor_interfaces.py
testinfra/functional/test_tor_interfaces.py
import os import re import pytest sdvars = pytest.securedrop_test_vars @pytest.mark.xfail @pytest.mark.parametrize('site', sdvars.tor_url_files) @pytest.mark.skipif(os.environ.get('FPF_CI', 'false') == "false", reason="Can only assure Tor is configured in CI atm") def test_www(Command, site): """ Ensure tor interface is reachable and returns expected content. """ # Extract Onion URL from saved onion file, fetched back from app-staging. onion_url_filepath = os.path.join( os.path.dirname(__file__), "../../install_files/ansible-base/{}".format(site['file']) ) onion_url_raw = open(onion_url_filepath, 'ro').read() onion_url = re.search("\w+\.onion", onion_url_raw).group() # Fetch Onion URL via curl to confirm interface is rendered correctly. curl_tor = 'curl -s --socks5-hostname "${{TOR_PROXY}}":9050 {}'.format( onion_url) curl_tor_status = '{} -o /dev/null -w "%{{http_code}}"'.format(curl_tor) site_scrape = Command.check_output(curl_tor) assert Command.check_output(curl_tor_status) == "200" assert site['check_string'] in site_scrape assert site['error_string'] not in site_scrape
import os import re import pytest sdvars = pytest.securedrop_test_vars @pytest.mark.parametrize('site', sdvars.tor_url_files) @pytest.mark.skipif(os.environ.get('FPF_CI', 'false') == "false", reason="Can only assure Tor is configured in CI atm") def test_www(Command, site): """ Ensure tor interface is reachable and returns expected content. """ # Extract Onion URL from saved onion file, fetched back from app-staging. onion_url_filepath = os.path.join( os.path.dirname(__file__), "../../install_files/ansible-base/{}".format(site['file']) ) onion_url_raw = open(onion_url_filepath, 'ro').read() onion_url = re.search("\w+\.onion", onion_url_raw).group() # Fetch Onion URL via curl to confirm interface is rendered correctly. curl_tor = 'curl -s --socks5-hostname "${{TOR_PROXY}}":9050 {}'.format( onion_url) curl_tor_status = '{} -o /dev/null -w "%{{http_code}}"'.format(curl_tor) site_scrape = Command.check_output(curl_tor) assert Command.check_output(curl_tor_status) == "200" assert site['check_string'] in site_scrape assert site['error_string'] not in site_scrape
Remove XFAIL on functional tor test
Remove XFAIL on functional tor test
Python
agpl-3.0
conorsch/securedrop,ehartsuyker/securedrop,garrettr/securedrop,ehartsuyker/securedrop,conorsch/securedrop,heartsucker/securedrop,garrettr/securedrop,ehartsuyker/securedrop,ehartsuyker/securedrop,conorsch/securedrop,ehartsuyker/securedrop,heartsucker/securedrop,conorsch/securedrop,heartsucker/securedrop,ehartsuyker/securedrop,heartsucker/securedrop,garrettr/securedrop,conorsch/securedrop,garrettr/securedrop,heartsucker/securedrop
import os import re import pytest sdvars = pytest.securedrop_test_vars - @pytest.mark.xfail @pytest.mark.parametrize('site', sdvars.tor_url_files) @pytest.mark.skipif(os.environ.get('FPF_CI', 'false') == "false", reason="Can only assure Tor is configured in CI atm") def test_www(Command, site): """ Ensure tor interface is reachable and returns expected content. """ # Extract Onion URL from saved onion file, fetched back from app-staging. onion_url_filepath = os.path.join( os.path.dirname(__file__), "../../install_files/ansible-base/{}".format(site['file']) ) onion_url_raw = open(onion_url_filepath, 'ro').read() onion_url = re.search("\w+\.onion", onion_url_raw).group() # Fetch Onion URL via curl to confirm interface is rendered correctly. curl_tor = 'curl -s --socks5-hostname "${{TOR_PROXY}}":9050 {}'.format( onion_url) curl_tor_status = '{} -o /dev/null -w "%{{http_code}}"'.format(curl_tor) site_scrape = Command.check_output(curl_tor) assert Command.check_output(curl_tor_status) == "200" assert site['check_string'] in site_scrape assert site['error_string'] not in site_scrape
Remove XFAIL on functional tor test
## Code Before: import os import re import pytest sdvars = pytest.securedrop_test_vars @pytest.mark.xfail @pytest.mark.parametrize('site', sdvars.tor_url_files) @pytest.mark.skipif(os.environ.get('FPF_CI', 'false') == "false", reason="Can only assure Tor is configured in CI atm") def test_www(Command, site): """ Ensure tor interface is reachable and returns expected content. """ # Extract Onion URL from saved onion file, fetched back from app-staging. onion_url_filepath = os.path.join( os.path.dirname(__file__), "../../install_files/ansible-base/{}".format(site['file']) ) onion_url_raw = open(onion_url_filepath, 'ro').read() onion_url = re.search("\w+\.onion", onion_url_raw).group() # Fetch Onion URL via curl to confirm interface is rendered correctly. curl_tor = 'curl -s --socks5-hostname "${{TOR_PROXY}}":9050 {}'.format( onion_url) curl_tor_status = '{} -o /dev/null -w "%{{http_code}}"'.format(curl_tor) site_scrape = Command.check_output(curl_tor) assert Command.check_output(curl_tor_status) == "200" assert site['check_string'] in site_scrape assert site['error_string'] not in site_scrape ## Instruction: Remove XFAIL on functional tor test ## Code After: import os import re import pytest sdvars = pytest.securedrop_test_vars @pytest.mark.parametrize('site', sdvars.tor_url_files) @pytest.mark.skipif(os.environ.get('FPF_CI', 'false') == "false", reason="Can only assure Tor is configured in CI atm") def test_www(Command, site): """ Ensure tor interface is reachable and returns expected content. """ # Extract Onion URL from saved onion file, fetched back from app-staging. onion_url_filepath = os.path.join( os.path.dirname(__file__), "../../install_files/ansible-base/{}".format(site['file']) ) onion_url_raw = open(onion_url_filepath, 'ro').read() onion_url = re.search("\w+\.onion", onion_url_raw).group() # Fetch Onion URL via curl to confirm interface is rendered correctly. curl_tor = 'curl -s --socks5-hostname "${{TOR_PROXY}}":9050 {}'.format( onion_url) curl_tor_status = '{} -o /dev/null -w "%{{http_code}}"'.format(curl_tor) site_scrape = Command.check_output(curl_tor) assert Command.check_output(curl_tor_status) == "200" assert site['check_string'] in site_scrape assert site['error_string'] not in site_scrape
// ... existing code ... @pytest.mark.parametrize('site', sdvars.tor_url_files) @pytest.mark.skipif(os.environ.get('FPF_CI', 'false') == "false", // ... rest of the code ...
548cfea821bf1b0b92ce09c54405554d264b5395
tests/integration/session/test_timeout.py
tests/integration/session/test_timeout.py
import time from app import settings from tests.integration.integration_test_case import IntegrationTestCase class TestTimeout(IntegrationTestCase): def setUp(self): settings.EQ_SESSION_TIMEOUT_SECONDS = 1 settings.EQ_SESSION_TIMEOUT_GRACE_PERIOD_SECONDS = 0 super().setUp() def test_timeout_continue_returns_200(self): self.launchSurvey('test', 'timeout') self.get('/timeout-continue') self.assertStatusOK() def test_when_session_times_out_server_side_401_is_returned(self): self.launchSurvey('test', 'timeout') time.sleep(2) self.get(self.last_url) self.assertStatusUnauthorised() def test_schema_defined_timeout_is_used(self): self.launchSurvey('test', 'timeout') self.assertInPage('window.__EQ_SESSION_TIMEOUT__ = 1') def test_schema_defined_timeout_cant_be_higher_than_server(self): self._application.config['EQ_SESSION_TIMEOUT_SECONDS'] = 10 self.launchSurvey('test', 'timeout') self.assertInPage('window.__EQ_SESSION_TIMEOUT__ = 6')
import time from app import settings from tests.integration.integration_test_case import IntegrationTestCase class TestTimeout(IntegrationTestCase): def setUp(self): settings.EQ_SESSION_TIMEOUT_SECONDS = 1 settings.EQ_SESSION_TIMEOUT_GRACE_PERIOD_SECONDS = 0 super().setUp() def tearDown(self): settings.EQ_SESSION_TIMEOUT_SECONDS = 45 * 60 settings.EQ_SESSION_TIMEOUT_GRACE_PERIOD_SECONDS = 30 super().tearDown() def test_timeout_continue_returns_200(self): self.launchSurvey('test', 'timeout') self.get('/timeout-continue') self.assertStatusOK() def test_when_session_times_out_server_side_401_is_returned(self): self.launchSurvey('test', 'timeout') time.sleep(2) self.get(self.last_url) self.assertStatusUnauthorised() def test_schema_defined_timeout_is_used(self): self.launchSurvey('test', 'timeout') self.assertInPage('window.__EQ_SESSION_TIMEOUT__ = 1') def test_schema_defined_timeout_cant_be_higher_than_server(self): self._application.config['EQ_SESSION_TIMEOUT_SECONDS'] = 10 self.launchSurvey('test', 'timeout') self.assertInPage('window.__EQ_SESSION_TIMEOUT__ = 6')
Fix CSRF missing errors that happen occasionally in tests
Fix CSRF missing errors that happen occasionally in tests
Python
mit
ONSdigital/eq-survey-runner,ONSdigital/eq-survey-runner,ONSdigital/eq-survey-runner,ONSdigital/eq-survey-runner
import time from app import settings from tests.integration.integration_test_case import IntegrationTestCase class TestTimeout(IntegrationTestCase): def setUp(self): settings.EQ_SESSION_TIMEOUT_SECONDS = 1 settings.EQ_SESSION_TIMEOUT_GRACE_PERIOD_SECONDS = 0 super().setUp() + + def tearDown(self): + settings.EQ_SESSION_TIMEOUT_SECONDS = 45 * 60 + settings.EQ_SESSION_TIMEOUT_GRACE_PERIOD_SECONDS = 30 + super().tearDown() def test_timeout_continue_returns_200(self): self.launchSurvey('test', 'timeout') self.get('/timeout-continue') self.assertStatusOK() def test_when_session_times_out_server_side_401_is_returned(self): self.launchSurvey('test', 'timeout') time.sleep(2) self.get(self.last_url) self.assertStatusUnauthorised() def test_schema_defined_timeout_is_used(self): self.launchSurvey('test', 'timeout') self.assertInPage('window.__EQ_SESSION_TIMEOUT__ = 1') def test_schema_defined_timeout_cant_be_higher_than_server(self): self._application.config['EQ_SESSION_TIMEOUT_SECONDS'] = 10 self.launchSurvey('test', 'timeout') self.assertInPage('window.__EQ_SESSION_TIMEOUT__ = 6')
Fix CSRF missing errors that happen occasionally in tests
## Code Before: import time from app import settings from tests.integration.integration_test_case import IntegrationTestCase class TestTimeout(IntegrationTestCase): def setUp(self): settings.EQ_SESSION_TIMEOUT_SECONDS = 1 settings.EQ_SESSION_TIMEOUT_GRACE_PERIOD_SECONDS = 0 super().setUp() def test_timeout_continue_returns_200(self): self.launchSurvey('test', 'timeout') self.get('/timeout-continue') self.assertStatusOK() def test_when_session_times_out_server_side_401_is_returned(self): self.launchSurvey('test', 'timeout') time.sleep(2) self.get(self.last_url) self.assertStatusUnauthorised() def test_schema_defined_timeout_is_used(self): self.launchSurvey('test', 'timeout') self.assertInPage('window.__EQ_SESSION_TIMEOUT__ = 1') def test_schema_defined_timeout_cant_be_higher_than_server(self): self._application.config['EQ_SESSION_TIMEOUT_SECONDS'] = 10 self.launchSurvey('test', 'timeout') self.assertInPage('window.__EQ_SESSION_TIMEOUT__ = 6') ## Instruction: Fix CSRF missing errors that happen occasionally in tests ## Code After: import time from app import settings from tests.integration.integration_test_case import IntegrationTestCase class TestTimeout(IntegrationTestCase): def setUp(self): settings.EQ_SESSION_TIMEOUT_SECONDS = 1 settings.EQ_SESSION_TIMEOUT_GRACE_PERIOD_SECONDS = 0 super().setUp() def tearDown(self): settings.EQ_SESSION_TIMEOUT_SECONDS = 45 * 60 settings.EQ_SESSION_TIMEOUT_GRACE_PERIOD_SECONDS = 30 super().tearDown() def test_timeout_continue_returns_200(self): self.launchSurvey('test', 'timeout') self.get('/timeout-continue') self.assertStatusOK() def test_when_session_times_out_server_side_401_is_returned(self): self.launchSurvey('test', 'timeout') time.sleep(2) self.get(self.last_url) self.assertStatusUnauthorised() def test_schema_defined_timeout_is_used(self): self.launchSurvey('test', 'timeout') self.assertInPage('window.__EQ_SESSION_TIMEOUT__ = 1') def test_schema_defined_timeout_cant_be_higher_than_server(self): self._application.config['EQ_SESSION_TIMEOUT_SECONDS'] = 10 self.launchSurvey('test', 'timeout') self.assertInPage('window.__EQ_SESSION_TIMEOUT__ = 6')
... settings.EQ_SESSION_TIMEOUT_GRACE_PERIOD_SECONDS = 0 super().setUp() def tearDown(self): settings.EQ_SESSION_TIMEOUT_SECONDS = 45 * 60 settings.EQ_SESSION_TIMEOUT_GRACE_PERIOD_SECONDS = 30 super().tearDown() def test_timeout_continue_returns_200(self): ...
2664e9124af6b0d8f6b2eacd50f4d7e93b91e931
examples/GoBot/gobot.py
examples/GoBot/gobot.py
from minibot.bot import Bot from minibot.hardware.rpi.gpio import PWM from minibot.interface.servo import Servo import math import time L_MOTOR_PIN = 12 R_MOTOR_PIN = 18 class GoBot(Bot): def __init__(self): Bot.__init__(self, "GoBot") self.l_motor = Servo(PWM(L_MOTOR_PIN, 2, 15)) self.r_motor = Servo(PWM(R_MOTOR_PIN, 2, 15)) self.l_motor.set(17) self.r_motor.set(13) def run(self): pass if __name__ == "__main__": bot = GoBot() while True: bot.run()
from minibot.bot import Bot from minibot.hardware.rpi.gpio import PWM from minibot.interface.servo import Servo L_MOTOR_PIN = 12 R_MOTOR_PIN = 18 class GoBot(Bot): """ GoBot """ def __init__(self): Bot.__init__(self, "GoBot") self.l_motor = Servo(PWM(L_MOTOR_PIN, 2, 15)) self.r_motor = Servo(PWM(R_MOTOR_PIN, 2, 15)) self.l_motor.set(17) self.r_motor.set(13) def run(self): pass if __name__ == "__main__": bot = GoBot() while True: bot.run()
Fix linting errors in GoBot
Fix linting errors in GoBot
Python
apache-2.0
cornell-cup/cs-minibot,cornell-cup/cs-minibot,cornell-cup/cs-minibot,cornell-cup/cs-minibot,cornell-cup/cs-minibot
+ from minibot.bot import Bot from minibot.hardware.rpi.gpio import PWM from minibot.interface.servo import Servo - - import math - import time L_MOTOR_PIN = 12 R_MOTOR_PIN = 18 class GoBot(Bot): + """ + GoBot + """ def __init__(self): Bot.__init__(self, "GoBot") self.l_motor = Servo(PWM(L_MOTOR_PIN, 2, 15)) self.r_motor = Servo(PWM(R_MOTOR_PIN, 2, 15)) self.l_motor.set(17) self.r_motor.set(13) def run(self): pass if __name__ == "__main__": bot = GoBot() while True: bot.run()
Fix linting errors in GoBot
## Code Before: from minibot.bot import Bot from minibot.hardware.rpi.gpio import PWM from minibot.interface.servo import Servo import math import time L_MOTOR_PIN = 12 R_MOTOR_PIN = 18 class GoBot(Bot): def __init__(self): Bot.__init__(self, "GoBot") self.l_motor = Servo(PWM(L_MOTOR_PIN, 2, 15)) self.r_motor = Servo(PWM(R_MOTOR_PIN, 2, 15)) self.l_motor.set(17) self.r_motor.set(13) def run(self): pass if __name__ == "__main__": bot = GoBot() while True: bot.run() ## Instruction: Fix linting errors in GoBot ## Code After: from minibot.bot import Bot from minibot.hardware.rpi.gpio import PWM from minibot.interface.servo import Servo L_MOTOR_PIN = 12 R_MOTOR_PIN = 18 class GoBot(Bot): """ GoBot """ def __init__(self): Bot.__init__(self, "GoBot") self.l_motor = Servo(PWM(L_MOTOR_PIN, 2, 15)) self.r_motor = Servo(PWM(R_MOTOR_PIN, 2, 15)) self.l_motor.set(17) self.r_motor.set(13) def run(self): pass if __name__ == "__main__": bot = GoBot() while True: bot.run()
... from minibot.bot import Bot from minibot.hardware.rpi.gpio import PWM from minibot.interface.servo import Servo L_MOTOR_PIN = 12 ... class GoBot(Bot): """ GoBot """ def __init__(self): ...
dcecdbae798e0a83afb17911ec459224790e51cd
launch_control/dashboard_app/tests.py
launch_control/dashboard_app/tests.py
from django.test import TestCase from django.db import IntegrityError from launch_control.dashboard_app.models import ( SoftwarePackage, ) class SoftwarePackageTestCase(TestCase): def test_creation_1(self): sw_package = SoftwarePackage.objects.create(name='libfoo', version='1.2.0') self.assertEqual(sw_package.name, 'libfoo') self.assertEqual(sw_package.version, '1.2.0') def test_uniqueness(self): SoftwarePackage.objects.create(name='a', version='0') self.assertRaises(IntegrityError, SoftwarePackage.objects.create, name='a', version='0')
from django.test import TestCase from django.db import IntegrityError from launch_control.utils.call_helper import ObjectFactoryMixIn from launch_control.dashboard_app.models import ( SoftwarePackage, ) class SoftwarePackageTestCase(TestCase, ObjectFactoryMixIn): class Dummy: class SoftwarePackage: name = 'libfoo' version = '1.2.0' def test_creation_1(self): dummy, sw_package = self.make_and_get_dummy(SoftwarePackage) self.assertEqual(sw_package.name, dummy.name) self.assertEqual(sw_package.version, dummy.version) def test_uniqueness(self): pkg1 = self.make(SoftwarePackage) pkg1.save() pkg2 = self.make(SoftwarePackage) self.assertRaises(IntegrityError, pkg2.save)
Update SoftwarePackageTestCase to use ObjectFactoryMixIn
Update SoftwarePackageTestCase to use ObjectFactoryMixIn
Python
agpl-3.0
OSSystems/lava-server,Linaro/lava-server,OSSystems/lava-server,Linaro/lava-server,Linaro/lava-server,Linaro/lava-server,OSSystems/lava-server
from django.test import TestCase from django.db import IntegrityError + from launch_control.utils.call_helper import ObjectFactoryMixIn from launch_control.dashboard_app.models import ( SoftwarePackage, ) - class SoftwarePackageTestCase(TestCase): + class SoftwarePackageTestCase(TestCase, ObjectFactoryMixIn): + + class Dummy: + class SoftwarePackage: + name = 'libfoo' + version = '1.2.0' def test_creation_1(self): - sw_package = SoftwarePackage.objects.create(name='libfoo', version='1.2.0') + dummy, sw_package = self.make_and_get_dummy(SoftwarePackage) - self.assertEqual(sw_package.name, 'libfoo') + self.assertEqual(sw_package.name, dummy.name) - self.assertEqual(sw_package.version, '1.2.0') + self.assertEqual(sw_package.version, dummy.version) def test_uniqueness(self): - SoftwarePackage.objects.create(name='a', version='0') - self.assertRaises(IntegrityError, SoftwarePackage.objects.create, - name='a', version='0') + pkg1 = self.make(SoftwarePackage) + pkg1.save() + pkg2 = self.make(SoftwarePackage) + self.assertRaises(IntegrityError, pkg2.save)
Update SoftwarePackageTestCase to use ObjectFactoryMixIn
## Code Before: from django.test import TestCase from django.db import IntegrityError from launch_control.dashboard_app.models import ( SoftwarePackage, ) class SoftwarePackageTestCase(TestCase): def test_creation_1(self): sw_package = SoftwarePackage.objects.create(name='libfoo', version='1.2.0') self.assertEqual(sw_package.name, 'libfoo') self.assertEqual(sw_package.version, '1.2.0') def test_uniqueness(self): SoftwarePackage.objects.create(name='a', version='0') self.assertRaises(IntegrityError, SoftwarePackage.objects.create, name='a', version='0') ## Instruction: Update SoftwarePackageTestCase to use ObjectFactoryMixIn ## Code After: from django.test import TestCase from django.db import IntegrityError from launch_control.utils.call_helper import ObjectFactoryMixIn from launch_control.dashboard_app.models import ( SoftwarePackage, ) class SoftwarePackageTestCase(TestCase, ObjectFactoryMixIn): class Dummy: class SoftwarePackage: name = 'libfoo' version = '1.2.0' def test_creation_1(self): dummy, sw_package = self.make_and_get_dummy(SoftwarePackage) self.assertEqual(sw_package.name, dummy.name) self.assertEqual(sw_package.version, dummy.version) def test_uniqueness(self): pkg1 = self.make(SoftwarePackage) pkg1.save() pkg2 = self.make(SoftwarePackage) self.assertRaises(IntegrityError, pkg2.save)
# ... existing code ... from django.db import IntegrityError from launch_control.utils.call_helper import ObjectFactoryMixIn from launch_control.dashboard_app.models import ( SoftwarePackage, # ... modified code ... class SoftwarePackageTestCase(TestCase, ObjectFactoryMixIn): class Dummy: class SoftwarePackage: name = 'libfoo' version = '1.2.0' def test_creation_1(self): dummy, sw_package = self.make_and_get_dummy(SoftwarePackage) self.assertEqual(sw_package.name, dummy.name) self.assertEqual(sw_package.version, dummy.version) def test_uniqueness(self): pkg1 = self.make(SoftwarePackage) pkg1.save() pkg2 = self.make(SoftwarePackage) self.assertRaises(IntegrityError, pkg2.save) # ... rest of the code ...
b1685dc4a0a2036378d47f07d7315e5b1935a4ad
hyrodactil/tests/openings/models.py
hyrodactil/tests/openings/models.py
from django.test import TestCase from ..factories._applications import ApplicationFactory from ..factories._companysettings import InterviewStageFactory from ..factories._openings import OpeningFactory class OpeningsModelsTests(TestCase): def test_applicants_stats(self): opening = OpeningFactory() self.assertEqual(opening.applicants_stats(), []) s1 = InterviewStageFactory(name="L220", initial=True, company=opening.company) s2 = InterviewStageFactory(name="L33", company=opening.company) self.assertEqual(opening.applicants_stats(), [[s1.name, 0], [s2.name, 0]]) application = ApplicationFactory.create(opening=opening) application.stage_transitions.create(stage=s1) self.assertEqual(opening.applicants_stats(), [[s1.name, 1], [s2.name, 0]])
from django.test import TestCase from ..factories._applications import ApplicationFactory from ..factories._companysettings import InterviewStageFactory from ..factories._openings import OpeningFactory class OpeningsModelsTests(TestCase): def test_applicants_stats(self): opening = OpeningFactory() self.assertEqual(opening.applicants_stats(), []) s1 = InterviewStageFactory(name="L220", company=opening.company) s2 = InterviewStageFactory(name="L33", company=opening.company) self.assertEqual(opening.applicants_stats(), [[s1.name, 0], [s2.name, 0]]) application = ApplicationFactory.create(opening=opening) application.stage_transitions.create(stage=s1) self.assertEqual(opening.applicants_stats(), [[s1.name, 1], [s2.name, 0]])
Fix tests for positioned stages
openings: Fix tests for positioned stages
Python
mit
hizardapp/Hizard,hizardapp/Hizard,hizardapp/Hizard
from django.test import TestCase from ..factories._applications import ApplicationFactory from ..factories._companysettings import InterviewStageFactory from ..factories._openings import OpeningFactory class OpeningsModelsTests(TestCase): def test_applicants_stats(self): opening = OpeningFactory() self.assertEqual(opening.applicants_stats(), []) - s1 = InterviewStageFactory(name="L220", initial=True, + s1 = InterviewStageFactory(name="L220", company=opening.company) s2 = InterviewStageFactory(name="L33", company=opening.company) self.assertEqual(opening.applicants_stats(), [[s1.name, 0], [s2.name, 0]]) application = ApplicationFactory.create(opening=opening) application.stage_transitions.create(stage=s1) self.assertEqual(opening.applicants_stats(), [[s1.name, 1], [s2.name, 0]])
Fix tests for positioned stages
## Code Before: from django.test import TestCase from ..factories._applications import ApplicationFactory from ..factories._companysettings import InterviewStageFactory from ..factories._openings import OpeningFactory class OpeningsModelsTests(TestCase): def test_applicants_stats(self): opening = OpeningFactory() self.assertEqual(opening.applicants_stats(), []) s1 = InterviewStageFactory(name="L220", initial=True, company=opening.company) s2 = InterviewStageFactory(name="L33", company=opening.company) self.assertEqual(opening.applicants_stats(), [[s1.name, 0], [s2.name, 0]]) application = ApplicationFactory.create(opening=opening) application.stage_transitions.create(stage=s1) self.assertEqual(opening.applicants_stats(), [[s1.name, 1], [s2.name, 0]]) ## Instruction: Fix tests for positioned stages ## Code After: from django.test import TestCase from ..factories._applications import ApplicationFactory from ..factories._companysettings import InterviewStageFactory from ..factories._openings import OpeningFactory class OpeningsModelsTests(TestCase): def test_applicants_stats(self): opening = OpeningFactory() self.assertEqual(opening.applicants_stats(), []) s1 = InterviewStageFactory(name="L220", company=opening.company) s2 = InterviewStageFactory(name="L33", company=opening.company) self.assertEqual(opening.applicants_stats(), [[s1.name, 0], [s2.name, 0]]) application = ApplicationFactory.create(opening=opening) application.stage_transitions.create(stage=s1) self.assertEqual(opening.applicants_stats(), [[s1.name, 1], [s2.name, 0]])
# ... existing code ... opening = OpeningFactory() self.assertEqual(opening.applicants_stats(), []) s1 = InterviewStageFactory(name="L220", company=opening.company) s2 = InterviewStageFactory(name="L33", # ... rest of the code ...
20696d6f236afc1bc0e2b3db570363540e70ca84
test/test_serve.py
test/test_serve.py
import unittest import asyncio import io import multiprocessing import urllib.request import time import grole def simple_server(): app = grole.Grole() @app.route('/') def hello(env, req): return 'Hello, World!' app.run() class TestServe(unittest.TestCase): def test_simple(self): p = multiprocessing.Process(target=simple_server) p.start() time.sleep(0.1) with urllib.request.urlopen('http://localhost:1234') as response: html = response.read() self.assertEqual(html, b'Hello, World!') p.terminate() def test_fileserver(self): p = multiprocessing.Process(target=grole.main, args=[[]]) p.start() time.sleep(0.1) with urllib.request.urlopen('http://localhost:1234/test/test.dat') as response: html = response.read() self.assertEqual(html, b'foo\n') p.terminate()
import unittest import asyncio import io import multiprocessing import urllib.request import time import grole def simple_server(): app = grole.Grole() @app.route('/') def hello(env, req): return 'Hello, World!' app.run(host='127.0.0.1') class TestServe(unittest.TestCase): def test_simple(self): p = multiprocessing.Process(target=simple_server) p.start() time.sleep(0.1) with urllib.request.urlopen('http://127.0.0.1:1234') as response: html = response.read() self.assertEqual(html, b'Hello, World!') p.terminate() def test_fileserver(self): p = multiprocessing.Process(target=grole.main, args=[['-a', '127.0.0.1']]) p.start() time.sleep(0.1) with urllib.request.urlopen('http://127.0.0.1:1234/test/test.dat') as response: html = response.read() self.assertEqual(html, b'foo\n') p.terminate()
Use ip instead of localhost for travis
Use ip instead of localhost for travis
Python
mit
witchard/grole
import unittest import asyncio import io import multiprocessing import urllib.request import time import grole def simple_server(): app = grole.Grole() @app.route('/') def hello(env, req): return 'Hello, World!' - app.run() + app.run(host='127.0.0.1') class TestServe(unittest.TestCase): def test_simple(self): p = multiprocessing.Process(target=simple_server) p.start() time.sleep(0.1) - with urllib.request.urlopen('http://localhost:1234') as response: + with urllib.request.urlopen('http://127.0.0.1:1234') as response: html = response.read() self.assertEqual(html, b'Hello, World!') p.terminate() def test_fileserver(self): - p = multiprocessing.Process(target=grole.main, args=[[]]) + p = multiprocessing.Process(target=grole.main, args=[['-a', '127.0.0.1']]) p.start() time.sleep(0.1) - with urllib.request.urlopen('http://localhost:1234/test/test.dat') as response: + with urllib.request.urlopen('http://127.0.0.1:1234/test/test.dat') as response: html = response.read() self.assertEqual(html, b'foo\n') p.terminate()
Use ip instead of localhost for travis
## Code Before: import unittest import asyncio import io import multiprocessing import urllib.request import time import grole def simple_server(): app = grole.Grole() @app.route('/') def hello(env, req): return 'Hello, World!' app.run() class TestServe(unittest.TestCase): def test_simple(self): p = multiprocessing.Process(target=simple_server) p.start() time.sleep(0.1) with urllib.request.urlopen('http://localhost:1234') as response: html = response.read() self.assertEqual(html, b'Hello, World!') p.terminate() def test_fileserver(self): p = multiprocessing.Process(target=grole.main, args=[[]]) p.start() time.sleep(0.1) with urllib.request.urlopen('http://localhost:1234/test/test.dat') as response: html = response.read() self.assertEqual(html, b'foo\n') p.terminate() ## Instruction: Use ip instead of localhost for travis ## Code After: import unittest import asyncio import io import multiprocessing import urllib.request import time import grole def simple_server(): app = grole.Grole() @app.route('/') def hello(env, req): return 'Hello, World!' app.run(host='127.0.0.1') class TestServe(unittest.TestCase): def test_simple(self): p = multiprocessing.Process(target=simple_server) p.start() time.sleep(0.1) with urllib.request.urlopen('http://127.0.0.1:1234') as response: html = response.read() self.assertEqual(html, b'Hello, World!') p.terminate() def test_fileserver(self): p = multiprocessing.Process(target=grole.main, args=[['-a', '127.0.0.1']]) p.start() time.sleep(0.1) with urllib.request.urlopen('http://127.0.0.1:1234/test/test.dat') as response: html = response.read() self.assertEqual(html, b'foo\n') p.terminate()
... return 'Hello, World!' app.run(host='127.0.0.1') class TestServe(unittest.TestCase): ... p.start() time.sleep(0.1) with urllib.request.urlopen('http://127.0.0.1:1234') as response: html = response.read() self.assertEqual(html, b'Hello, World!') ... def test_fileserver(self): p = multiprocessing.Process(target=grole.main, args=[['-a', '127.0.0.1']]) p.start() time.sleep(0.1) with urllib.request.urlopen('http://127.0.0.1:1234/test/test.dat') as response: html = response.read() self.assertEqual(html, b'foo\n') ...
c07234bb3142df96dc9e02a236975bc3de2415cc
nailgun/nailgun/test/test_plugin.py
nailgun/nailgun/test/test_plugin.py
from nailgun.test.base import BaseHandlers class TestPluginStateMachine(BaseHandlers): def test_attrs_creation(self): pass
from nailgun.test.base import BaseHandlers from nailgun.plugin.process import get_queue, PluginProcessor from nailgun.api.models import Task class TestPluginProcess(BaseHandlers): def setUp(self): super(TestPluginProcess, self).setUp() self.plugin_processor = PluginProcessor() self.plugin_processor.start() def tearDown(self): super(TestPluginProcess, self).tearDown() self.plugin_processor.terminate() def test_task_set_to_error_when_exception_raised(self): queue = get_queue() task = Task(name='install_plugin', cache={'plugin_id': -1}) self.env.db.add(task) self.env.db.commit() queue.put(task.uuid) def check_task_status_is_error(): self.env.db.refresh(task) return task.status == 'error' self.env.wait_for_true(check_task_status_is_error, timeout=2) self.assertEquals(task.progress, 100)
Implement plugin test on exception handling
Implement plugin test on exception handling
Python
apache-2.0
SmartInfrastructures/fuel-main-dev,ddepaoli3/fuel-main-dev,zhaochao/fuel-main,zhaochao/fuel-main,huntxu/fuel-main,prmtl/fuel-web,huntxu/fuel-web,huntxu/fuel-main,SmartInfrastructures/fuel-main-dev,huntxu/fuel-web,teselkin/fuel-main,ddepaoli3/fuel-main-dev,teselkin/fuel-main,SmartInfrastructures/fuel-web-dev,SergK/fuel-main,dancn/fuel-main-dev,SmartInfrastructures/fuel-web-dev,nebril/fuel-web,SergK/fuel-main,zhaochao/fuel-web,dancn/fuel-main-dev,nebril/fuel-web,dancn/fuel-main-dev,AnselZhangGit/fuel-main,Fiware/ops.Fuel-main-dev,AnselZhangGit/fuel-main,nebril/fuel-web,SmartInfrastructures/fuel-main-dev,eayunstack/fuel-web,AnselZhangGit/fuel-main,eayunstack/fuel-web,SmartInfrastructures/fuel-web-dev,eayunstack/fuel-main,SergK/fuel-main,prmtl/fuel-web,zhaochao/fuel-web,eayunstack/fuel-main,huntxu/fuel-web,zhaochao/fuel-web,SmartInfrastructures/fuel-web-dev,koder-ua/nailgun-fcert,koder-ua/nailgun-fcert,ddepaoli3/fuel-main-dev,SmartInfrastructures/fuel-web-dev,zhaochao/fuel-web,zhaochao/fuel-web,teselkin/fuel-main,stackforge/fuel-main,prmtl/fuel-web,AnselZhangGit/fuel-main,zhaochao/fuel-main,eayunstack/fuel-web,eayunstack/fuel-main,koder-ua/nailgun-fcert,stackforge/fuel-web,nebril/fuel-web,stackforge/fuel-web,zhaochao/fuel-main,ddepaoli3/fuel-main-dev,Fiware/ops.Fuel-main-dev,prmtl/fuel-web,SmartInfrastructures/fuel-main-dev,Fiware/ops.Fuel-main-dev,zhaochao/fuel-main,koder-ua/nailgun-fcert,huntxu/fuel-web,dancn/fuel-main-dev,stackforge/fuel-main,eayunstack/fuel-web,prmtl/fuel-web,stackforge/fuel-web,huntxu/fuel-web,huntxu/fuel-main,Fiware/ops.Fuel-main-dev,teselkin/fuel-main,nebril/fuel-web,eayunstack/fuel-web,stackforge/fuel-main
from nailgun.test.base import BaseHandlers + from nailgun.plugin.process import get_queue, PluginProcessor + from nailgun.api.models import Task + class TestPluginProcess(BaseHandlers): + def setUp(self): + super(TestPluginProcess, self).setUp() + self.plugin_processor = PluginProcessor() + self.plugin_processor.start() - class TestPluginStateMachine(BaseHandlers): + def tearDown(self): + super(TestPluginProcess, self).tearDown() + self.plugin_processor.terminate() - def test_attrs_creation(self): - pass + def test_task_set_to_error_when_exception_raised(self): + queue = get_queue() + task = Task(name='install_plugin', cache={'plugin_id': -1}) + self.env.db.add(task) + self.env.db.commit() + queue.put(task.uuid) + + def check_task_status_is_error(): + self.env.db.refresh(task) + return task.status == 'error' + + self.env.wait_for_true(check_task_status_is_error, timeout=2) + self.assertEquals(task.progress, 100) +
Implement plugin test on exception handling
## Code Before: from nailgun.test.base import BaseHandlers class TestPluginStateMachine(BaseHandlers): def test_attrs_creation(self): pass ## Instruction: Implement plugin test on exception handling ## Code After: from nailgun.test.base import BaseHandlers from nailgun.plugin.process import get_queue, PluginProcessor from nailgun.api.models import Task class TestPluginProcess(BaseHandlers): def setUp(self): super(TestPluginProcess, self).setUp() self.plugin_processor = PluginProcessor() self.plugin_processor.start() def tearDown(self): super(TestPluginProcess, self).tearDown() self.plugin_processor.terminate() def test_task_set_to_error_when_exception_raised(self): queue = get_queue() task = Task(name='install_plugin', cache={'plugin_id': -1}) self.env.db.add(task) self.env.db.commit() queue.put(task.uuid) def check_task_status_is_error(): self.env.db.refresh(task) return task.status == 'error' self.env.wait_for_true(check_task_status_is_error, timeout=2) self.assertEquals(task.progress, 100)
// ... existing code ... from nailgun.test.base import BaseHandlers from nailgun.plugin.process import get_queue, PluginProcessor from nailgun.api.models import Task class TestPluginProcess(BaseHandlers): def setUp(self): super(TestPluginProcess, self).setUp() self.plugin_processor = PluginProcessor() self.plugin_processor.start() def tearDown(self): super(TestPluginProcess, self).tearDown() self.plugin_processor.terminate() def test_task_set_to_error_when_exception_raised(self): queue = get_queue() task = Task(name='install_plugin', cache={'plugin_id': -1}) self.env.db.add(task) self.env.db.commit() queue.put(task.uuid) def check_task_status_is_error(): self.env.db.refresh(task) return task.status == 'error' self.env.wait_for_true(check_task_status_is_error, timeout=2) self.assertEquals(task.progress, 100) // ... rest of the code ...
f37d26541d6baf3da47a8f373a8c7a65177067db
push/modules/push_notification.py
push/modules/push_notification.py
import time, os, json from apns import APNs, Frame, Payload from push.models import DevelopFileModel, ProductFileModel from django.conf import settings PEM_FILE_DIR = settings.BASE_DIR + '/push/files/' def execute(device_token_lists, notification): if notification.is_production: pem_file_name = ProductFileModel.objects.all()[0].production_file_name apns = APNs(use_sandbox = False, cert_file = PEM_FILE_DIR + pem_file_name, enhanced = True) else: pem_file_name = DevelopFileModel.objects.all()[0].development_file_name apns = APNs(use_sandbox = True, cert_file = PEM_FILE_DIR + pem_file_name, enhanced = True) token_hex = [] for token in device_token_lists: token_hex.append(token) json_data = '' if notification.json != '': json_data = json.loads(notification.json) payload = Payload(alert = notification.message, sound = notification.sound, badge = notification.badge, custom = json_data) frame = Frame() identifier = 1 expiry = time.time() + 3600 priority = 10 for token in token_hex: frame.add_item(token, payload, identifier, expiry, priority) apns.gateway_server.send_notification_multiple(frame)
import time, os, json from apns import APNs, Frame, Payload from push.models import DevelopFileModel, ProductFileModel from django.conf import settings PEM_FILE_DIR = settings.BASE_DIR + '/push/files/' def execute(device_token_lists, notification): if notification.is_production: pem_file_name = ProductFileModel.objects.all()[0].production_file_name apns = APNs(use_sandbox = False, cert_file = PEM_FILE_DIR + pem_file_name, enhanced = True) else: pem_file_name = DevelopFileModel.objects.all()[0].development_file_name apns = APNs(use_sandbox = True, cert_file = PEM_FILE_DIR + pem_file_name, enhanced = True) token_hex = [] for token in device_token_lists: token_hex.append(token) json_data = '' if notification.json != '': json_data = json.loads(notification.json) payload = Payload(alert = notification.message, sound = notification.sound, badge = notification.badge, custom = json_data) frame = Frame() identifier = 1 expiry = time.time() + 3600 priority = 10 for token in token_hex: frame.add_item(token, payload, identifier, expiry, priority) apns.gateway_server.send_notification_multiple(frame) notification.is_sent = True notification.save()
Send flag when success push notifications
Send flag when success push notifications
Python
apache-2.0
nnsnodnb/django-mbaas,nnsnodnb/django-mbaas,nnsnodnb/django-mbaas
import time, os, json from apns import APNs, Frame, Payload from push.models import DevelopFileModel, ProductFileModel from django.conf import settings PEM_FILE_DIR = settings.BASE_DIR + '/push/files/' def execute(device_token_lists, notification): if notification.is_production: pem_file_name = ProductFileModel.objects.all()[0].production_file_name apns = APNs(use_sandbox = False, cert_file = PEM_FILE_DIR + pem_file_name, enhanced = True) else: pem_file_name = DevelopFileModel.objects.all()[0].development_file_name apns = APNs(use_sandbox = True, cert_file = PEM_FILE_DIR + pem_file_name, enhanced = True) token_hex = [] for token in device_token_lists: token_hex.append(token) json_data = '' if notification.json != '': json_data = json.loads(notification.json) payload = Payload(alert = notification.message, sound = notification.sound, badge = notification.badge, custom = json_data) frame = Frame() identifier = 1 expiry = time.time() + 3600 priority = 10 for token in token_hex: frame.add_item(token, payload, identifier, expiry, priority) apns.gateway_server.send_notification_multiple(frame) + notification.is_sent = True + notification.save() +
Send flag when success push notifications
## Code Before: import time, os, json from apns import APNs, Frame, Payload from push.models import DevelopFileModel, ProductFileModel from django.conf import settings PEM_FILE_DIR = settings.BASE_DIR + '/push/files/' def execute(device_token_lists, notification): if notification.is_production: pem_file_name = ProductFileModel.objects.all()[0].production_file_name apns = APNs(use_sandbox = False, cert_file = PEM_FILE_DIR + pem_file_name, enhanced = True) else: pem_file_name = DevelopFileModel.objects.all()[0].development_file_name apns = APNs(use_sandbox = True, cert_file = PEM_FILE_DIR + pem_file_name, enhanced = True) token_hex = [] for token in device_token_lists: token_hex.append(token) json_data = '' if notification.json != '': json_data = json.loads(notification.json) payload = Payload(alert = notification.message, sound = notification.sound, badge = notification.badge, custom = json_data) frame = Frame() identifier = 1 expiry = time.time() + 3600 priority = 10 for token in token_hex: frame.add_item(token, payload, identifier, expiry, priority) apns.gateway_server.send_notification_multiple(frame) ## Instruction: Send flag when success push notifications ## Code After: import time, os, json from apns import APNs, Frame, Payload from push.models import DevelopFileModel, ProductFileModel from django.conf import settings PEM_FILE_DIR = settings.BASE_DIR + '/push/files/' def execute(device_token_lists, notification): if notification.is_production: pem_file_name = ProductFileModel.objects.all()[0].production_file_name apns = APNs(use_sandbox = False, cert_file = PEM_FILE_DIR + pem_file_name, enhanced = True) else: pem_file_name = DevelopFileModel.objects.all()[0].development_file_name apns = APNs(use_sandbox = True, cert_file = PEM_FILE_DIR + pem_file_name, enhanced = True) token_hex = [] for token in device_token_lists: token_hex.append(token) json_data = '' if notification.json != '': json_data = json.loads(notification.json) payload = Payload(alert = notification.message, sound = notification.sound, badge = notification.badge, custom = json_data) frame = Frame() identifier = 1 expiry = time.time() + 3600 priority = 10 for token in token_hex: frame.add_item(token, payload, identifier, expiry, priority) apns.gateway_server.send_notification_multiple(frame) notification.is_sent = True notification.save()
... apns.gateway_server.send_notification_multiple(frame) notification.is_sent = True notification.save() ...
8623aae8778307648e4a0380d84ca7dc7a63f3f2
oneflow/core/context_processors.py
oneflow/core/context_processors.py
from .models.nonrel import User def mongodb_user(request): if request.user.is_anonymous(): return {u'mongodb_user': None} try: mongodb_user = User.objects.get(id=request.session[u'mongodb_user_id']) except KeyError: mongodb_user = User.objects.get(django_user=request.user.id) # Cache it for next time. request.session[u'mongodb_user_id'] = mongodb_user.id return {u'mongodb_user': mongodb_user}
def mongodb_user(request): """ not the most usefull context manager in the world. """ if request.user.is_anonymous(): return {u'mongodb_user': None} return {u'mongodb_user': request.user.mongo}
Simplify the context processor. Not very useful anymore, in fact.
Simplify the context processor. Not very useful anymore, in fact.
Python
agpl-3.0
1flow/1flow,1flow/1flow,WillianPaiva/1flow,WillianPaiva/1flow,1flow/1flow,WillianPaiva/1flow,WillianPaiva/1flow,1flow/1flow,1flow/1flow,WillianPaiva/1flow
- - from .models.nonrel import User def mongodb_user(request): + """ not the most usefull context manager in the world. """ if request.user.is_anonymous(): return {u'mongodb_user': None} + return {u'mongodb_user': request.user.mongo} - try: - mongodb_user = User.objects.get(id=request.session[u'mongodb_user_id']) - except KeyError: - mongodb_user = User.objects.get(django_user=request.user.id) - - # Cache it for next time. - request.session[u'mongodb_user_id'] = mongodb_user.id - - return {u'mongodb_user': mongodb_user} -
Simplify the context processor. Not very useful anymore, in fact.
## Code Before: from .models.nonrel import User def mongodb_user(request): if request.user.is_anonymous(): return {u'mongodb_user': None} try: mongodb_user = User.objects.get(id=request.session[u'mongodb_user_id']) except KeyError: mongodb_user = User.objects.get(django_user=request.user.id) # Cache it for next time. request.session[u'mongodb_user_id'] = mongodb_user.id return {u'mongodb_user': mongodb_user} ## Instruction: Simplify the context processor. Not very useful anymore, in fact. ## Code After: def mongodb_user(request): """ not the most usefull context manager in the world. """ if request.user.is_anonymous(): return {u'mongodb_user': None} return {u'mongodb_user': request.user.mongo}
... def mongodb_user(request): """ not the most usefull context manager in the world. """ if request.user.is_anonymous(): ... return {u'mongodb_user': None} return {u'mongodb_user': request.user.mongo} ...
f5cc3275a11c809bb6f5ab097414d0a5ccda2341
main.py
main.py
def main(): website = input("Input website(cnn, nytimes, bbc, nzherald): ") url = input("Input url: ") scraper(website, url) def scraper(website, url): print("%s, %s" % (website, url)) if __name__ == '__main__': main()
def main(): website = input("Input website(cnn, nytimes, bbc, nzherald): ") url = input("Input url: ") scraper(website, url) def scraper(website, url): if ".com" not in url: print("Invalid url") exit() print("%s, %s" % (website, url)) if __name__ == '__main__': main()
Check for .com in url
Check for .com in url
Python
mit
Alex-Gurung/ScrapeTheNews
def main(): website = input("Input website(cnn, nytimes, bbc, nzherald): ") url = input("Input url: ") scraper(website, url) def scraper(website, url): + if ".com" not in url: + print("Invalid url") + exit() print("%s, %s" % (website, url)) if __name__ == '__main__': main()
Check for .com in url
## Code Before: def main(): website = input("Input website(cnn, nytimes, bbc, nzherald): ") url = input("Input url: ") scraper(website, url) def scraper(website, url): print("%s, %s" % (website, url)) if __name__ == '__main__': main() ## Instruction: Check for .com in url ## Code After: def main(): website = input("Input website(cnn, nytimes, bbc, nzherald): ") url = input("Input url: ") scraper(website, url) def scraper(website, url): if ".com" not in url: print("Invalid url") exit() print("%s, %s" % (website, url)) if __name__ == '__main__': main()
// ... existing code ... scraper(website, url) def scraper(website, url): if ".com" not in url: print("Invalid url") exit() print("%s, %s" % (website, url)) if __name__ == '__main__': // ... rest of the code ...
c496be720461722ce482c981b4915365dd0df8ab
events/views.py
events/views.py
from django.contrib import messages from django.contrib.auth.decorators import login_required from django.views.generic.list import ListView from django.views.generic.detail import DetailView from base.util import class_view_decorator from base.views import RedirectBackView from .models import Event, EventUserRegistration class EventListView(ListView): model = Event context_object_name = 'events' class EventDetailView(DetailView): model = Event context_object_name = 'event' @class_view_decorator(login_required) class EventUserRegisterView(RedirectBackView): default_return_view = 'events_event_list' def dispatch(self, request, *args, **kwargs): event = Event.objects.get(pk=kwargs['event_id']) if event.registration_open(): registration = EventUserRegistration(user=request.user, event=event) registration.save() message = 'Successfully registered to the %s' % event messages.add_message(request, messages.INFO, message) else: message = 'Registration to the %s is not open.' % event messages.add_message(request, messages.ERROR, message) return super(EventUserRegisterView, self).dispatch(request, *args, **kwargs)
from django.contrib import messages from django.contrib.auth.decorators import login_required from django.utils.translation import ugettext_lazy as _ from django.views.generic.list import ListView from django.views.generic.detail import DetailView from base.util import class_view_decorator from base.views import RedirectBackView from .models import Event, EventUserRegistration class EventListView(ListView): model = Event context_object_name = 'events' class EventDetailView(DetailView): model = Event context_object_name = 'event' @class_view_decorator(login_required) class EventUserRegisterView(RedirectBackView): default_return_view = 'events_event_list' def dispatch(self, request, *args, **kwargs): event = Event.objects.get(pk=kwargs['event_id']) # Check if user is not already registered registrations = EventUserRegistration.objects.filter( user=request.user, event=event).count() if registrations: message = _('You are already registered to the %s') % event messages.add_message(request, messages.ERROR, message) return super(EventUserRegisterView, self).dispatch(request, *args, **kwargs) if event.registration_open(): registration = EventUserRegistration(user=request.user, event=event) registration.save() message = _('Successfully registered to the %s') % event messages.add_message(request, messages.INFO, message) else: message = _('Registration to the %s is not open.') % event messages.add_message(request, messages.ERROR, message) return super(EventUserRegisterView, self).dispatch(request, *args, **kwargs)
Raise error when user is registering to the event multiple times
events: Raise error when user is registering to the event multiple times
Python
mit
matus-stehlik/roots,rtrembecky/roots,tbabej/roots,rtrembecky/roots,matus-stehlik/roots,rtrembecky/roots,tbabej/roots,tbabej/roots,matus-stehlik/roots
from django.contrib import messages from django.contrib.auth.decorators import login_required + from django.utils.translation import ugettext_lazy as _ from django.views.generic.list import ListView from django.views.generic.detail import DetailView from base.util import class_view_decorator from base.views import RedirectBackView from .models import Event, EventUserRegistration class EventListView(ListView): model = Event context_object_name = 'events' class EventDetailView(DetailView): model = Event context_object_name = 'event' @class_view_decorator(login_required) class EventUserRegisterView(RedirectBackView): default_return_view = 'events_event_list' def dispatch(self, request, *args, **kwargs): event = Event.objects.get(pk=kwargs['event_id']) + # Check if user is not already registered + registrations = EventUserRegistration.objects.filter( + user=request.user, + event=event).count() + + if registrations: + message = _('You are already registered to the %s') % event + messages.add_message(request, messages.ERROR, message) + return super(EventUserRegisterView, self).dispatch(request, + *args, + **kwargs) + if event.registration_open(): registration = EventUserRegistration(user=request.user, event=event) registration.save() - message = 'Successfully registered to the %s' % event + message = _('Successfully registered to the %s') % event messages.add_message(request, messages.INFO, message) else: - message = 'Registration to the %s is not open.' % event + message = _('Registration to the %s is not open.') % event messages.add_message(request, messages.ERROR, message) return super(EventUserRegisterView, self).dispatch(request, *args, **kwargs)
Raise error when user is registering to the event multiple times
## Code Before: from django.contrib import messages from django.contrib.auth.decorators import login_required from django.views.generic.list import ListView from django.views.generic.detail import DetailView from base.util import class_view_decorator from base.views import RedirectBackView from .models import Event, EventUserRegistration class EventListView(ListView): model = Event context_object_name = 'events' class EventDetailView(DetailView): model = Event context_object_name = 'event' @class_view_decorator(login_required) class EventUserRegisterView(RedirectBackView): default_return_view = 'events_event_list' def dispatch(self, request, *args, **kwargs): event = Event.objects.get(pk=kwargs['event_id']) if event.registration_open(): registration = EventUserRegistration(user=request.user, event=event) registration.save() message = 'Successfully registered to the %s' % event messages.add_message(request, messages.INFO, message) else: message = 'Registration to the %s is not open.' % event messages.add_message(request, messages.ERROR, message) return super(EventUserRegisterView, self).dispatch(request, *args, **kwargs) ## Instruction: Raise error when user is registering to the event multiple times ## Code After: from django.contrib import messages from django.contrib.auth.decorators import login_required from django.utils.translation import ugettext_lazy as _ from django.views.generic.list import ListView from django.views.generic.detail import DetailView from base.util import class_view_decorator from base.views import RedirectBackView from .models import Event, EventUserRegistration class EventListView(ListView): model = Event context_object_name = 'events' class EventDetailView(DetailView): model = Event context_object_name = 'event' @class_view_decorator(login_required) class EventUserRegisterView(RedirectBackView): default_return_view = 'events_event_list' def dispatch(self, request, *args, **kwargs): event = Event.objects.get(pk=kwargs['event_id']) # Check if user is not already registered registrations = EventUserRegistration.objects.filter( user=request.user, event=event).count() if registrations: message = _('You are already registered to the %s') % event messages.add_message(request, messages.ERROR, message) return super(EventUserRegisterView, self).dispatch(request, *args, **kwargs) if event.registration_open(): registration = EventUserRegistration(user=request.user, event=event) registration.save() message = _('Successfully registered to the %s') % event messages.add_message(request, messages.INFO, message) else: message = _('Registration to the %s is not open.') % event messages.add_message(request, messages.ERROR, message) return super(EventUserRegisterView, self).dispatch(request, *args, **kwargs)
... from django.contrib import messages from django.contrib.auth.decorators import login_required from django.utils.translation import ugettext_lazy as _ from django.views.generic.list import ListView from django.views.generic.detail import DetailView ... event = Event.objects.get(pk=kwargs['event_id']) # Check if user is not already registered registrations = EventUserRegistration.objects.filter( user=request.user, event=event).count() if registrations: message = _('You are already registered to the %s') % event messages.add_message(request, messages.ERROR, message) return super(EventUserRegisterView, self).dispatch(request, *args, **kwargs) if event.registration_open(): registration = EventUserRegistration(user=request.user, event=event) ... registration.save() message = _('Successfully registered to the %s') % event messages.add_message(request, messages.INFO, message) else: message = _('Registration to the %s is not open.') % event messages.add_message(request, messages.ERROR, message) ...
60743b33e5034776576073b151c7a02dc0a40b7e
tests/unit_project/test_fields.py
tests/unit_project/test_fields.py
from djangosanetesting.cases import DatabaseTestCase from djangomarkup.fields import RichTextField from djangomarkup.models import SourceText from exampleapp.models import Article class TestRichTextField(DatabaseTestCase): def setUp(self): super(TestRichTextField, self).setUp() self.field = RichTextField( instance = Article(), model = Article, syntax_processor_name = "markdown", field_name = "text", required = True, label = "Text" ) def test_retrieve_empty_source_for_empty_article(self): self.assert_equals(u'', self.field.get_source().content) def test_source_available_for_empty_article(self): self.assert_equals(u'', self.field.get_source_text()) def test_render_available_for_empty_article(self): self.assert_equals(u'<p></p>', self.field.get_rendered_text().strip())
from djangosanetesting.cases import UnitTestCase from djangomarkup.fields import RichTextField from exampleapp.models import Article class TestRichTextField(UnitTestCase): def setUp(self): super(TestRichTextField, self).setUp() self.field = RichTextField( instance = Article(), model = Article, syntax_processor_name = "markdown", field_name = "text", required = True, label = "Text" ) def test_retrieve_empty_source_for_empty_article(self): self.assert_equals(u'', self.field.get_source().content) def test_source_available_for_empty_article(self): self.assert_equals(u'', self.field.get_source_text()) def test_render_available_for_empty_article(self): self.assert_equals(u'<p></p>', self.field.get_rendered_text().strip()) def test_value_error_raised_when_accessing_source_without_instance(self): field = RichTextField( instance = None, model = Article, syntax_processor_name = "markdown", field_name = "text", required = True, label = "Text" ) self.assert_raises(ValueError, field.get_source)
Check proper error when accessing source without instance
Check proper error when accessing source without instance
Python
bsd-3-clause
ella/django-markup
- from djangosanetesting.cases import DatabaseTestCase + from djangosanetesting.cases import UnitTestCase from djangomarkup.fields import RichTextField - from djangomarkup.models import SourceText from exampleapp.models import Article - class TestRichTextField(DatabaseTestCase): + class TestRichTextField(UnitTestCase): def setUp(self): super(TestRichTextField, self).setUp() self.field = RichTextField( instance = Article(), model = Article, syntax_processor_name = "markdown", field_name = "text", required = True, label = "Text" ) def test_retrieve_empty_source_for_empty_article(self): self.assert_equals(u'', self.field.get_source().content) def test_source_available_for_empty_article(self): self.assert_equals(u'', self.field.get_source_text()) def test_render_available_for_empty_article(self): self.assert_equals(u'<p></p>', self.field.get_rendered_text().strip()) + def test_value_error_raised_when_accessing_source_without_instance(self): + field = RichTextField( + instance = None, + model = Article, + syntax_processor_name = "markdown", + field_name = "text", + required = True, + label = "Text" + ) + self.assert_raises(ValueError, field.get_source)
Check proper error when accessing source without instance
## Code Before: from djangosanetesting.cases import DatabaseTestCase from djangomarkup.fields import RichTextField from djangomarkup.models import SourceText from exampleapp.models import Article class TestRichTextField(DatabaseTestCase): def setUp(self): super(TestRichTextField, self).setUp() self.field = RichTextField( instance = Article(), model = Article, syntax_processor_name = "markdown", field_name = "text", required = True, label = "Text" ) def test_retrieve_empty_source_for_empty_article(self): self.assert_equals(u'', self.field.get_source().content) def test_source_available_for_empty_article(self): self.assert_equals(u'', self.field.get_source_text()) def test_render_available_for_empty_article(self): self.assert_equals(u'<p></p>', self.field.get_rendered_text().strip()) ## Instruction: Check proper error when accessing source without instance ## Code After: from djangosanetesting.cases import UnitTestCase from djangomarkup.fields import RichTextField from exampleapp.models import Article class TestRichTextField(UnitTestCase): def setUp(self): super(TestRichTextField, self).setUp() self.field = RichTextField( instance = Article(), model = Article, syntax_processor_name = "markdown", field_name = "text", required = True, label = "Text" ) def test_retrieve_empty_source_for_empty_article(self): self.assert_equals(u'', self.field.get_source().content) def test_source_available_for_empty_article(self): self.assert_equals(u'', self.field.get_source_text()) def test_render_available_for_empty_article(self): self.assert_equals(u'<p></p>', self.field.get_rendered_text().strip()) def test_value_error_raised_when_accessing_source_without_instance(self): field = RichTextField( instance = None, model = Article, syntax_processor_name = "markdown", field_name = "text", required = True, label = "Text" ) self.assert_raises(ValueError, field.get_source)
# ... existing code ... from djangosanetesting.cases import UnitTestCase from djangomarkup.fields import RichTextField from exampleapp.models import Article class TestRichTextField(UnitTestCase): def setUp(self): # ... modified code ... self.assert_equals(u'<p></p>', self.field.get_rendered_text().strip()) def test_value_error_raised_when_accessing_source_without_instance(self): field = RichTextField( instance = None, model = Article, syntax_processor_name = "markdown", field_name = "text", required = True, label = "Text" ) self.assert_raises(ValueError, field.get_source) # ... rest of the code ...
252cfa3baa7973a923952ecb3c83cdfb9f28ab67
l10n_br_account/models/fiscal_document.py
l10n_br_account/models/fiscal_document.py
from odoo import api, models class FiscalDocument(models.Model): _inherit = 'l10n_br_fiscal.document' @api.multi def unlink(self): invoices = self.env['account.invoice'].search( [('fiscal_document_id', 'in', self.ids)]) invoices.unlink() return super().unlink()
from odoo import _, api, models from odoo.exceptions import UserError from odoo.addons.l10n_br_fiscal.constants.fiscal import ( SITUACAO_EDOC_EM_DIGITACAO, ) class FiscalDocument(models.Model): _inherit = 'l10n_br_fiscal.document' @api.multi def unlink(self): draft_documents = self.filtered( lambda d: d.state == SITUACAO_EDOC_EM_DIGITACAO) if draft_documents: UserError(_("You cannot delete a fiscal document " "which is not draft state.")) invoices = self.env['account.invoice'].search( [('fiscal_document_id', 'in', self.ids)]) invoices.unlink() return super().unlink()
Allow delete only fiscal documents with draft state
[REF] Allow delete only fiscal documents with draft state
Python
agpl-3.0
OCA/l10n-brazil,akretion/l10n-brazil,akretion/l10n-brazil,akretion/l10n-brazil,OCA/l10n-brazil,OCA/l10n-brazil
- from odoo import api, models + from odoo import _, api, models + from odoo.exceptions import UserError + + from odoo.addons.l10n_br_fiscal.constants.fiscal import ( + SITUACAO_EDOC_EM_DIGITACAO, + ) class FiscalDocument(models.Model): _inherit = 'l10n_br_fiscal.document' @api.multi def unlink(self): + draft_documents = self.filtered( + lambda d: d.state == SITUACAO_EDOC_EM_DIGITACAO) + + if draft_documents: + UserError(_("You cannot delete a fiscal document " + "which is not draft state.")) + invoices = self.env['account.invoice'].search( [('fiscal_document_id', 'in', self.ids)]) invoices.unlink() return super().unlink()
Allow delete only fiscal documents with draft state
## Code Before: from odoo import api, models class FiscalDocument(models.Model): _inherit = 'l10n_br_fiscal.document' @api.multi def unlink(self): invoices = self.env['account.invoice'].search( [('fiscal_document_id', 'in', self.ids)]) invoices.unlink() return super().unlink() ## Instruction: Allow delete only fiscal documents with draft state ## Code After: from odoo import _, api, models from odoo.exceptions import UserError from odoo.addons.l10n_br_fiscal.constants.fiscal import ( SITUACAO_EDOC_EM_DIGITACAO, ) class FiscalDocument(models.Model): _inherit = 'l10n_br_fiscal.document' @api.multi def unlink(self): draft_documents = self.filtered( lambda d: d.state == SITUACAO_EDOC_EM_DIGITACAO) if draft_documents: UserError(_("You cannot delete a fiscal document " "which is not draft state.")) invoices = self.env['account.invoice'].search( [('fiscal_document_id', 'in', self.ids)]) invoices.unlink() return super().unlink()
// ... existing code ... from odoo import _, api, models from odoo.exceptions import UserError from odoo.addons.l10n_br_fiscal.constants.fiscal import ( SITUACAO_EDOC_EM_DIGITACAO, ) // ... modified code ... @api.multi def unlink(self): draft_documents = self.filtered( lambda d: d.state == SITUACAO_EDOC_EM_DIGITACAO) if draft_documents: UserError(_("You cannot delete a fiscal document " "which is not draft state.")) invoices = self.env['account.invoice'].search( [('fiscal_document_id', 'in', self.ids)]) // ... rest of the code ...
1ed7d695eff134557990d8b1a5dffa51b6d1d2f6
distarray/run_tests.py
distarray/run_tests.py
from __future__ import print_function import os import sys import shlex import subprocess import distarray def _run_shell_command(specific_cmd): """Run a command with subprocess and pass the results through to stdout. First, change directory to the project directory. """ path = os.path.split(os.path.split(distarray.__file__)[0])[0] os.chdir(path) proc = subprocess.Popen(shlex.split(specific_cmd), stdout=subprocess.PIPE, stderr=subprocess.STDOUT) while True: char = proc.stdout.read(1).decode() if not char: break else: print(char, end="") sys.stdout.flush() def test(): """Run all DistArray tests.""" cmd = "make test" _run_shell_command(cmd) if __name__ == "__main__": test()
from __future__ import print_function import os import sys import shlex import subprocess import distarray def _run_shell_command(specific_cmd): """Run a command with subprocess and pass the results through to stdout. First, change directory to the project directory. """ path = os.path.split(os.path.split(distarray.__file__)[0])[0] os.chdir(path) proc = subprocess.Popen(shlex.split(specific_cmd), stdout=subprocess.PIPE, stderr=subprocess.STDOUT) while True: char = proc.stdout.read(1).decode() if not char: return proc.wait() else: print(char, end="") sys.stdout.flush() def test(): """Run all DistArray tests.""" cmd = "make test" return _run_shell_command(cmd) if __name__ == "__main__": sys.exit(test())
Return returncode from shell command.
Return returncode from shell command.
Python
bsd-3-clause
enthought/distarray,RaoUmer/distarray,RaoUmer/distarray,enthought/distarray
from __future__ import print_function import os import sys import shlex import subprocess import distarray def _run_shell_command(specific_cmd): """Run a command with subprocess and pass the results through to stdout. First, change directory to the project directory. """ path = os.path.split(os.path.split(distarray.__file__)[0])[0] os.chdir(path) proc = subprocess.Popen(shlex.split(specific_cmd), stdout=subprocess.PIPE, stderr=subprocess.STDOUT) while True: char = proc.stdout.read(1).decode() if not char: - break + return proc.wait() else: print(char, end="") sys.stdout.flush() def test(): """Run all DistArray tests.""" cmd = "make test" - _run_shell_command(cmd) + return _run_shell_command(cmd) if __name__ == "__main__": - test() + sys.exit(test())
Return returncode from shell command.
## Code Before: from __future__ import print_function import os import sys import shlex import subprocess import distarray def _run_shell_command(specific_cmd): """Run a command with subprocess and pass the results through to stdout. First, change directory to the project directory. """ path = os.path.split(os.path.split(distarray.__file__)[0])[0] os.chdir(path) proc = subprocess.Popen(shlex.split(specific_cmd), stdout=subprocess.PIPE, stderr=subprocess.STDOUT) while True: char = proc.stdout.read(1).decode() if not char: break else: print(char, end="") sys.stdout.flush() def test(): """Run all DistArray tests.""" cmd = "make test" _run_shell_command(cmd) if __name__ == "__main__": test() ## Instruction: Return returncode from shell command. ## Code After: from __future__ import print_function import os import sys import shlex import subprocess import distarray def _run_shell_command(specific_cmd): """Run a command with subprocess and pass the results through to stdout. First, change directory to the project directory. """ path = os.path.split(os.path.split(distarray.__file__)[0])[0] os.chdir(path) proc = subprocess.Popen(shlex.split(specific_cmd), stdout=subprocess.PIPE, stderr=subprocess.STDOUT) while True: char = proc.stdout.read(1).decode() if not char: return proc.wait() else: print(char, end="") sys.stdout.flush() def test(): """Run all DistArray tests.""" cmd = "make test" return _run_shell_command(cmd) if __name__ == "__main__": sys.exit(test())
# ... existing code ... char = proc.stdout.read(1).decode() if not char: return proc.wait() else: print(char, end="") # ... modified code ... """Run all DistArray tests.""" cmd = "make test" return _run_shell_command(cmd) if __name__ == "__main__": sys.exit(test()) # ... rest of the code ...
73aa38a5d481a26278dd29364f16839cad0f22cf
manager/projects/ui/views/files.py
manager/projects/ui/views/files.py
from django.contrib.auth.decorators import login_required from django.http import HttpRequest, HttpResponse from django.shortcuts import render from projects.api.views.files import ProjectsFilesViewSet @login_required def list(request: HttpRequest, *args, **kwargs) -> HttpResponse: """ Get a list of project files. The trailing part of the URL becomes the `prefix` query parameter, consistent with API ending e.g. /<account>/<project>/files/sub?search=foo is equivalent to: /api/projects/<project>/files?prefix=sub&search=foo """ prefix = kwargs.get("prefix") if prefix and not prefix.endswith("/"): prefix += "/" request.GET = request.GET.copy() request.GET["prefix"] = prefix request.GET["aggregate"] = True viewset = ProjectsFilesViewSet.init("list", request, args, kwargs) project = viewset.get_project() files = viewset.get_queryset(project) # List of tuples for directory breadcrumbs dirs = [("root", "")] path = "" for name in prefix.split("/"): if name: path += name + "/" dirs.append((name, path)) return render( request, "projects/files/list.html", dict(prefix=prefix, dirs=dirs, files=files, project=project,), )
from django.contrib.auth.decorators import login_required from django.http import HttpRequest, HttpResponse from django.shortcuts import render from projects.api.views.files import ProjectsFilesViewSet @login_required def list(request: HttpRequest, *args, **kwargs) -> HttpResponse: """ Get a list of project files. The trailing part of the URL becomes the `prefix` query parameter, consistent with API ending e.g. /<account>/<project>/files/sub?search=foo is equivalent to: /api/projects/<project>/files?prefix=sub&search=foo """ request.GET = request.GET.copy() request.GET["prefix"] = kwargs.get("prefix") viewset = ProjectsFilesViewSet.init("list", request, args, kwargs) project = viewset.get_project() files = viewset.get_queryset(project) context = viewset.get_response_context(queryset=files) return render( request, "projects/files/list.html", dict(project=project, **context), )
Update view for change in viewset
refactor(Files): Update view for change in viewset
Python
apache-2.0
stencila/hub,stencila/hub,stencila/hub,stencila/hub,stencila/hub
from django.contrib.auth.decorators import login_required from django.http import HttpRequest, HttpResponse from django.shortcuts import render from projects.api.views.files import ProjectsFilesViewSet @login_required def list(request: HttpRequest, *args, **kwargs) -> HttpResponse: """ Get a list of project files. The trailing part of the URL becomes the `prefix` query parameter, consistent with API ending e.g. /<account>/<project>/files/sub?search=foo is equivalent to: /api/projects/<project>/files?prefix=sub&search=foo """ - prefix = kwargs.get("prefix") - if prefix and not prefix.endswith("/"): - prefix += "/" - request.GET = request.GET.copy() - request.GET["prefix"] = prefix + request.GET["prefix"] = kwargs.get("prefix") - request.GET["aggregate"] = True viewset = ProjectsFilesViewSet.init("list", request, args, kwargs) project = viewset.get_project() files = viewset.get_queryset(project) + context = viewset.get_response_context(queryset=files) - - # List of tuples for directory breadcrumbs - dirs = [("root", "")] - path = "" - for name in prefix.split("/"): - if name: - path += name + "/" - dirs.append((name, path)) return render( + request, "projects/files/list.html", dict(project=project, **context), - request, - "projects/files/list.html", - dict(prefix=prefix, dirs=dirs, files=files, project=project,), )
Update view for change in viewset
## Code Before: from django.contrib.auth.decorators import login_required from django.http import HttpRequest, HttpResponse from django.shortcuts import render from projects.api.views.files import ProjectsFilesViewSet @login_required def list(request: HttpRequest, *args, **kwargs) -> HttpResponse: """ Get a list of project files. The trailing part of the URL becomes the `prefix` query parameter, consistent with API ending e.g. /<account>/<project>/files/sub?search=foo is equivalent to: /api/projects/<project>/files?prefix=sub&search=foo """ prefix = kwargs.get("prefix") if prefix and not prefix.endswith("/"): prefix += "/" request.GET = request.GET.copy() request.GET["prefix"] = prefix request.GET["aggregate"] = True viewset = ProjectsFilesViewSet.init("list", request, args, kwargs) project = viewset.get_project() files = viewset.get_queryset(project) # List of tuples for directory breadcrumbs dirs = [("root", "")] path = "" for name in prefix.split("/"): if name: path += name + "/" dirs.append((name, path)) return render( request, "projects/files/list.html", dict(prefix=prefix, dirs=dirs, files=files, project=project,), ) ## Instruction: Update view for change in viewset ## Code After: from django.contrib.auth.decorators import login_required from django.http import HttpRequest, HttpResponse from django.shortcuts import render from projects.api.views.files import ProjectsFilesViewSet @login_required def list(request: HttpRequest, *args, **kwargs) -> HttpResponse: """ Get a list of project files. The trailing part of the URL becomes the `prefix` query parameter, consistent with API ending e.g. /<account>/<project>/files/sub?search=foo is equivalent to: /api/projects/<project>/files?prefix=sub&search=foo """ request.GET = request.GET.copy() request.GET["prefix"] = kwargs.get("prefix") viewset = ProjectsFilesViewSet.init("list", request, args, kwargs) project = viewset.get_project() files = viewset.get_queryset(project) context = viewset.get_response_context(queryset=files) return render( request, "projects/files/list.html", dict(project=project, **context), )
// ... existing code ... /api/projects/<project>/files?prefix=sub&search=foo """ request.GET = request.GET.copy() request.GET["prefix"] = kwargs.get("prefix") viewset = ProjectsFilesViewSet.init("list", request, args, kwargs) // ... modified code ... project = viewset.get_project() files = viewset.get_queryset(project) context = viewset.get_response_context(queryset=files) return render( request, "projects/files/list.html", dict(project=project, **context), ) // ... rest of the code ...
154b64b2ee56fa4391251268ba4a85d178bedd60
djangoautoconf/urls.py
djangoautoconf/urls.py
from django.conf.urls import patterns, include, url from django.conf import settings from django.conf.urls.static import static # Uncomment the next two lines to enable the admin: from django.contrib import admin # from mezzanine.core.views import direct_to_template admin.autodiscover() # Must be defined before auto discover and urlpatterns var. So when there is root url # injection, we first insert root url to this, then the last line will insert it to real urlpatterns default_app_url_patterns = [] from djangoautoconf import auto_conf_urls auto_conf_urls.autodiscover() urlpatterns = patterns('', # Examples: # url(r'^default_django_15_and_below/', include('default_django_15_and_below.foo.urls')), # Uncomment the admin/doc line below to enable admin documentation: url(r'^admin/doc/', include('django.contrib.admindocs.urls')), # Uncomment the next line to enable the admin: url(r'^admin/', include(admin.site.urls)), # url(r'^', include('demo.urls')), # url(r'^obj_sys/', include('obj_sys.urls')), # url("^$", direct_to_template, {"template": "index.html"}, name="home"), ) urlpatterns = [ # ... the rest of your URLconf goes here ... ] + static(settings.MEDIA_URL, document_root=settings.MEDIA_ROOT) urlpatterns += default_app_url_patterns
from django.conf.urls import patterns, include, url from django.conf import settings from django.conf.urls.static import static # Uncomment the next two lines to enable the admin: from django.contrib import admin # from mezzanine.core.views import direct_to_template admin.autodiscover() # Must be defined before auto discover and urlpatterns var. So when there is root url # injection, we first insert root url to this, then the last line will insert it to real urlpatterns default_app_url_patterns = [] from djangoautoconf import auto_conf_urls auto_conf_urls.autodiscover() urlpatterns = patterns('', # Examples: # url(r'^default_django_15_and_below/', include('default_django_15_and_below.foo.urls')), # Uncomment the admin/doc line below to enable admin documentation: url(r'^admin/doc/', include('django.contrib.admindocs.urls')), # Uncomment the next line to enable the admin: url(r'^admin/', include(admin.site.urls)), # url(r'^', include('demo.urls')), # url(r'^obj_sys/', include('obj_sys.urls')), # url("^$", direct_to_template, {"template": "index.html"}, name="home"), ) urlpatterns += static(settings.MEDIA_URL, document_root=settings.MEDIA_ROOT) urlpatterns += default_app_url_patterns
Fix the issue of override url by mistake.
Fix the issue of override url by mistake.
Python
bsd-3-clause
weijia/djangoautoconf,weijia/djangoautoconf
from django.conf.urls import patterns, include, url from django.conf import settings from django.conf.urls.static import static # Uncomment the next two lines to enable the admin: from django.contrib import admin # from mezzanine.core.views import direct_to_template admin.autodiscover() # Must be defined before auto discover and urlpatterns var. So when there is root url # injection, we first insert root url to this, then the last line will insert it to real urlpatterns default_app_url_patterns = [] from djangoautoconf import auto_conf_urls auto_conf_urls.autodiscover() urlpatterns = patterns('', # Examples: # url(r'^default_django_15_and_below/', include('default_django_15_and_below.foo.urls')), # Uncomment the admin/doc line below to enable admin documentation: url(r'^admin/doc/', include('django.contrib.admindocs.urls')), # Uncomment the next line to enable the admin: url(r'^admin/', include(admin.site.urls)), # url(r'^', include('demo.urls')), # url(r'^obj_sys/', include('obj_sys.urls')), # url("^$", direct_to_template, {"template": "index.html"}, name="home"), ) - urlpatterns = [ - # ... the rest of your URLconf goes here ... - ] + static(settings.MEDIA_URL, document_root=settings.MEDIA_ROOT) + urlpatterns += static(settings.MEDIA_URL, document_root=settings.MEDIA_ROOT) urlpatterns += default_app_url_patterns
Fix the issue of override url by mistake.
## Code Before: from django.conf.urls import patterns, include, url from django.conf import settings from django.conf.urls.static import static # Uncomment the next two lines to enable the admin: from django.contrib import admin # from mezzanine.core.views import direct_to_template admin.autodiscover() # Must be defined before auto discover and urlpatterns var. So when there is root url # injection, we first insert root url to this, then the last line will insert it to real urlpatterns default_app_url_patterns = [] from djangoautoconf import auto_conf_urls auto_conf_urls.autodiscover() urlpatterns = patterns('', # Examples: # url(r'^default_django_15_and_below/', include('default_django_15_and_below.foo.urls')), # Uncomment the admin/doc line below to enable admin documentation: url(r'^admin/doc/', include('django.contrib.admindocs.urls')), # Uncomment the next line to enable the admin: url(r'^admin/', include(admin.site.urls)), # url(r'^', include('demo.urls')), # url(r'^obj_sys/', include('obj_sys.urls')), # url("^$", direct_to_template, {"template": "index.html"}, name="home"), ) urlpatterns = [ # ... the rest of your URLconf goes here ... ] + static(settings.MEDIA_URL, document_root=settings.MEDIA_ROOT) urlpatterns += default_app_url_patterns ## Instruction: Fix the issue of override url by mistake. ## Code After: from django.conf.urls import patterns, include, url from django.conf import settings from django.conf.urls.static import static # Uncomment the next two lines to enable the admin: from django.contrib import admin # from mezzanine.core.views import direct_to_template admin.autodiscover() # Must be defined before auto discover and urlpatterns var. So when there is root url # injection, we first insert root url to this, then the last line will insert it to real urlpatterns default_app_url_patterns = [] from djangoautoconf import auto_conf_urls auto_conf_urls.autodiscover() urlpatterns = patterns('', # Examples: # url(r'^default_django_15_and_below/', include('default_django_15_and_below.foo.urls')), # Uncomment the admin/doc line below to enable admin documentation: url(r'^admin/doc/', include('django.contrib.admindocs.urls')), # Uncomment the next line to enable the admin: url(r'^admin/', include(admin.site.urls)), # url(r'^', include('demo.urls')), # url(r'^obj_sys/', include('obj_sys.urls')), # url("^$", direct_to_template, {"template": "index.html"}, name="home"), ) urlpatterns += static(settings.MEDIA_URL, document_root=settings.MEDIA_ROOT) urlpatterns += default_app_url_patterns
... # url("^$", direct_to_template, {"template": "index.html"}, name="home"), ) urlpatterns += static(settings.MEDIA_URL, document_root=settings.MEDIA_ROOT) urlpatterns += default_app_url_patterns ...
0213bbb8f8075b2dc36a33380a66932c9d541f63
src/sphobjinv/__init__.py
src/sphobjinv/__init__.py
from sphobjinv.data import DataFields, DataObjBytes, DataObjStr from sphobjinv.enum import HeaderFields, SourceTypes from sphobjinv.error import ( SphobjinvError, VersionError, ) from sphobjinv.fileops import readbytes, readjson, urlwalk, writebytes, writejson from sphobjinv.inventory import Inventory from sphobjinv.re import p_data, pb_comments, pb_data, pb_project, pb_version from sphobjinv.schema import json_schema from sphobjinv.version import __version__ from sphobjinv.zlib import compress, decompress
from sphobjinv.data import DataFields, DataObjBytes, DataObjStr from sphobjinv.enum import HeaderFields, SourceTypes from sphobjinv.error import SphobjinvError, VersionError from sphobjinv.fileops import readbytes, readjson, urlwalk, writebytes, writejson from sphobjinv.inventory import Inventory from sphobjinv.re import p_data, pb_comments, pb_data, pb_project, pb_version from sphobjinv.schema import json_schema from sphobjinv.version import __version__ from sphobjinv.zlib import compress, decompress
Clean up the error imports
Clean up the error imports The new errors that had been added for _intersphinx.py had left the sphobjinv.error import line split. No need, when it all fits on one line.
Python
mit
bskinn/sphobjinv
from sphobjinv.data import DataFields, DataObjBytes, DataObjStr from sphobjinv.enum import HeaderFields, SourceTypes + from sphobjinv.error import SphobjinvError, VersionError - from sphobjinv.error import ( - SphobjinvError, - VersionError, - ) from sphobjinv.fileops import readbytes, readjson, urlwalk, writebytes, writejson from sphobjinv.inventory import Inventory from sphobjinv.re import p_data, pb_comments, pb_data, pb_project, pb_version from sphobjinv.schema import json_schema from sphobjinv.version import __version__ from sphobjinv.zlib import compress, decompress
Clean up the error imports
## Code Before: from sphobjinv.data import DataFields, DataObjBytes, DataObjStr from sphobjinv.enum import HeaderFields, SourceTypes from sphobjinv.error import ( SphobjinvError, VersionError, ) from sphobjinv.fileops import readbytes, readjson, urlwalk, writebytes, writejson from sphobjinv.inventory import Inventory from sphobjinv.re import p_data, pb_comments, pb_data, pb_project, pb_version from sphobjinv.schema import json_schema from sphobjinv.version import __version__ from sphobjinv.zlib import compress, decompress ## Instruction: Clean up the error imports ## Code After: from sphobjinv.data import DataFields, DataObjBytes, DataObjStr from sphobjinv.enum import HeaderFields, SourceTypes from sphobjinv.error import SphobjinvError, VersionError from sphobjinv.fileops import readbytes, readjson, urlwalk, writebytes, writejson from sphobjinv.inventory import Inventory from sphobjinv.re import p_data, pb_comments, pb_data, pb_project, pb_version from sphobjinv.schema import json_schema from sphobjinv.version import __version__ from sphobjinv.zlib import compress, decompress
... from sphobjinv.data import DataFields, DataObjBytes, DataObjStr from sphobjinv.enum import HeaderFields, SourceTypes from sphobjinv.error import SphobjinvError, VersionError from sphobjinv.fileops import readbytes, readjson, urlwalk, writebytes, writejson from sphobjinv.inventory import Inventory ...
785fcdca3c9bfb908444d3b9339457c616761f2c
tests/flights_to_test.py
tests/flights_to_test.py
import unittest import datetime import json import sys sys.path.append('..') import sabre_dev_studio import sabre_dev_studio.sabre_exceptions as sabre_exceptions ''' requires config.json in the same directory for api authentication { "sabre_client_id": -----, "sabre_client_secret": ----- } ''' class TestBasicInstaflights(unittest.TestCase): def read_config(self): raw_data = open('config.json').read() data = json.loads(raw_data) client_secret = data['sabre_client_secret'] client_id = data['sabre_client_id'] return (client_id, client_secret) def setUp(self): # Read from config self.client_id, self.client_secret = self.read_config() self.sds = sabre_dev_studio.SabreDevStudio() self.sds.set_credentials(self.client_id, self.client_secret) self.sds.authenticate() def test_basic_request(self): city = 'YTO' instaf = self.sds.flights_to(city) self.assertIsNotNone(instaf) def test_no_authorization(self): sds = sabre_dev_studio.SabreDevStudio() with self.assertRaises(sabre_exceptions.NotAuthorizedError): resp = sds.flights_to('YTO') if __name__ == '__main__': unittest.main()
import unittest import datetime import json import sys sys.path.append('..') import sabre_dev_studio import sabre_dev_studio.sabre_exceptions as sabre_exceptions ''' requires config.json in the same directory for api authentication { "sabre_client_id": -----, "sabre_client_secret": ----- } ''' class TestBasicFlightsTo(unittest.TestCase): def read_config(self): raw_data = open('config.json').read() data = json.loads(raw_data) client_secret = data['sabre_client_secret'] client_id = data['sabre_client_id'] return (client_id, client_secret) def setUp(self): # Read from config self.client_id, self.client_secret = self.read_config() self.sds = sabre_dev_studio.SabreDevStudio() self.sds.set_credentials(self.client_id, self.client_secret) self.sds.authenticate() def test_basic_request(self): city = 'YTO' flights_to_city = self.sds.flights_to(city) print(flights_to_city) self.assertIsNotNone(flights_to_city) def test_no_authorization(self): sds = sabre_dev_studio.SabreDevStudio() with self.assertRaises(sabre_exceptions.NotAuthorizedError): resp = sds.flights_to('YTO') if __name__ == '__main__': unittest.main()
Change instaflights name in flights_to tests
Change instaflights name in flights_to tests
Python
mit
Jamil/sabre_dev_studio
import unittest import datetime import json import sys sys.path.append('..') import sabre_dev_studio import sabre_dev_studio.sabre_exceptions as sabre_exceptions ''' requires config.json in the same directory for api authentication { "sabre_client_id": -----, "sabre_client_secret": ----- } ''' - class TestBasicInstaflights(unittest.TestCase): + class TestBasicFlightsTo(unittest.TestCase): def read_config(self): raw_data = open('config.json').read() data = json.loads(raw_data) client_secret = data['sabre_client_secret'] client_id = data['sabre_client_id'] return (client_id, client_secret) def setUp(self): # Read from config self.client_id, self.client_secret = self.read_config() self.sds = sabre_dev_studio.SabreDevStudio() self.sds.set_credentials(self.client_id, self.client_secret) self.sds.authenticate() def test_basic_request(self): city = 'YTO' - instaf = self.sds.flights_to(city) + flights_to_city = self.sds.flights_to(city) + print(flights_to_city) - self.assertIsNotNone(instaf) + self.assertIsNotNone(flights_to_city) def test_no_authorization(self): sds = sabre_dev_studio.SabreDevStudio() with self.assertRaises(sabre_exceptions.NotAuthorizedError): resp = sds.flights_to('YTO') if __name__ == '__main__': unittest.main()
Change instaflights name in flights_to tests
## Code Before: import unittest import datetime import json import sys sys.path.append('..') import sabre_dev_studio import sabre_dev_studio.sabre_exceptions as sabre_exceptions ''' requires config.json in the same directory for api authentication { "sabre_client_id": -----, "sabre_client_secret": ----- } ''' class TestBasicInstaflights(unittest.TestCase): def read_config(self): raw_data = open('config.json').read() data = json.loads(raw_data) client_secret = data['sabre_client_secret'] client_id = data['sabre_client_id'] return (client_id, client_secret) def setUp(self): # Read from config self.client_id, self.client_secret = self.read_config() self.sds = sabre_dev_studio.SabreDevStudio() self.sds.set_credentials(self.client_id, self.client_secret) self.sds.authenticate() def test_basic_request(self): city = 'YTO' instaf = self.sds.flights_to(city) self.assertIsNotNone(instaf) def test_no_authorization(self): sds = sabre_dev_studio.SabreDevStudio() with self.assertRaises(sabre_exceptions.NotAuthorizedError): resp = sds.flights_to('YTO') if __name__ == '__main__': unittest.main() ## Instruction: Change instaflights name in flights_to tests ## Code After: import unittest import datetime import json import sys sys.path.append('..') import sabre_dev_studio import sabre_dev_studio.sabre_exceptions as sabre_exceptions ''' requires config.json in the same directory for api authentication { "sabre_client_id": -----, "sabre_client_secret": ----- } ''' class TestBasicFlightsTo(unittest.TestCase): def read_config(self): raw_data = open('config.json').read() data = json.loads(raw_data) client_secret = data['sabre_client_secret'] client_id = data['sabre_client_id'] return (client_id, client_secret) def setUp(self): # Read from config self.client_id, self.client_secret = self.read_config() self.sds = sabre_dev_studio.SabreDevStudio() self.sds.set_credentials(self.client_id, self.client_secret) self.sds.authenticate() def test_basic_request(self): city = 'YTO' flights_to_city = self.sds.flights_to(city) print(flights_to_city) self.assertIsNotNone(flights_to_city) def test_no_authorization(self): sds = sabre_dev_studio.SabreDevStudio() with self.assertRaises(sabre_exceptions.NotAuthorizedError): resp = sds.flights_to('YTO') if __name__ == '__main__': unittest.main()
... ''' class TestBasicFlightsTo(unittest.TestCase): def read_config(self): raw_data = open('config.json').read() ... def test_basic_request(self): city = 'YTO' flights_to_city = self.sds.flights_to(city) print(flights_to_city) self.assertIsNotNone(flights_to_city) def test_no_authorization(self): ...
a24b2b303c1cd5e9f43353d55cc6b9d07b37b7f4
ephemeral-cluster.py
ephemeral-cluster.py
import subprocess import sys import uuid usage = """\ Run a command using a temporary docker-compose cluster, removing all containers \ and images after command completion (regardless of success or failure.) Generally, this would be used with the ``run`` command to provide a clean room \ testing environment. """ if not sys.argv[1:]: sys.stderr.write(usage) sys.exit(1) project = uuid.uuid1().hex sys.stderr.write('Starting ephemeral cluster: {0}\n'.format(project)) try: sys.exit(subprocess.check_call(['docker-compose', '-p', project] + sys.argv[1:])) finally: subprocess.check_call(['docker-compose', '-p', project, 'stop']) subprocess.check_call(['docker-compose', '-p', project, 'rm', '-f', '-v'])
import subprocess import sys import uuid usage = """\ Run a command using a temporary docker-compose cluster, removing all containers \ and associated volumes after command completion (regardless of success or \ failure.) Generally, this would be used with the ``run`` command to provide a clean room \ testing environment. """ if not sys.argv[1:]: sys.stderr.write(usage) sys.exit(1) project = uuid.uuid1().hex sys.stderr.write('Setting up ephemeral cluster ({0})...\n'.format(project)) try: subprocess.check_call(['docker-compose', '-p', project] + sys.argv[1:]) except subprocess.CalledProcessError as error: raise SystemExit(error.returncode) finally: sys.stderr.write('\nCleaning up ephemeral cluster ({0})...\n'.format(project)) subprocess.check_call(['docker-compose', '-p', project, 'stop']) subprocess.check_call(['docker-compose', '-p', project, 'rm', '-f', '-v'])
Fix forwarding ephemeral cluster exit code.
Fix forwarding ephemeral cluster exit code. Summary: Also improves logging a little bit. Test Plan: $ python ephemeral-cluster.py run --rm --entrypoint=bash pgshovel -c "exit 10" $ test $? -eq 10 Reviewers: jeff, tail Reviewed By: tail Differential Revision: http://phabricator.local.disqus.net/D19564
Python
apache-2.0
fuziontech/pgshovel,disqus/pgshovel,fuziontech/pgshovel,fuziontech/pgshovel,disqus/pgshovel
import subprocess import sys import uuid usage = """\ Run a command using a temporary docker-compose cluster, removing all containers \ - and images after command completion (regardless of success or failure.) + and associated volumes after command completion (regardless of success or \ + failure.) Generally, this would be used with the ``run`` command to provide a clean room \ testing environment. """ if not sys.argv[1:]: sys.stderr.write(usage) sys.exit(1) project = uuid.uuid1().hex - sys.stderr.write('Starting ephemeral cluster: {0}\n'.format(project)) + sys.stderr.write('Setting up ephemeral cluster ({0})...\n'.format(project)) try: - sys.exit(subprocess.check_call(['docker-compose', '-p', project] + sys.argv[1:])) + subprocess.check_call(['docker-compose', '-p', project] + sys.argv[1:]) + except subprocess.CalledProcessError as error: + raise SystemExit(error.returncode) finally: + sys.stderr.write('\nCleaning up ephemeral cluster ({0})...\n'.format(project)) subprocess.check_call(['docker-compose', '-p', project, 'stop']) subprocess.check_call(['docker-compose', '-p', project, 'rm', '-f', '-v'])
Fix forwarding ephemeral cluster exit code.
## Code Before: import subprocess import sys import uuid usage = """\ Run a command using a temporary docker-compose cluster, removing all containers \ and images after command completion (regardless of success or failure.) Generally, this would be used with the ``run`` command to provide a clean room \ testing environment. """ if not sys.argv[1:]: sys.stderr.write(usage) sys.exit(1) project = uuid.uuid1().hex sys.stderr.write('Starting ephemeral cluster: {0}\n'.format(project)) try: sys.exit(subprocess.check_call(['docker-compose', '-p', project] + sys.argv[1:])) finally: subprocess.check_call(['docker-compose', '-p', project, 'stop']) subprocess.check_call(['docker-compose', '-p', project, 'rm', '-f', '-v']) ## Instruction: Fix forwarding ephemeral cluster exit code. ## Code After: import subprocess import sys import uuid usage = """\ Run a command using a temporary docker-compose cluster, removing all containers \ and associated volumes after command completion (regardless of success or \ failure.) Generally, this would be used with the ``run`` command to provide a clean room \ testing environment. """ if not sys.argv[1:]: sys.stderr.write(usage) sys.exit(1) project = uuid.uuid1().hex sys.stderr.write('Setting up ephemeral cluster ({0})...\n'.format(project)) try: subprocess.check_call(['docker-compose', '-p', project] + sys.argv[1:]) except subprocess.CalledProcessError as error: raise SystemExit(error.returncode) finally: sys.stderr.write('\nCleaning up ephemeral cluster ({0})...\n'.format(project)) subprocess.check_call(['docker-compose', '-p', project, 'stop']) subprocess.check_call(['docker-compose', '-p', project, 'rm', '-f', '-v'])
# ... existing code ... usage = """\ Run a command using a temporary docker-compose cluster, removing all containers \ and associated volumes after command completion (regardless of success or \ failure.) Generally, this would be used with the ``run`` command to provide a clean room \ # ... modified code ... project = uuid.uuid1().hex sys.stderr.write('Setting up ephemeral cluster ({0})...\n'.format(project)) try: subprocess.check_call(['docker-compose', '-p', project] + sys.argv[1:]) except subprocess.CalledProcessError as error: raise SystemExit(error.returncode) finally: sys.stderr.write('\nCleaning up ephemeral cluster ({0})...\n'.format(project)) subprocess.check_call(['docker-compose', '-p', project, 'stop']) subprocess.check_call(['docker-compose', '-p', project, 'rm', '-f', '-v']) # ... rest of the code ...
a3c52c84da93c3e3007fa291213b97fd7d5b0e8f
tests.py
tests.py
import TwitterSA import unittest class TwitterSATestCase(unittest.TestCase): def setUp(self): TwitterSA.app.config['TESTING'] = True self.app = TwitterSA.app.test_client() def tearDown(self): pass def test_invalid_search_query(self): """Test for invalid search queries""" rv = self.app.get('/search?q=') assert 'Invalid search query' in rv.data rv = self.app.get('/search?nonsense=nonsense') assert 'Invalid search query' in rv.data def test_invalid_user_id(self): """Test for invalid user ids""" rv = self.app.get('/user?uid=') assert 'Invalid user id' in rv.data rv = self.app.get('/user?nonsense=nonsense') assert 'Invalid user id' in rv.data if __name__ == '__main__': unittest.main()
import TwitterSA import unittest class TwitterSATestCase(unittest.TestCase): def setUp(self): TwitterSA.app.config['TESTING'] = True self.app = TwitterSA.app.test_client() def tearDown(self): pass def test_twitter_api(self): """Test to make sure the API is getting tweets""" tweets = TwitterSA.api.search(q='hello') assert tweets and len(tweets) def test_invalid_search_query(self): """Test for invalid search queries""" rv = self.app.get('/search?q=') assert 'Invalid search query' in rv.data rv = self.app.get('/search?nonsense=nonsense') assert 'Invalid search query' in rv.data def test_invalid_user_id(self): """Test for invalid user ids""" rv = self.app.get('/user?uid=') assert 'Invalid user id' in rv.data rv = self.app.get('/user?nonsense=nonsense') assert 'Invalid user id' in rv.data if __name__ == '__main__': unittest.main()
Add twitter API functionality test
Add twitter API functionality test
Python
mit
jayelm/twittersa,jayelm/twittersa
import TwitterSA import unittest class TwitterSATestCase(unittest.TestCase): def setUp(self): TwitterSA.app.config['TESTING'] = True self.app = TwitterSA.app.test_client() def tearDown(self): pass + + def test_twitter_api(self): + """Test to make sure the API is getting tweets""" + tweets = TwitterSA.api.search(q='hello') + assert tweets and len(tweets) def test_invalid_search_query(self): """Test for invalid search queries""" rv = self.app.get('/search?q=') assert 'Invalid search query' in rv.data rv = self.app.get('/search?nonsense=nonsense') assert 'Invalid search query' in rv.data def test_invalid_user_id(self): """Test for invalid user ids""" rv = self.app.get('/user?uid=') assert 'Invalid user id' in rv.data rv = self.app.get('/user?nonsense=nonsense') assert 'Invalid user id' in rv.data if __name__ == '__main__': unittest.main()
Add twitter API functionality test
## Code Before: import TwitterSA import unittest class TwitterSATestCase(unittest.TestCase): def setUp(self): TwitterSA.app.config['TESTING'] = True self.app = TwitterSA.app.test_client() def tearDown(self): pass def test_invalid_search_query(self): """Test for invalid search queries""" rv = self.app.get('/search?q=') assert 'Invalid search query' in rv.data rv = self.app.get('/search?nonsense=nonsense') assert 'Invalid search query' in rv.data def test_invalid_user_id(self): """Test for invalid user ids""" rv = self.app.get('/user?uid=') assert 'Invalid user id' in rv.data rv = self.app.get('/user?nonsense=nonsense') assert 'Invalid user id' in rv.data if __name__ == '__main__': unittest.main() ## Instruction: Add twitter API functionality test ## Code After: import TwitterSA import unittest class TwitterSATestCase(unittest.TestCase): def setUp(self): TwitterSA.app.config['TESTING'] = True self.app = TwitterSA.app.test_client() def tearDown(self): pass def test_twitter_api(self): """Test to make sure the API is getting tweets""" tweets = TwitterSA.api.search(q='hello') assert tweets and len(tweets) def test_invalid_search_query(self): """Test for invalid search queries""" rv = self.app.get('/search?q=') assert 'Invalid search query' in rv.data rv = self.app.get('/search?nonsense=nonsense') assert 'Invalid search query' in rv.data def test_invalid_user_id(self): """Test for invalid user ids""" rv = self.app.get('/user?uid=') assert 'Invalid user id' in rv.data rv = self.app.get('/user?nonsense=nonsense') assert 'Invalid user id' in rv.data if __name__ == '__main__': unittest.main()
// ... existing code ... def tearDown(self): pass def test_twitter_api(self): """Test to make sure the API is getting tweets""" tweets = TwitterSA.api.search(q='hello') assert tweets and len(tweets) def test_invalid_search_query(self): // ... rest of the code ...
e4ee7034291fbeda48efa0d1c617be8a20eb49bd
algorithms/python/496_next_greater_element.py
algorithms/python/496_next_greater_element.py
class Solution(object): def nextGreaterElement(self, findNums, nums): """ :type findNums: List[int] :type nums: List[int] :rtype: List[int] """ results = [] for findNum in findNums: index = nums.index(findNum) result = index + 1 for candidate in nums[index + 1:]: if candidate > findNum: results.append(candidate) break else: result += 1 if result >= len(nums): results.append(-1) return results
class Solution(object): def nextGreaterElement(self, findNums, nums): """ :type findNums: List[int] :type nums: List[int] :rtype: List[int] """ results = [] for findNum in findNums: index = nums.index(findNum) result = index + 1 for candidate in nums[index + 1:]: if candidate > findNum: results.append(candidate) break else: result += 1 if result >= len(nums): results.append(-1) return results # Solution 2 class Solution(object): def nextGreaterElement(self, findNums, nums): """ :type findNums: List[int] :type nums: List[int] :rtype: List[int] """ result_hash = {} stack = [] for num in nums: while stack and num > stack[-1]: result_hash[stack.pop()] = num stack.append(num) return [result_hash.get(x, -1) for x in findNums]
Add another solution for 496 next greater element
Add another solution for 496 next greater element
Python
mit
ruichao-factual/leetcode
class Solution(object): def nextGreaterElement(self, findNums, nums): """ :type findNums: List[int] :type nums: List[int] :rtype: List[int] """ results = [] for findNum in findNums: index = nums.index(findNum) result = index + 1 for candidate in nums[index + 1:]: if candidate > findNum: results.append(candidate) break else: result += 1 if result >= len(nums): results.append(-1) return results + + # Solution 2 + + class Solution(object): + def nextGreaterElement(self, findNums, nums): + """ + :type findNums: List[int] + :type nums: List[int] + :rtype: List[int] + """ + result_hash = {} + stack = [] + + for num in nums: + while stack and num > stack[-1]: + result_hash[stack.pop()] = num + stack.append(num) + + return [result_hash.get(x, -1) for x in findNums] +
Add another solution for 496 next greater element
## Code Before: class Solution(object): def nextGreaterElement(self, findNums, nums): """ :type findNums: List[int] :type nums: List[int] :rtype: List[int] """ results = [] for findNum in findNums: index = nums.index(findNum) result = index + 1 for candidate in nums[index + 1:]: if candidate > findNum: results.append(candidate) break else: result += 1 if result >= len(nums): results.append(-1) return results ## Instruction: Add another solution for 496 next greater element ## Code After: class Solution(object): def nextGreaterElement(self, findNums, nums): """ :type findNums: List[int] :type nums: List[int] :rtype: List[int] """ results = [] for findNum in findNums: index = nums.index(findNum) result = index + 1 for candidate in nums[index + 1:]: if candidate > findNum: results.append(candidate) break else: result += 1 if result >= len(nums): results.append(-1) return results # Solution 2 class Solution(object): def nextGreaterElement(self, findNums, nums): """ :type findNums: List[int] :type nums: List[int] :rtype: List[int] """ result_hash = {} stack = [] for num in nums: while stack and num > stack[-1]: result_hash[stack.pop()] = num stack.append(num) return [result_hash.get(x, -1) for x in findNums]
... return results # Solution 2 class Solution(object): def nextGreaterElement(self, findNums, nums): """ :type findNums: List[int] :type nums: List[int] :rtype: List[int] """ result_hash = {} stack = [] for num in nums: while stack and num > stack[-1]: result_hash[stack.pop()] = num stack.append(num) return [result_hash.get(x, -1) for x in findNums] ...
a633fd37a4d795e7b565254ef10aaa0f2ad77f31
vcontrol/rest/machines/shutdown.py
vcontrol/rest/machines/shutdown.py
from ..helpers import get_allowed import subprocess import web class ShutdownMachineR: """ This endpoint is for shutting down a running machine. """ allow_origin, rest_url = get_allowed.get_allowed() def GET(self, machine): web.header('Access-Control-Allow-Origin', self.allow_origin) try: out = subprocess.check_output("/usr/local/bin/docker-machine stop "+machine, shell=True) except: out = "unable to stop machine" return str(out)
from ..helpers import get_allowed import subprocess import web class ShutdownMachineR: """ This endpoint is for shutting down a running machine. """ allow_origin, rest_url = get_allowed.get_allowed() def GET(self, machine): try: web.header('Access-Control-Allow-Origin', self.allow_origin) except Exception as e: # no pragma print(e.message) try: out = subprocess.check_output("/usr/local/bin/docker-machine stop "+machine, shell=True) except: out = "unable to stop machine" return str(out)
Put the web.header function in a try/except block
Put the web.header function in a try/except block
Python
apache-2.0
cglewis/vcontrol,CyberReboot/vcontrol,CyberReboot/vcontrol,cglewis/vcontrol,CyberReboot/vcontrol,cglewis/vcontrol
from ..helpers import get_allowed import subprocess import web + class ShutdownMachineR: """ This endpoint is for shutting down a running machine. """ allow_origin, rest_url = get_allowed.get_allowed() + def GET(self, machine): + try: - web.header('Access-Control-Allow-Origin', self.allow_origin) + web.header('Access-Control-Allow-Origin', self.allow_origin) + except Exception as e: # no pragma + print(e.message) try: out = subprocess.check_output("/usr/local/bin/docker-machine stop "+machine, shell=True) except: out = "unable to stop machine" return str(out)
Put the web.header function in a try/except block
## Code Before: from ..helpers import get_allowed import subprocess import web class ShutdownMachineR: """ This endpoint is for shutting down a running machine. """ allow_origin, rest_url = get_allowed.get_allowed() def GET(self, machine): web.header('Access-Control-Allow-Origin', self.allow_origin) try: out = subprocess.check_output("/usr/local/bin/docker-machine stop "+machine, shell=True) except: out = "unable to stop machine" return str(out) ## Instruction: Put the web.header function in a try/except block ## Code After: from ..helpers import get_allowed import subprocess import web class ShutdownMachineR: """ This endpoint is for shutting down a running machine. """ allow_origin, rest_url = get_allowed.get_allowed() def GET(self, machine): try: web.header('Access-Control-Allow-Origin', self.allow_origin) except Exception as e: # no pragma print(e.message) try: out = subprocess.check_output("/usr/local/bin/docker-machine stop "+machine, shell=True) except: out = "unable to stop machine" return str(out)
... import subprocess import web class ShutdownMachineR: ... """ allow_origin, rest_url = get_allowed.get_allowed() def GET(self, machine): try: web.header('Access-Control-Allow-Origin', self.allow_origin) except Exception as e: # no pragma print(e.message) try: out = subprocess.check_output("/usr/local/bin/docker-machine stop "+machine, shell=True) ...
34f8c0a4a0a9f78c124cd07b121ce5b2fbf00136
onadata/libs/utils/csv_import.py
onadata/libs/utils/csv_import.py
import unicodecsv as ucsv from cStringIO import StringIO from ondata.apps.api.viewsets.xform_submission_api import dict_lists2strings from onadata.libs.utils.logger_tools import dict2xform, safe_create_instance def submit_csv(username, request, csv_data): if isinstance(csv_data, (str, unicode)): csv_data = StringIO(csv_data) elif not isinstance(csv_data, file): raise TypeError(u'Invalid param type for `csv_data`. ' 'Expected file, String or Unicode but ' 'got {} instead.'.format(type(csv_data).__name__)) csv_reader = ucsv.DictReader(csv_data) for row in csv_reader: xml_file = StringIO(dict2xform(dict_lists2strings(row), row.get('_uuid'))) safe_create_instance(username, xml_file, [], None, None)
import unicodecsv as ucsv from cStringIO import StringIO from ondata.apps.api.viewsets.xform_submission_api import dict_lists2strings from onadata.libs.utils.logger_tools import dict2xform, safe_create_instance from django.db import transaction class CSVImportException(Exception): pass def submit_csv(username, request, csv_data): if isinstance(csv_data, (str, unicode)): csv_data = StringIO(csv_data) elif not isinstance(csv_data, file): raise TypeError(u'Invalid param type for `csv_data`. ' 'Expected file, String or Unicode but ' 'got {} instead.'.format(type(csv_data).__name__)) csv_reader = ucsv.DictReader(csv_data) with transaction.atomic(): for row in csv_reader: # fetch submission uuid before nuking row metadata _uuid = row.get('_uuid') # nuke metadata (keys starting with '_') for key in row.keys(): if key.startswith('_'): del row[key] xml_file = StringIO(dict2xform(dict_lists2strings(row), _uuid)) error, instance = safe_create_instance( username, xml_file, [], None, None) if error is None: raise CSVImportException(error)
Implement atomicity for CSV imports
JZ: Implement atomicity for CSV imports CSV imports should happen for all rows or nothing at all! Use `django.transactions` for rollbacks on submission on errors Also remove metadata from CSV rows before submitting
Python
bsd-2-clause
awemulya/fieldsight-kobocat,mainakibui/kobocat,qlands/onadata,smn/onadata,sounay/flaminggo-test,piqoni/onadata,qlands/onadata,jomolinare/kobocat,sounay/flaminggo-test,mainakibui/kobocat,kobotoolbox/kobocat,kobotoolbox/kobocat,smn/onadata,piqoni/onadata,hnjamba/onaclone,awemulya/fieldsight-kobocat,smn/onadata,jomolinare/kobocat,jomolinare/kobocat,smn/onadata,spatialdev/onadata,qlands/onadata,GeoODK/onadata,kobotoolbox/kobocat,spatialdev/onadata,hnjamba/onaclone,GeoODK/onadata,spatialdev/onadata,awemulya/fieldsight-kobocat,sounay/flaminggo-test,awemulya/fieldsight-kobocat,sounay/flaminggo-test,kobotoolbox/kobocat,piqoni/onadata,GeoODK/onadata,hnjamba/onaclone,qlands/onadata,hnjamba/onaclone,GeoODK/onadata,piqoni/onadata,spatialdev/onadata,mainakibui/kobocat,mainakibui/kobocat,jomolinare/kobocat
import unicodecsv as ucsv from cStringIO import StringIO from ondata.apps.api.viewsets.xform_submission_api import dict_lists2strings from onadata.libs.utils.logger_tools import dict2xform, safe_create_instance + from django.db import transaction + + + class CSVImportException(Exception): + pass def submit_csv(username, request, csv_data): if isinstance(csv_data, (str, unicode)): csv_data = StringIO(csv_data) elif not isinstance(csv_data, file): raise TypeError(u'Invalid param type for `csv_data`. ' 'Expected file, String or Unicode but ' 'got {} instead.'.format(type(csv_data).__name__)) csv_reader = ucsv.DictReader(csv_data) + with transaction.atomic(): - for row in csv_reader: + for row in csv_reader: + # fetch submission uuid before nuking row metadata + _uuid = row.get('_uuid') + # nuke metadata (keys starting with '_') + for key in row.keys(): + if key.startswith('_'): + del row[key] - xml_file = StringIO(dict2xform(dict_lists2strings(row), + xml_file = StringIO(dict2xform(dict_lists2strings(row), _uuid)) - row.get('_uuid'))) - safe_create_instance(username, xml_file, [], None, None) + error, instance = safe_create_instance( + username, xml_file, [], None, None) + if error is None: + raise CSVImportException(error)
Implement atomicity for CSV imports
## Code Before: import unicodecsv as ucsv from cStringIO import StringIO from ondata.apps.api.viewsets.xform_submission_api import dict_lists2strings from onadata.libs.utils.logger_tools import dict2xform, safe_create_instance def submit_csv(username, request, csv_data): if isinstance(csv_data, (str, unicode)): csv_data = StringIO(csv_data) elif not isinstance(csv_data, file): raise TypeError(u'Invalid param type for `csv_data`. ' 'Expected file, String or Unicode but ' 'got {} instead.'.format(type(csv_data).__name__)) csv_reader = ucsv.DictReader(csv_data) for row in csv_reader: xml_file = StringIO(dict2xform(dict_lists2strings(row), row.get('_uuid'))) safe_create_instance(username, xml_file, [], None, None) ## Instruction: Implement atomicity for CSV imports ## Code After: import unicodecsv as ucsv from cStringIO import StringIO from ondata.apps.api.viewsets.xform_submission_api import dict_lists2strings from onadata.libs.utils.logger_tools import dict2xform, safe_create_instance from django.db import transaction class CSVImportException(Exception): pass def submit_csv(username, request, csv_data): if isinstance(csv_data, (str, unicode)): csv_data = StringIO(csv_data) elif not isinstance(csv_data, file): raise TypeError(u'Invalid param type for `csv_data`. ' 'Expected file, String or Unicode but ' 'got {} instead.'.format(type(csv_data).__name__)) csv_reader = ucsv.DictReader(csv_data) with transaction.atomic(): for row in csv_reader: # fetch submission uuid before nuking row metadata _uuid = row.get('_uuid') # nuke metadata (keys starting with '_') for key in row.keys(): if key.startswith('_'): del row[key] xml_file = StringIO(dict2xform(dict_lists2strings(row), _uuid)) error, instance = safe_create_instance( username, xml_file, [], None, None) if error is None: raise CSVImportException(error)
# ... existing code ... from ondata.apps.api.viewsets.xform_submission_api import dict_lists2strings from onadata.libs.utils.logger_tools import dict2xform, safe_create_instance from django.db import transaction class CSVImportException(Exception): pass # ... modified code ... csv_reader = ucsv.DictReader(csv_data) with transaction.atomic(): for row in csv_reader: # fetch submission uuid before nuking row metadata _uuid = row.get('_uuid') # nuke metadata (keys starting with '_') for key in row.keys(): if key.startswith('_'): del row[key] xml_file = StringIO(dict2xform(dict_lists2strings(row), _uuid)) error, instance = safe_create_instance( username, xml_file, [], None, None) if error is None: raise CSVImportException(error) # ... rest of the code ...
a27b03a89af6442dc8e1be3d310a8fc046a98ed4
foampy/tests.py
foampy/tests.py
from .core import * from .dictionaries import * from .types import * from .foil import *
"""Tests for foamPy.""" from .core import * from .dictionaries import * from .types import * from .foil import * def test_load_all_torque_drag(): """Test the `load_all_torque_drag` function.""" t, torque, drag = load_all_torque_drag(casedir="test") assert t.max() == 4.0
Add test for loading all torque and drag data
Add test for loading all torque and drag data
Python
mit
petebachant/foamPy,petebachant/foamPy,petebachant/foamPy
+ """Tests for foamPy.""" from .core import * from .dictionaries import * from .types import * from .foil import * + + def test_load_all_torque_drag(): + """Test the `load_all_torque_drag` function.""" + t, torque, drag = load_all_torque_drag(casedir="test") + assert t.max() == 4.0 +
Add test for loading all torque and drag data
## Code Before: from .core import * from .dictionaries import * from .types import * from .foil import * ## Instruction: Add test for loading all torque and drag data ## Code After: """Tests for foamPy.""" from .core import * from .dictionaries import * from .types import * from .foil import * def test_load_all_torque_drag(): """Test the `load_all_torque_drag` function.""" t, torque, drag = load_all_torque_drag(casedir="test") assert t.max() == 4.0
# ... existing code ... """Tests for foamPy.""" from .core import * # ... modified code ... from .types import * from .foil import * def test_load_all_torque_drag(): """Test the `load_all_torque_drag` function.""" t, torque, drag = load_all_torque_drag(casedir="test") assert t.max() == 4.0 # ... rest of the code ...
c755934a9bc9f15f1e7dcf6d337c3dd3acf4e824
checks/check_solarize.py
checks/check_solarize.py
import imgaug as ia import imgaug.augmenters as iaa def main(): image = ia.quokka_square((128, 128)) images_aug = iaa.Solarize(1.0)(images=[image] * (5*5)) ia.imshow(ia.draw_grid(images_aug)) if __name__ == "__main__": main()
from __future__ import print_function, division, absolute_import import imgaug as ia import imgaug.augmenters as iaa import timeit def main(): for size in [64, 128, 256, 512, 1024]: for threshold in [64, 128, 192]: time_iaa = timeit.timeit( "iaa.solarize(image, %d)" % (threshold,), number=1000, setup=( "import imgaug as ia; " "import imgaug.augmenters as iaa; " "image = ia.quokka_square((%d, %d))" % (size, size)) ) time_pil = timeit.timeit( "np.asarray(" "PIL.ImageOps.solarize(PIL.Image.fromarray(image), %d)" ")" % (threshold,), number=1000, setup=( "import numpy as np; " "import PIL.Image; " "import PIL.ImageOps; " "import imgaug as ia; " "image = ia.quokka_square((%d, %d))" % (size, size)) ) print("[size=%04d, thresh=%03d] iaa=%.4f pil=%.4f" % ( size, threshold, time_iaa, time_pil)) image = ia.quokka_square((128, 128)) images_aug = iaa.Solarize(1.0)(images=[image] * (5*5)) ia.imshow(ia.draw_grid(images_aug)) if __name__ == "__main__": main()
Add performance comparison with PIL
Add performance comparison with PIL
Python
mit
aleju/ImageAugmenter,aleju/imgaug,aleju/imgaug
+ from __future__ import print_function, division, absolute_import import imgaug as ia import imgaug.augmenters as iaa + import timeit def main(): + for size in [64, 128, 256, 512, 1024]: + for threshold in [64, 128, 192]: + time_iaa = timeit.timeit( + "iaa.solarize(image, %d)" % (threshold,), + number=1000, + setup=( + "import imgaug as ia; " + "import imgaug.augmenters as iaa; " + "image = ia.quokka_square((%d, %d))" % (size, size)) + ) + time_pil = timeit.timeit( + "np.asarray(" + "PIL.ImageOps.solarize(PIL.Image.fromarray(image), %d)" + ")" % (threshold,), + number=1000, + setup=( + "import numpy as np; " + "import PIL.Image; " + "import PIL.ImageOps; " + "import imgaug as ia; " + "image = ia.quokka_square((%d, %d))" % (size, size)) + ) + print("[size=%04d, thresh=%03d] iaa=%.4f pil=%.4f" % ( + size, threshold, time_iaa, time_pil)) + image = ia.quokka_square((128, 128)) images_aug = iaa.Solarize(1.0)(images=[image] * (5*5)) ia.imshow(ia.draw_grid(images_aug)) if __name__ == "__main__": main()
Add performance comparison with PIL
## Code Before: import imgaug as ia import imgaug.augmenters as iaa def main(): image = ia.quokka_square((128, 128)) images_aug = iaa.Solarize(1.0)(images=[image] * (5*5)) ia.imshow(ia.draw_grid(images_aug)) if __name__ == "__main__": main() ## Instruction: Add performance comparison with PIL ## Code After: from __future__ import print_function, division, absolute_import import imgaug as ia import imgaug.augmenters as iaa import timeit def main(): for size in [64, 128, 256, 512, 1024]: for threshold in [64, 128, 192]: time_iaa = timeit.timeit( "iaa.solarize(image, %d)" % (threshold,), number=1000, setup=( "import imgaug as ia; " "import imgaug.augmenters as iaa; " "image = ia.quokka_square((%d, %d))" % (size, size)) ) time_pil = timeit.timeit( "np.asarray(" "PIL.ImageOps.solarize(PIL.Image.fromarray(image), %d)" ")" % (threshold,), number=1000, setup=( "import numpy as np; " "import PIL.Image; " "import PIL.ImageOps; " "import imgaug as ia; " "image = ia.quokka_square((%d, %d))" % (size, size)) ) print("[size=%04d, thresh=%03d] iaa=%.4f pil=%.4f" % ( size, threshold, time_iaa, time_pil)) image = ia.quokka_square((128, 128)) images_aug = iaa.Solarize(1.0)(images=[image] * (5*5)) ia.imshow(ia.draw_grid(images_aug)) if __name__ == "__main__": main()
# ... existing code ... from __future__ import print_function, division, absolute_import import imgaug as ia import imgaug.augmenters as iaa import timeit def main(): for size in [64, 128, 256, 512, 1024]: for threshold in [64, 128, 192]: time_iaa = timeit.timeit( "iaa.solarize(image, %d)" % (threshold,), number=1000, setup=( "import imgaug as ia; " "import imgaug.augmenters as iaa; " "image = ia.quokka_square((%d, %d))" % (size, size)) ) time_pil = timeit.timeit( "np.asarray(" "PIL.ImageOps.solarize(PIL.Image.fromarray(image), %d)" ")" % (threshold,), number=1000, setup=( "import numpy as np; " "import PIL.Image; " "import PIL.ImageOps; " "import imgaug as ia; " "image = ia.quokka_square((%d, %d))" % (size, size)) ) print("[size=%04d, thresh=%03d] iaa=%.4f pil=%.4f" % ( size, threshold, time_iaa, time_pil)) image = ia.quokka_square((128, 128)) images_aug = iaa.Solarize(1.0)(images=[image] * (5*5)) # ... rest of the code ...
b5e13cb92f539545873d59553c03e1523eac1dbb
recipes/kaleido-core/run_test.py
recipes/kaleido-core/run_test.py
from subprocess import Popen, PIPE import json import platform # Remove "sys.exit" after feedstock creation when running # on linux-anvil-cos7-x86_64 image if platform.system() == "Linux": import sys sys.exit(0) if platform.system() == "Windows": ext = ".cmd" else: ext = "" p = Popen( ['kaleido' + ext, "plotly", "--disable-gpu"], stdout=PIPE, stdin=PIPE, stderr=PIPE, text=True ) stdout_data = p.communicate( input=json.dumps({"data": {"data": []}, "format": "png"}) )[0] assert "iVBORw" in stdout_data
from subprocess import Popen, PIPE import json import platform # Remove "sys.exit" after feedstock creation when running # on linux-anvil-cos7-x86_64 image if platform.system() == "Linux": import sys sys.exit(0) if platform.system() == "Windows": ext = ".cmd" else: ext = "" p = Popen( ['kaleido' + ext, "plotly", "--disable-gpu", "--no-sandbox", "--disable-breakpad"], stdout=PIPE, stdin=PIPE, stderr=PIPE, text=True ) stdout_data = p.communicate( input=json.dumps({"data": {"data": []}, "format": "png"}) )[0] assert "iVBORw" in stdout_data
Fix hanging test on MacOS
Fix hanging test on MacOS
Python
bsd-3-clause
kwilcox/staged-recipes,ocefpaf/staged-recipes,stuertz/staged-recipes,johanneskoester/staged-recipes,jochym/staged-recipes,SylvainCorlay/staged-recipes,goanpeca/staged-recipes,igortg/staged-recipes,mariusvniekerk/staged-recipes,conda-forge/staged-recipes,patricksnape/staged-recipes,goanpeca/staged-recipes,johanneskoester/staged-recipes,scopatz/staged-recipes,ocefpaf/staged-recipes,ReimarBauer/staged-recipes,scopatz/staged-recipes,hadim/staged-recipes,hadim/staged-recipes,jochym/staged-recipes,patricksnape/staged-recipes,mariusvniekerk/staged-recipes,conda-forge/staged-recipes,stuertz/staged-recipes,jakirkham/staged-recipes,ReimarBauer/staged-recipes,igortg/staged-recipes,jakirkham/staged-recipes,SylvainCorlay/staged-recipes,kwilcox/staged-recipes
from subprocess import Popen, PIPE import json import platform # Remove "sys.exit" after feedstock creation when running # on linux-anvil-cos7-x86_64 image if platform.system() == "Linux": import sys sys.exit(0) if platform.system() == "Windows": ext = ".cmd" else: ext = "" p = Popen( - ['kaleido' + ext, "plotly", "--disable-gpu"], + ['kaleido' + ext, "plotly", "--disable-gpu", "--no-sandbox", "--disable-breakpad"], stdout=PIPE, stdin=PIPE, stderr=PIPE, text=True ) stdout_data = p.communicate( input=json.dumps({"data": {"data": []}, "format": "png"}) )[0] assert "iVBORw" in stdout_data
Fix hanging test on MacOS
## Code Before: from subprocess import Popen, PIPE import json import platform # Remove "sys.exit" after feedstock creation when running # on linux-anvil-cos7-x86_64 image if platform.system() == "Linux": import sys sys.exit(0) if platform.system() == "Windows": ext = ".cmd" else: ext = "" p = Popen( ['kaleido' + ext, "plotly", "--disable-gpu"], stdout=PIPE, stdin=PIPE, stderr=PIPE, text=True ) stdout_data = p.communicate( input=json.dumps({"data": {"data": []}, "format": "png"}) )[0] assert "iVBORw" in stdout_data ## Instruction: Fix hanging test on MacOS ## Code After: from subprocess import Popen, PIPE import json import platform # Remove "sys.exit" after feedstock creation when running # on linux-anvil-cos7-x86_64 image if platform.system() == "Linux": import sys sys.exit(0) if platform.system() == "Windows": ext = ".cmd" else: ext = "" p = Popen( ['kaleido' + ext, "plotly", "--disable-gpu", "--no-sandbox", "--disable-breakpad"], stdout=PIPE, stdin=PIPE, stderr=PIPE, text=True ) stdout_data = p.communicate( input=json.dumps({"data": {"data": []}, "format": "png"}) )[0] assert "iVBORw" in stdout_data
... p = Popen( ['kaleido' + ext, "plotly", "--disable-gpu", "--no-sandbox", "--disable-breakpad"], stdout=PIPE, stdin=PIPE, stderr=PIPE, text=True ...
d42b9da06d5cde89a6116d711fc6ae216256cabc
shell/view/home/IconLayout.py
shell/view/home/IconLayout.py
import random class IconLayout: def __init__(self, width, height): self._icons = [] self._width = width self._height = height def add_icon(self, icon): self._icons.append(icon) self._layout_icon(icon) def remove_icon(self, icon): self._icons.remove(icon) def _is_valid_position(self, icon, x, y): icon_size = icon.props.size border = 20 if not (border < x < self._width - icon_size - border and \ border < y < self._height - icon_size - border): return False return True def _layout_icon(self, icon): while True: x = random.random() * self._width y = random.random() * self._height if self._is_valid_position(icon, x, y): break icon.props.x = x icon.props.y = y
import random class IconLayout: def __init__(self, width, height): self._icons = [] self._width = width self._height = height def add_icon(self, icon): self._icons.append(icon) self._layout_icon(icon) def remove_icon(self, icon): self._icons.remove(icon) def _is_valid_position(self, icon, x, y): icon_size = icon.get_property('size') border = 20 if not (border < x < self._width - icon_size - border and \ border < y < self._height - icon_size - border): return False return True def _layout_icon(self, icon): while True: x = random.random() * self._width y = random.random() * self._height if self._is_valid_position(icon, x, y): break icon.set_property('x', x) icon.set_property('y', y)
Use get/set_property rather than direct accessors
Use get/set_property rather than direct accessors
Python
lgpl-2.1
Daksh/sugar-toolkit-gtk3,manuq/sugar-toolkit-gtk3,tchx84/sugar-toolkit-gtk3,quozl/sugar-toolkit-gtk3,tchx84/sugar-toolkit-gtk3,i5o/sugar-toolkit-gtk3,puneetgkaur/backup_sugar_sugartoolkit,quozl/sugar-toolkit-gtk3,godiard/sugar-toolkit-gtk3,tchx84/debian-pkg-sugar-toolkit,tchx84/debian-pkg-sugar-toolkit,ceibal-tatu/sugar-toolkit,gusDuarte/sugar-toolkit-gtk3,tchx84/debian-pkg-sugar-toolkit,tchx84/debian-pkg-sugar-toolkit-gtk3,tchx84/debian-pkg-sugar-toolkit-gtk3,ceibal-tatu/sugar-toolkit-gtk3,manuq/sugar-toolkit-gtk3,samdroid-apps/sugar-toolkit-gtk3,puneetgkaur/sugar-toolkit-gtk3,sugarlabs/sugar-toolkit-gtk3,samdroid-apps/sugar-toolkit-gtk3,godiard/sugar-toolkit-gtk3,sugarlabs/sugar-toolkit,i5o/sugar-toolkit-gtk3,Daksh/sugar-toolkit-gtk3,godiard/sugar-toolkit-gtk3,tchx84/debian-pkg-sugar-toolkit-gtk3,ceibal-tatu/sugar-toolkit-gtk3,ceibal-tatu/sugar-toolkit,gusDuarte/sugar-toolkit-gtk3,puneetgkaur/sugar-toolkit-gtk3,quozl/sugar-toolkit-gtk3,i5o/sugar-toolkit-gtk3,i5o/sugar-toolkit-gtk3,Daksh/sugar-toolkit-gtk3,sugarlabs/sugar-toolkit-gtk3,samdroid-apps/sugar-toolkit-gtk3,puneetgkaur/sugar-toolkit-gtk3,samdroid-apps/sugar-toolkit-gtk3,gusDuarte/sugar-toolkit-gtk3,gusDuarte/sugar-toolkit-gtk3,ceibal-tatu/sugar-toolkit-gtk3,sugarlabs/sugar-toolkit,sugarlabs/sugar-toolkit,sugarlabs/sugar-toolkit-gtk3,quozl/sugar-toolkit-gtk3,manuq/sugar-toolkit-gtk3,puneetgkaur/backup_sugar_sugartoolkit,puneetgkaur/backup_sugar_sugartoolkit,sugarlabs/sugar-toolkit,ceibal-tatu/sugar-toolkit,tchx84/sugar-toolkit-gtk3
import random class IconLayout: def __init__(self, width, height): self._icons = [] self._width = width self._height = height def add_icon(self, icon): self._icons.append(icon) self._layout_icon(icon) def remove_icon(self, icon): self._icons.remove(icon) def _is_valid_position(self, icon, x, y): - icon_size = icon.props.size + icon_size = icon.get_property('size') border = 20 if not (border < x < self._width - icon_size - border and \ border < y < self._height - icon_size - border): return False return True def _layout_icon(self, icon): while True: x = random.random() * self._width y = random.random() * self._height if self._is_valid_position(icon, x, y): break - icon.props.x = x - icon.props.y = y + icon.set_property('x', x) + icon.set_property('y', y)
Use get/set_property rather than direct accessors
## Code Before: import random class IconLayout: def __init__(self, width, height): self._icons = [] self._width = width self._height = height def add_icon(self, icon): self._icons.append(icon) self._layout_icon(icon) def remove_icon(self, icon): self._icons.remove(icon) def _is_valid_position(self, icon, x, y): icon_size = icon.props.size border = 20 if not (border < x < self._width - icon_size - border and \ border < y < self._height - icon_size - border): return False return True def _layout_icon(self, icon): while True: x = random.random() * self._width y = random.random() * self._height if self._is_valid_position(icon, x, y): break icon.props.x = x icon.props.y = y ## Instruction: Use get/set_property rather than direct accessors ## Code After: import random class IconLayout: def __init__(self, width, height): self._icons = [] self._width = width self._height = height def add_icon(self, icon): self._icons.append(icon) self._layout_icon(icon) def remove_icon(self, icon): self._icons.remove(icon) def _is_valid_position(self, icon, x, y): icon_size = icon.get_property('size') border = 20 if not (border < x < self._width - icon_size - border and \ border < y < self._height - icon_size - border): return False return True def _layout_icon(self, icon): while True: x = random.random() * self._width y = random.random() * self._height if self._is_valid_position(icon, x, y): break icon.set_property('x', x) icon.set_property('y', y)
... def _is_valid_position(self, icon, x, y): icon_size = icon.get_property('size') border = 20 ... break icon.set_property('x', x) icon.set_property('y', y) ...
aabf28c02a4dff593e5e4b156052adb9b81a70c7
skflow/ops/tests/test_dropout_ops.py
skflow/ops/tests/test_dropout_ops.py
import numpy as np import tensorflow as tf from skflow import ops class DropoutTest(tf.test.TestCase): def test_dropout_float(self): with self.test_session(): x = tf.placeholder(tf.float32, [5, 5]) y = ops.dropout(x, 0.5) probs = tf.get_collection(ops.DROPOUTS) self.assertEqual(len(probs), 1) if __name__ == '__main__': tf.test.main()
import numpy as np import tensorflow as tf from skflow import ops class DropoutTest(tf.test.TestCase): def test_dropout_float(self): with self.test_session() as session: x = tf.placeholder(tf.float32, [5, 5]) y = ops.dropout(x, 0.5) probs = tf.get_collection(ops.DROPOUTS) session.run(tf.initialize_all_variables()) self.assertEqual(len(probs), 1) self.assertEqual(session.run(probs[0]), 0.5) def test_dropout_tensor(self): with self.test_session(): x = tf.placeholder(tf.float32, [5, 5]) y = tf.get_variable("prob", [], initializer=tf.constant_initializer(0.5)) z = ops.dropout(x, y) probs = tf.get_collection(ops.DROPOUTS) self.assertEqual(probs, [y]) if __name__ == '__main__': tf.test.main()
Test for dropout probability be a tensor
Test for dropout probability be a tensor
Python
apache-2.0
handroissuazo/tensorflow,elingg/tensorflow,awni/tensorflow,aselle/tensorflow,theflofly/tensorflow,ishay2b/tensorflow,XueqingLin/tensorflow,DavidNorman/tensorflow,odejesush/tensorflow,taknevski/tensorflow-xsmm,AndreasMadsen/tensorflow,Kongsea/tensorflow,hfp/tensorflow-xsmm,jhaux/tensorflow,JingJunYin/tensorflow,tensorflow/tensorflow,mixturemodel-flow/tensorflow,aam-at/tensorflow,ibab/tensorflow,tensorflow/tensorflow-pywrap_saved_model,Carmezim/tensorflow,yongtang/tensorflow,ageron/tensorflow,mrry/tensorflow,HKUST-SING/tensorflow,av8ramit/tensorflow,whn09/tensorflow,tensorflow/tensorflow-pywrap_tf_optimizer,ppwwyyxx/tensorflow,anilmuthineni/tensorflow,laosiaudi/tensorflow,gnieboer/tensorflow,cxxgtxy/tensorflow,gautam1858/tensorflow,gnieboer/tensorflow,annarev/tensorflow,haeusser/tensorflow,JingJunYin/tensorflow,snnn/tensorflow,lukeiwanski/tensorflow-opencl,ibab/tensorflow,alshedivat/tensorflow,AnishShah/tensorflow,anand-c-goog/tensorflow,DCSaunders/tensorflow,seanli9jan/tensorflow,scenarios/tensorflow,xodus7/tensorflow,rabipanda/tensorflow,LUTAN/tensorflow,dancingdan/tensorflow,jart/tensorflow,alisidd/tensorflow,zycdragonball/tensorflow,ghchinoy/tensorflow,tensorflow/tensorflow-pywrap_tf_optimizer,handroissuazo/tensorflow,nolanliou/tensorflow,tensorflow/tensorflow-experimental_link_static_libraries_once,vrv/tensorflow,andrewcmyers/tensorflow,girving/tensorflow,nanditav/15712-TensorFlow,AndreasMadsen/tensorflow,LUTAN/tensorflow,ivano666/tensorflow,asadziach/tensorflow,tntnatbry/tensorflow,DavidNorman/tensorflow,dancingdan/tensorflow,paolodedios/tensorflow,sandeepdsouza93/TensorFlow-15712,dongjoon-hyun/tensorflow,thesuperzapper/tensorflow,dyoung418/tensorflow,alivecor/tensorflow,asimshankar/tensorflow,seanli9jan/tensorflow,ppwwyyxx/tensorflow,hsaputra/tensorflow,strint/tensorflow,gautam1858/tensorflow,benoitsteiner/tensorflow-opencl,wangyum/tensorflow,benoitsteiner/tensorflow-opencl,johndpope/tensorflow,jalexvig/tensorflow,freedomtan/tensorflow,TakayukiSakai/tensorflow,Xeralux/tensorflow,dendisuhubdy/tensorflow,kevin-coder/tensorflow-fork,bowang/tensorflow,renyi533/tensorflow,tensorflow/tensorflow-pywrap_tf_optimizer,alsrgv/tensorflow,benoitsteiner/tensorflow,dyoung418/tensorflow,chris-chris/tensorflow,mavenlin/tensorflow,wangyum/tensorflow,mengxn/tensorflow,horance-liu/tensorflow,aam-at/tensorflow,ZhangXinNan/tensorflow,kevin-coder/tensorflow-fork,lakshayg/tensorflow,calebfoss/tensorflow,mavenlin/tensorflow,a-doumoulakis/tensorflow,tongwang01/tensorflow,gojira/tensorflow,tongwang01/tensorflow,with-git/tensorflow,llhe/tensorflow,jhaux/tensorflow,nburn42/tensorflow,Carmezim/tensorflow,dhalleine/tensorflow,jhseu/tensorflow,aam-at/tensorflow,johndpope/tensorflow,annarev/tensorflow,adit-chandra/tensorflow,cg31/tensorflow,AnishShah/tensorflow,brchiu/tensorflow,manipopopo/tensorflow,dyoung418/tensorflow,abhitopia/tensorflow,kevin-coder/tensorflow-fork,Bismarrck/tensorflow,karllessard/tensorflow,MoamerEncsConcordiaCa/tensorflow,ZhangXinNan/tensorflow,meteorcloudy/tensorflow,frreiss/tensorflow-fred,ychfan/tensorflow,chenjun0210/tensorflow,alheinecke/tensorflow-xsmm,Mazecreator/tensorflow,nburn42/tensorflow,renyi533/tensorflow,chemelnucfin/tensorflow,AnishShah/tensorflow,eadgarchen/tensorflow,alshedivat/tensorflow,lukeiwanski/tensorflow-opencl,martinwicke/tensorflow,hsaputra/tensorflow,LUTAN/tensorflow,theflofly/tensorflow,alistairlow/tensorflow,jendap/tensorflow,mavenlin/tensorflow,alivecor/tensorflow,panmari/tensorflow,jhseu/tensorflow,Mistobaan/tensorflow,arborh/tensorflow,asimshankar/tensorflow,alsrgv/tensorflow,unsiloai/syntaxnet-ops-hack,manazhao/tf_recsys,karllessard/tensorflow,LUTAN/tensorflow,laosiaudi/tensorflow,jendap/tensorflow,aldian/tensorflow,xodus7/tensorflow,sandeepgupta2k4/tensorflow,jhseu/tensorflow,MycChiu/tensorflow,snnn/tensorflow,seanli9jan/tensorflow,strint/tensorflow,cancan101/tensorflow,aldian/tensorflow,theflofly/tensorflow,ppwwyyxx/tensorflow,brchiu/tensorflow,apark263/tensorflow,chemelnucfin/tensorflow,gibiansky/tensorflow,dyoung418/tensorflow,TakayukiSakai/tensorflow,freedomtan/tensorflow,cxxgtxy/tensorflow,ibab/tensorflow,manazhao/tf_recsys,Xeralux/tensorflow,cxxgtxy/tensorflow,tillahoffmann/tensorflow,ivano666/tensorflow,gunan/tensorflow,seanli9jan/tensorflow,code-sauce/tensorflow,ivano666/tensorflow,adamtiger/tensorflow,haeusser/tensorflow,dancingdan/tensorflow,Bulochkin/tensorflow_pack,eadgarchen/tensorflow,ibmsoe/tensorflow,calebfoss/tensorflow,caisq/tensorflow,ppwwyyxx/tensorflow,arborh/tensorflow,nikste/tensorflow,manazhao/tf_recsys,petewarden/tensorflow_makefile,jostep/tensorflow,hsaputra/tensorflow,suiyuan2009/tensorflow,davidzchen/tensorflow,tensorflow/tensorflow-pywrap_tf_optimizer,jwlawson/tensorflow,freedomtan/tensorflow,nanditav/15712-TensorFlow,ninotoshi/tensorflow,pierreg/tensorflow,whn09/tensorflow,freedomtan/tensorflow,dansbecker/skflow,alistairlow/tensorflow,brchiu/tensorflow,elingg/tensorflow,EvenStrangest/tensorflow,lukeiwanski/tensorflow-opencl,DavidNorman/tensorflow,Moriadry/tensorflow,paolodedios/tensorflow,a-doumoulakis/tensorflow,alshedivat/tensorflow,Kongsea/tensorflow,asadziach/tensorflow,chemelnucfin/tensorflow,annarev/tensorflow,dancingdan/tensorflow,MycChiu/tensorflow,davidzchen/tensorflow,kobejean/tensorflow,eadgarchen/tensorflow,gunan/tensorflow,Carmezim/tensorflow,gnieboer/tensorflow,mengxn/tensorflow,ppwwyyxx/tensorflow,xzturn/tensorflow,jwlawson/tensorflow,Mazecreator/tensorflow,alsrgv/tensorflow,lukeiwanski/tensorflow,drpngx/tensorflow,thesuperzapper/tensorflow,LUTAN/tensorflow,alisidd/tensorflow,rabipanda/tensorflow,adit-chandra/tensorflow,rabipanda/tensorflow,alheinecke/tensorflow-xsmm,alisidd/tensorflow,jbedorf/tensorflow,SnakeJenny/TensorFlow,ghchinoy/tensorflow,alsrgv/tensorflow,davidzchen/tensorflow,kevin-coder/tensorflow-fork,peterbraden/tensorflow,Xeralux/tensorflow,ravindrapanda/tensorflow,peterbraden/tensorflow,memo/tensorflow,wangyum/tensorflow,taknevski/tensorflow-xsmm,allenlavoie/tensorflow,krikru/tensorflow-opencl,MoamerEncsConcordiaCa/tensorflow,llhe/tensorflow,kevin-coder/tensorflow-fork,SnakeJenny/TensorFlow,taknevski/tensorflow-xsmm,kchodorow/tensorflow,hsaputra/tensorflow,jbedorf/tensorflow,aam-at/tensorflow,yaroslavvb/tensorflow,SnakeJenny/TensorFlow,ishay2b/tensorflow,sjperkins/tensorflow,jeffzheng1/tensorflow,lakshayg/tensorflow,tensorflow/tensorflow-pywrap_saved_model,tomasreimers/tensorflow-emscripten,tensorflow/tensorflow-experimental_link_static_libraries_once,ppwwyyxx/tensorflow,cancan101/tensorflow,chemelnucfin/tensorflow,snnn/tensorflow,tensorflow/tensorflow-experimental_link_static_libraries_once,benoitsteiner/tensorflow-opencl,ibab/tensorflow,codrut3/tensorflow,wchan/tensorflow,ychfan/tensorflow,jalexvig/tensorflow,kobejean/tensorflow,asadziach/tensorflow,tomasreimers/tensorflow-emscripten,pavelchristof/gomoku-ai,Bulochkin/tensorflow_pack,chenjun0210/tensorflow,dyoung418/tensorflow,code-sauce/tensorflow,thjashin/tensorflow,taknevski/tensorflow-xsmm,ageron/tensorflow,anand-c-goog/tensorflow,ninotoshi/tensorflow,yanchen036/tensorflow,mortada/tensorflow,tomasreimers/tensorflow-emscripten,kamcpp/tensorflow,thesuperzapper/tensorflow,tensorflow/tensorflow-pywrap_saved_model,neilhan/tensorflow,admcrae/tensorflow,drpngx/tensorflow,frreiss/tensorflow-fred,yaroslavvb/tensorflow,xzturn/tensorflow,benoitsteiner/tensorflow,adamtiger/tensorflow,gojira/tensorflow,caisq/tensorflow,lukeiwanski/tensorflow-opencl,DCSaunders/tensorflow,ppries/tensorflow,davidzchen/tensorflow,benoitsteiner/tensorflow,Xeralux/tensorflow,gojira/tensorflow,raymondxyang/tensorflow,jhseu/tensorflow,strint/tensorflow,alivecor/tensorflow,drpngx/tensorflow,JingJunYin/tensorflow,snnn/tensorflow,guschmue/tensorflow,chemelnucfin/tensorflow,tensorflow/tensorflow-pywrap_saved_model,arborh/tensorflow,gnieboer/tensorflow,haeusser/tensorflow,raymondxyang/tensorflow,anilmuthineni/tensorflow,benoitsteiner/tensorflow-xsmm,manipopopo/tensorflow,ppries/tensorflow,asimshankar/tensorflow,chenjun0210/tensorflow,pcm17/tensorflow,vrv/tensorflow,xodus7/tensorflow,nolanliou/tensorflow,code-sauce/tensorflow,eerwitt/tensorflow,Bismarrck/tensorflow,DavidNorman/tensorflow,ravindrapanda/tensorflow,Mistobaan/tensorflow,jbedorf/tensorflow,girving/tensorflow,sarvex/tensorflow,jalexvig/tensorflow,hfp/tensorflow-xsmm,petewarden/tensorflow,dancingdan/tensorflow,ivano666/tensorflow,codrut3/tensorflow,DCSaunders/tensorflow,mixturemodel-flow/tensorflow,jendap/tensorflow,DCSaunders/tensorflow,MycChiu/tensorflow,AnishShah/tensorflow,MycChiu/tensorflow,gibiansky/tensorflow,Intel-tensorflow/tensorflow,JVillella/tensorflow,markslwong/tensorflow,jwlawson/tensorflow,mrry/tensorflow,nburn42/tensorflow,admcrae/tensorflow,jbedorf/tensorflow,ravindrapanda/tensorflow,seanli9jan/tensorflow,eaplatanios/tensorflow,hsaputra/tensorflow,EvenStrangest/tensorflow,gnieboer/tensorflow,adit-chandra/tensorflow,wchan/tensorflow,pcm17/tensorflow,thesuperzapper/tensorflow,martinwicke/tensorflow,seaotterman/tensorflow,gnieboer/tensorflow,lukeiwanski/tensorflow,sandeepdsouza93/TensorFlow-15712,asimshankar/tensorflow,Intel-tensorflow/tensorflow,guschmue/tensorflow,ibab/tensorflow,jbedorf/tensorflow,chemelnucfin/tensorflow,gautam1858/tensorflow,raymondxyang/tensorflow,HKUST-SING/tensorflow,lukeiwanski/tensorflow-opencl,ppries/tensorflow,sandeepdsouza93/TensorFlow-15712,pavelchristof/gomoku-ai,ishay2b/tensorflow,whn09/tensorflow,ibmsoe/tensorflow,meteorcloudy/tensorflow,Carmezim/tensorflow,whn09/tensorflow,a-doumoulakis/tensorflow,theflofly/tensorflow,jhseu/tensorflow,ArtsiomCh/tensorflow,ppries/tensorflow,jart/tensorflow,llhe/tensorflow,dhalleine/tensorflow,adit-chandra/tensorflow,DavidNorman/tensorflow,HKUST-SING/tensorflow,rdipietro/tensorflow,ppries/tensorflow,nolanliou/tensorflow,benoitsteiner/tensorflow,ghchinoy/tensorflow,sarvex/tensorflow,seanli9jan/tensorflow,Carmezim/tensorflow,chris-chris/tensorflow,davidzchen/tensorflow,eaplatanios/tensorflow,pavelchristof/gomoku-ai,thjashin/tensorflow,mortada/tensorflow,neilhan/tensorflow,Xeralux/tensorflow,mdrumond/tensorflow,paolodedios/tensorflow,cancan101/tensorflow,jhaux/tensorflow,hehongliang/tensorflow,dendisuhubdy/tensorflow,Bismarrck/tensorflow,renyi533/tensorflow,renyi533/tensorflow,tillahoffmann/tensorflow,apark263/tensorflow,chenjun0210/tensorflow,anilmuthineni/tensorflow,Intel-Corporation/tensorflow,theflofly/tensorflow,manipopopo/tensorflow,manjunaths/tensorflow,alshedivat/tensorflow,nightjean/Deep-Learning,lakshayg/tensorflow,kobejean/tensorflow,cg31/tensorflow,handroissuazo/tensorflow,xzturn/tensorflow,tensorflow/tensorflow-experimental_link_static_libraries_once,Carmezim/tensorflow,aldian/tensorflow,panmari/tensorflow,frreiss/tensorflow-fred,odejesush/tensorflow,tillahoffmann/tensorflow,ppwwyyxx/tensorflow,Moriadry/tensorflow,xzturn/tensorflow,ravindrapanda/tensorflow,jendap/tensorflow,jart/tensorflow,paolodedios/tensorflow,manjunaths/tensorflow,yufengg/tensorflow,johndpope/tensorflow,jhaux/tensorflow,caisq/tensorflow,jalexvig/tensorflow,annarev/tensorflow,naturali/tensorflow,handroissuazo/tensorflow,mrry/tensorflow,jwlawson/tensorflow,strint/tensorflow,maciekcc/tensorflow,nolanliou/tensorflow,meteorcloudy/tensorflow,ageron/tensorflow,lakshayg/tensorflow,odejesush/tensorflow,adit-chandra/tensorflow,sarvex/tensorflow,tensorflow/tensorflow-experimental_link_static_libraries_once,yongtang/tensorflow,karllessard/tensorflow,DCSaunders/tensorflow,snnn/tensorflow,yufengg/tensorflow,Bismarrck/tensorflow,scenarios/tensorflow,chenjun0210/tensorflow,LUTAN/tensorflow,thjashin/tensorflow,tongwang01/tensorflow,awni/tensorflow,ville-k/tensorflow,Kongsea/tensorflow,code-sauce/tensorflow,ran5515/DeepDecision,nanditav/15712-TensorFlow,apark263/tensorflow,AndreasMadsen/tensorflow,seaotterman/tensorflow,tomasreimers/tensorflow-emscripten,RapidApplicationDevelopment/tensorflow,nburn42/tensorflow,ghchinoy/tensorflow,bowang/tensorflow,alshedivat/tensorflow,ArtsiomCh/tensorflow,alivecor/tensorflow,ran5515/DeepDecision,dendisuhubdy/tensorflow,TakayukiSakai/tensorflow,Moriadry/tensorflow,karllessard/tensorflow,XueqingLin/tensorflow,tensorflow/tensorflow-experimental_link_static_libraries_once,manipopopo/tensorflow,AndreasMadsen/tensorflow,unsiloai/syntaxnet-ops-hack,ninotoshi/tensorflow,MostafaGazar/tensorflow,jwlawson/tensorflow,manjunaths/tensorflow,ZhangXinNan/tensorflow,nanditav/15712-TensorFlow,tomasreimers/tensorflow-emscripten,tiagofrepereira2012/tensorflow,panmari/tensorflow,abhitopia/tensorflow,kobejean/tensorflow,DavidNorman/tensorflow,dongjoon-hyun/tensorflow,strint/tensorflow,pierreg/tensorflow,tensorflow/tensorflow,horance-liu/tensorflow,drpngx/tensorflow,dyoung418/tensorflow,tornadozou/tensorflow,nburn42/tensorflow,XueqingLin/tensorflow,drpngx/tensorflow,Mazecreator/tensorflow,Intel-Corporation/tensorflow,abhitopia/tensorflow,cxxgtxy/tensorflow,aselle/tensorflow,asimshankar/tensorflow,Bulochkin/tensorflow_pack,seaotterman/tensorflow,jhseu/tensorflow,seanli9jan/tensorflow,zasdfgbnm/tensorflow,MoamerEncsConcordiaCa/tensorflow,tntnatbry/tensorflow,girving/tensorflow,MostafaGazar/tensorflow,cancan101/tensorflow,calebfoss/tensorflow,tensorflow/tensorflow,ghchinoy/tensorflow,with-git/tensorflow,haeusser/tensorflow,maciekcc/tensorflow,krikru/tensorflow-opencl,ZhangXinNan/tensorflow,codrut3/tensorflow,ychfan/tensorflow,hehongliang/tensorflow,pierreg/tensorflow,bowang/tensorflow,girving/tensorflow,girving/tensorflow,Intel-tensorflow/tensorflow,theflofly/tensorflow,manazhao/tf_recsys,admcrae/tensorflow,chenjun0210/tensorflow,jendap/tensorflow,benoitsteiner/tensorflow-opencl,ville-k/tensorflow,chenjun0210/tensorflow,thjashin/tensorflow,lukeiwanski/tensorflow,arborh/tensorflow,admcrae/tensorflow,Mistobaan/tensorflow,cancan101/tensorflow,ageron/tensorflow,dongjoon-hyun/tensorflow,hehongliang/tensorflow,vrv/tensorflow,odejesush/tensorflow,MostafaGazar/tensorflow,alistairlow/tensorflow,eerwitt/tensorflow,eaplatanios/tensorflow,xodus7/tensorflow,kamcpp/tensorflow,alistairlow/tensorflow,dancingdan/tensorflow,mengxn/tensorflow,rdipietro/tensorflow,ibmsoe/tensorflow,codrut3/tensorflow,drpngx/tensorflow,xzturn/tensorflow,jeffzheng1/tensorflow,HaebinShin/tensorflow,bowang/tensorflow,yongtang/tensorflow,eerwitt/tensorflow,ville-k/tensorflow,lukeiwanski/tensorflow,ibab/tensorflow,Carmezim/tensorflow,kchodorow/tensorflow,laszlocsomor/tensorflow,juharris/tensorflow,Intel-Corporation/tensorflow,guschmue/tensorflow,nightjean/Deep-Learning,gojira/tensorflow,zasdfgbnm/tensorflow,karllessard/tensorflow,Xeralux/tensorflow,frreiss/tensorflow-fred,tntnatbry/tensorflow,dendisuhubdy/tensorflow,ghchinoy/tensorflow,alistairlow/tensorflow,XueqingLin/tensorflow,Intel-tensorflow/tensorflow,taknevski/tensorflow-xsmm,suiyuan2009/tensorflow,nburn42/tensorflow,ravindrapanda/tensorflow,nightjean/Deep-Learning,brchiu/tensorflow,hsaputra/tensorflow,handroissuazo/tensorflow,wchan/tensorflow,horance-liu/tensorflow,tntnatbry/tensorflow,benoitsteiner/tensorflow-opencl,jostep/tensorflow,yanchen036/tensorflow,scenarios/tensorflow,kevin-coder/tensorflow-fork,paolodedios/tensorflow,pcm17/tensorflow,annarev/tensorflow,wangyum/tensorflow,gojira/tensorflow,calebfoss/tensorflow,asimshankar/tensorflow,nikste/tensorflow,martinwicke/tensorflow,jwlawson/tensorflow,jbedorf/tensorflow,alivecor/tensorflow,llhe/tensorflow,av8ramit/tensorflow,benoitsteiner/tensorflow-opencl,XueqingLin/tensorflow,ravindrapanda/tensorflow,vrv/tensorflow,ivano666/tensorflow,mdrumond/tensorflow,pierreg/tensorflow,vrv/tensorflow,Bulochkin/tensorflow_pack,seaotterman/tensorflow,alistairlow/tensorflow,hfp/tensorflow-xsmm,martinwicke/tensorflow,raymondxyang/tensorflow,mavenlin/tensorflow,adit-chandra/tensorflow,Bulochkin/tensorflow_pack,chemelnucfin/tensorflow,calebfoss/tensorflow,alheinecke/tensorflow-xsmm,yaroslavvb/tensorflow,ghchinoy/tensorflow,neilhan/tensorflow,nikste/tensorflow,anand-c-goog/tensorflow,memo/tensorflow,krikru/tensorflow-opencl,alheinecke/tensorflow-xsmm,adamtiger/tensorflow,jalexvig/tensorflow,frreiss/tensorflow-fred,kamcpp/tensorflow,aldian/tensorflow,ville-k/tensorflow,pavelchristof/gomoku-ai,Intel-tensorflow/tensorflow,alivecor/tensorflow,allenlavoie/tensorflow,av8ramit/tensorflow,petewarden/tensorflow_makefile,memo/tensorflow,sandeepgupta2k4/tensorflow,alisidd/tensorflow,manazhao/tf_recsys,jart/tensorflow,paolodedios/tensorflow,XueqingLin/tensorflow,haeusser/tensorflow,jwlawson/tensorflow,krikru/tensorflow-opencl,jwlawson/tensorflow,av8ramit/tensorflow,chenjun0210/tensorflow,caisq/tensorflow,lukeiwanski/tensorflow-opencl,cg31/tensorflow,ninotoshi/tensorflow,ravindrapanda/tensorflow,nikste/tensorflow,mengxn/tensorflow,davidzchen/tensorflow,jeffzheng1/tensorflow,Bulochkin/tensorflow_pack,ppwwyyxx/tensorflow,eerwitt/tensorflow,zycdragonball/tensorflow,snnn/tensorflow,tensorflow/tensorflow-pywrap_tf_optimizer,dancingdan/tensorflow,raymondxyang/tensorflow,eerwitt/tensorflow,mixturemodel-flow/tensorflow,ageron/tensorflow,kevin-coder/tensorflow-fork,yufengg/tensorflow,zasdfgbnm/tensorflow,manipopopo/tensorflow,ppwwyyxx/tensorflow,krikru/tensorflow-opencl,allenlavoie/tensorflow,HaebinShin/tensorflow,anilmuthineni/tensorflow,ageron/tensorflow,jart/tensorflow,dhalleine/tensorflow,petewarden/tensorflow_makefile,laszlocsomor/tensorflow,tensorflow/skflow,JVillella/tensorflow,davidzchen/tensorflow,odejesush/tensorflow,lukeiwanski/tensorflow-opencl,kamcpp/tensorflow,jalexvig/tensorflow,anand-c-goog/tensorflow,eadgarchen/tensorflow,eaplatanios/tensorflow,manipopopo/tensorflow,nburn42/tensorflow,jwlawson/tensorflow,manjunaths/tensorflow,ychfan/tensorflow,AnishShah/tensorflow,juharris/tensorflow,jeffzheng1/tensorflow,bowang/tensorflow,Mistobaan/tensorflow,Xeralux/tensorflow,neilhan/tensorflow,laszlocsomor/tensorflow,nanditav/15712-TensorFlow,hfp/tensorflow-xsmm,tillahoffmann/tensorflow,HaebinShin/tensorflow,naturali/tensorflow,markslwong/tensorflow,renyi533/tensorflow,nightjean/Deep-Learning,sandeepdsouza93/TensorFlow-15712,odejesush/tensorflow,HaebinShin/tensorflow,ZhangXinNan/tensorflow,haeusser/tensorflow,suiyuan2009/tensorflow,allenlavoie/tensorflow,codrut3/tensorflow,nikste/tensorflow,gojira/tensorflow,JVillella/tensorflow,paolodedios/tensorflow,sandeepdsouza93/TensorFlow-15712,nburn42/tensorflow,SnakeJenny/TensorFlow,gautam1858/tensorflow,sarvex/tensorflow,TakayukiSakai/tensorflow,yongtang/tensorflow,apark263/tensorflow,petewarden/tensorflow_makefile,av8ramit/tensorflow,admcrae/tensorflow,tomasreimers/tensorflow-emscripten,martinwicke/tensorflow,gautam1858/tensorflow,JingJunYin/tensorflow,aam-at/tensorflow,strint/tensorflow,dendisuhubdy/tensorflow,tensorflow/tensorflow,tensorflow/tensorflow-experimental_link_static_libraries_once,benoitsteiner/tensorflow-xsmm,annarev/tensorflow,tillahoffmann/tensorflow,Moriadry/tensorflow,dongjoon-hyun/tensorflow,kchodorow/tensorflow,EvenStrangest/tensorflow,mrry/tensorflow,yufengg/tensorflow,pierreg/tensorflow,anilmuthineni/tensorflow,maciekcc/tensorflow,renyi533/tensorflow,vrv/tensorflow,maciekcc/tensorflow,Carmezim/tensorflow,anilmuthineni/tensorflow,tensorflow/tensorflow-pywrap_tf_optimizer,elingg/tensorflow,sjperkins/tensorflow,lukeiwanski/tensorflow-opencl,laosiaudi/tensorflow,cg31/tensorflow,ArtsiomCh/tensorflow,asadziach/tensorflow,alistairlow/tensorflow,ghchinoy/tensorflow,lukeiwanski/tensorflow,with-git/tensorflow,EvenStrangest/tensorflow,kobejean/tensorflow,sandeepdsouza93/TensorFlow-15712,pavelchristof/gomoku-ai,vrv/tensorflow,ravindrapanda/tensorflow,aam-at/tensorflow,jalexvig/tensorflow,tongwang01/tensorflow,Bismarrck/tensorflow,jhseu/tensorflow,Intel-tensorflow/tensorflow,seaotterman/tensorflow,meteorcloudy/tensorflow,ishay2b/tensorflow,jendap/tensorflow,tongwang01/tensorflow,apark263/tensorflow,EvenStrangest/tensorflow,horance-liu/tensorflow,xzturn/tensorflow,guschmue/tensorflow,bowang/tensorflow,adit-chandra/tensorflow,jhaux/tensorflow,ageron/tensorflow,andrewcmyers/tensorflow,yanchen036/tensorflow,gunan/tensorflow,handroissuazo/tensorflow,JingJunYin/tensorflow,ibmsoe/tensorflow,Intel-Corporation/tensorflow,ravindrapanda/tensorflow,mavenlin/tensorflow,rdipietro/tensorflow,jhaux/tensorflow,mdrumond/tensorflow,tongwang01/tensorflow,petewarden/tensorflow,zycdragonball/tensorflow,ivano666/tensorflow,AnishShah/tensorflow,alsrgv/tensorflow,ppries/tensorflow,Xeralux/tensorflow,mrry/tensorflow,tensorflow/tensorflow-pywrap_saved_model,gunan/tensorflow,benoitsteiner/tensorflow-opencl,kobejean/tensorflow,Moriadry/tensorflow,gojira/tensorflow,JVillella/tensorflow,alheinecke/tensorflow-xsmm,renyi533/tensorflow,xzturn/tensorflow,tiagofrepereira2012/tensorflow,tongwang01/tensorflow,karllessard/tensorflow,xodus7/tensorflow,nanditav/15712-TensorFlow,freedomtan/tensorflow,ZhangXinNan/tensorflow,rabipanda/tensorflow,apark263/tensorflow,gnieboer/tensorflow,RapidApplicationDevelopment/tensorflow,Mistobaan/tensorflow,meteorcloudy/tensorflow,Moriadry/tensorflow,zasdfgbnm/tensorflow,ageron/tensorflow,kobejean/tensorflow,chris-chris/tensorflow,tillahoffmann/tensorflow,yaroslavvb/tensorflow,ran5515/DeepDecision,alheinecke/tensorflow-xsmm,mrry/tensorflow,laszlocsomor/tensorflow,apark263/tensorflow,nikste/tensorflow,sarvex/tensorflow,jbedorf/tensorflow,XueqingLin/tensorflow,markslwong/tensorflow,alheinecke/tensorflow-xsmm,anand-c-goog/tensorflow,allenlavoie/tensorflow,johndpope/tensorflow,karllessard/tensorflow,cg31/tensorflow,bowang/tensorflow,DCSaunders/tensorflow,ran5515/DeepDecision,horance-liu/tensorflow,MoamerEncsConcordiaCa/tensorflow,lukeiwanski/tensorflow,seanli9jan/tensorflow,ibab/tensorflow,anand-c-goog/tensorflow,neilhan/tensorflow,hfp/tensorflow-xsmm,DavidNorman/tensorflow,mengxn/tensorflow,cancan101/tensorflow,kamcpp/tensorflow,pavelchristof/gomoku-ai,renyi533/tensorflow,lukeiwanski/tensorflow-opencl,andrewcmyers/tensorflow,with-git/tensorflow,JVillella/tensorflow,mengxn/tensorflow,admcrae/tensorflow,paolodedios/tensorflow,xodus7/tensorflow,peterbraden/tensorflow,freedomtan/tensorflow,paolodedios/tensorflow,freedomtan/tensorflow,jeffzheng1/tensorflow,snnn/tensorflow,cg31/tensorflow,jhseu/tensorflow,nolanliou/tensorflow,manipopopo/tensorflow,ibmsoe/tensorflow,DavidNorman/tensorflow,tensorflow/tensorflow-pywrap_saved_model,nightjean/Deep-Learning,brchiu/tensorflow,gunan/tensorflow,lakshayg/tensorflow,eerwitt/tensorflow,xodus7/tensorflow,adit-chandra/tensorflow,zasdfgbnm/tensorflow,allenlavoie/tensorflow,ychfan/tensorflow,dancingdan/tensorflow,frreiss/tensorflow-fred,jbedorf/tensorflow,tornadozou/tensorflow,gojira/tensorflow,jbedorf/tensorflow,naturali/tensorflow,caisq/tensorflow,ninotoshi/tensorflow,lakshayg/tensorflow,hfp/tensorflow-xsmm,nolanliou/tensorflow,laosiaudi/tensorflow,caisq/tensorflow,suiyuan2009/tensorflow,jendap/tensorflow,LUTAN/tensorflow,jbedorf/tensorflow,memo/tensorflow,xzturn/tensorflow,Intel-tensorflow/tensorflow,gibiansky/tensorflow,kchodorow/tensorflow,laszlocsomor/tensorflow,with-git/tensorflow,tornadozou/tensorflow,rabipanda/tensorflow,memo/tensorflow,alsrgv/tensorflow,tntnatbry/tensorflow,sjperkins/tensorflow,whn09/tensorflow,neilhan/tensorflow,gautam1858/tensorflow,Bismarrck/tensorflow,mengxn/tensorflow,pierreg/tensorflow,kchodorow/tensorflow,with-git/tensorflow,llhe/tensorflow,wchan/tensorflow,ZhangXinNan/tensorflow,jostep/tensorflow,calebfoss/tensorflow,davidzchen/tensorflow,yanchen036/tensorflow,guschmue/tensorflow,anand-c-goog/tensorflow,meteorcloudy/tensorflow,awni/tensorflow,ville-k/tensorflow,jeffzheng1/tensorflow,jostep/tensorflow,eadgarchen/tensorflow,mengxn/tensorflow,Bismarrck/tensorflow,yongtang/tensorflow,nightjean/Deep-Learning,zasdfgbnm/tensorflow,ville-k/tensorflow,yanchen036/tensorflow,tomasreimers/tensorflow-emscripten,unsiloai/syntaxnet-ops-hack,manjunaths/tensorflow,admcrae/tensorflow,memo/tensorflow,dhalleine/tensorflow,brchiu/tensorflow,tntnatbry/tensorflow,krikru/tensorflow-opencl,benoitsteiner/tensorflow-xsmm,abhitopia/tensorflow,asimshankar/tensorflow,nanditav/15712-TensorFlow,jostep/tensorflow,chemelnucfin/tensorflow,kevin-coder/tensorflow-fork,XueqingLin/tensorflow,naturali/tensorflow,theflofly/tensorflow,Mistobaan/tensorflow,ran5515/DeepDecision,alshedivat/tensorflow,adamtiger/tensorflow,ageron/tensorflow,tntnatbry/tensorflow,eadgarchen/tensorflow,Mazecreator/tensorflow,mavenlin/tensorflow,seanli9jan/tensorflow,abhitopia/tensorflow,tensorflow/tensorflow-pywrap_saved_model,Xeralux/tensorflow,whn09/tensorflow,manipopopo/tensorflow,HKUST-SING/tensorflow,cxxgtxy/tensorflow,MostafaGazar/tensorflow,ychfan/tensorflow,martinwicke/tensorflow,elingg/tensorflow,johndpope/tensorflow,cancan101/tensorflow,abhitopia/tensorflow,girving/tensorflow,AnishShah/tensorflow,mortada/tensorflow,Bismarrck/tensorflow,alisidd/tensorflow,manazhao/tf_recsys,seaotterman/tensorflow,renyi533/tensorflow,asadziach/tensorflow,a-doumoulakis/tensorflow,mortada/tensorflow,gautam1858/tensorflow,tiagofrepereira2012/tensorflow,alsrgv/tensorflow,seanli9jan/tensorflow,AnishShah/tensorflow,snnn/tensorflow,codrut3/tensorflow,a-doumoulakis/tensorflow,rdipietro/tensorflow,tomasreimers/tensorflow-emscripten,benoitsteiner/tensorflow-xsmm,jhaux/tensorflow,elingg/tensorflow,laosiaudi/tensorflow,aselle/tensorflow,gunan/tensorflow,krikru/tensorflow-opencl,gnieboer/tensorflow,xodus7/tensorflow,snnn/tensorflow,petewarden/tensorflow,sjperkins/tensorflow,petewarden/tensorflow,naturali/tensorflow,hfp/tensorflow-xsmm,abhitopia/tensorflow,DavidNorman/tensorflow,llhe/tensorflow,annarev/tensorflow,yongtang/tensorflow,laosiaudi/tensorflow,llhe/tensorflow,anand-c-goog/tensorflow,wchan/tensorflow,hsaputra/tensorflow,yongtang/tensorflow,hfp/tensorflow-xsmm,mdrumond/tensorflow,chenjun0210/tensorflow,TakayukiSakai/tensorflow,whn09/tensorflow,seaotterman/tensorflow,gibiansky/tensorflow,hsaputra/tensorflow,Mistobaan/tensorflow,allenlavoie/tensorflow,hfp/tensorflow-xsmm,adit-chandra/tensorflow,maciekcc/tensorflow,eaplatanios/tensorflow,neilhan/tensorflow,DavidNorman/tensorflow,peterbraden/tensorflow,mrry/tensorflow,gunan/tensorflow,xzturn/tensorflow,JVillella/tensorflow,AnishShah/tensorflow,jart/tensorflow,scenarios/tensorflow,taknevski/tensorflow-xsmm,RapidApplicationDevelopment/tensorflow,mixturemodel-flow/tensorflow,taknevski/tensorflow-xsmm,lukeiwanski/tensorflow,a-doumoulakis/tensorflow,jostep/tensorflow,panmari/tensorflow,nikste/tensorflow,tensorflow/tensorflow-pywrap_saved_model,ychfan/tensorflow,ZhangXinNan/tensorflow,cancan101/tensorflow,thesuperzapper/tensorflow,awni/tensorflow,awni/tensorflow,ran5515/DeepDecision,aldian/tensorflow,aam-at/tensorflow,alsrgv/tensorflow,maciekcc/tensorflow,lakshayg/tensorflow,Moriadry/tensorflow,kchodorow/tensorflow,ghchinoy/tensorflow,benoitsteiner/tensorflow-xsmm,frreiss/tensorflow-fred,dongjoon-hyun/tensorflow,karllessard/tensorflow,handroissuazo/tensorflow,mortada/tensorflow,apark263/tensorflow,ppries/tensorflow,hfp/tensorflow-xsmm,arborh/tensorflow,Mistobaan/tensorflow,nightjean/Deep-Learning,ninotoshi/tensorflow,frreiss/tensorflow-fred,ibmsoe/tensorflow,kevin-coder/tensorflow-fork,codrut3/tensorflow,raymondxyang/tensorflow,Carmezim/tensorflow,eaplatanios/tensorflow,tensorflow/tensorflow-experimental_link_static_libraries_once,gunan/tensorflow,codrut3/tensorflow,raymondxyang/tensorflow,wangyum/tensorflow,rabipanda/tensorflow,av8ramit/tensorflow,chemelnucfin/tensorflow,ZhangXinNan/tensorflow,guschmue/tensorflow,tiagofrepereira2012/tensorflow,annarev/tensorflow,whn09/tensorflow,MoamerEncsConcordiaCa/tensorflow,ppries/tensorflow,jwlawson/tensorflow,benoitsteiner/tensorflow,manjunaths/tensorflow,Intel-tensorflow/tensorflow,eerwitt/tensorflow,eerwitt/tensorflow,RapidApplicationDevelopment/tensorflow,ishay2b/tensorflow,chris-chris/tensorflow,andrewcmyers/tensorflow,ychfan/tensorflow,sandeepdsouza93/TensorFlow-15712,laosiaudi/tensorflow,zasdfgbnm/tensorflow,ivano666/tensorflow,jalexvig/tensorflow,sandeepdsouza93/TensorFlow-15712,girving/tensorflow,dancingdan/tensorflow,karllessard/tensorflow,arborh/tensorflow,benoitsteiner/tensorflow-opencl,HKUST-SING/tensorflow,sarvex/tensorflow,jbedorf/tensorflow,gautam1858/tensorflow,handroissuazo/tensorflow,jostep/tensorflow,code-sauce/tensorflow,johndpope/tensorflow,alshedivat/tensorflow,ville-k/tensorflow,jendap/tensorflow,sjperkins/tensorflow,dendisuhubdy/tensorflow,drpngx/tensorflow,unsiloai/syntaxnet-ops-hack,laszlocsomor/tensorflow,chris-chris/tensorflow,zasdfgbnm/tensorflow,Intel-Corporation/tensorflow,juharris/tensorflow,yufengg/tensorflow,cxxgtxy/tensorflow,pcm17/tensorflow,yufengg/tensorflow,Intel-tensorflow/tensorflow,Mistobaan/tensorflow,yufengg/tensorflow,jeffzheng1/tensorflow,nolanliou/tensorflow,tornadozou/tensorflow,tiagofrepereira2012/tensorflow,XueqingLin/tensorflow,zasdfgbnm/tensorflow,ville-k/tensorflow,frreiss/tensorflow-fred,yongtang/tensorflow,jalexvig/tensorflow,aselle/tensorflow,chris-chris/tensorflow,hfp/tensorflow-xsmm,MoamerEncsConcordiaCa/tensorflow,tensorflow/tensorflow-pywrap_saved_model,mengxn/tensorflow,apark263/tensorflow,awni/tensorflow,ninotoshi/tensorflow,AnishShah/tensorflow,suiyuan2009/tensorflow,benoitsteiner/tensorflow-xsmm,gojira/tensorflow,yaroslavvb/tensorflow,adit-chandra/tensorflow,xzturn/tensorflow,whn09/tensorflow,aldian/tensorflow,brchiu/tensorflow,mortada/tensorflow,manjunaths/tensorflow,guschmue/tensorflow,gibiansky/tensorflow,theflofly/tensorflow,Mazecreator/tensorflow,Mazecreator/tensorflow,seaotterman/tensorflow,eadgarchen/tensorflow,sjperkins/tensorflow,manazhao/tf_recsys,ppwwyyxx/tensorflow,meteorcloudy/tensorflow,ArtsiomCh/tensorflow,thesuperzapper/tensorflow,martinwicke/tensorflow,mdrumond/tensorflow,panmari/tensorflow,nburn42/tensorflow,chris-chris/tensorflow,ZhangXinNan/tensorflow,JingJunYin/tensorflow,elingg/tensorflow,a-doumoulakis/tensorflow,manipopopo/tensorflow,arborh/tensorflow,rdipietro/tensorflow,dongjoon-hyun/tensorflow,naturali/tensorflow,SnakeJenny/TensorFlow,ghchinoy/tensorflow,asimshankar/tensorflow,benoitsteiner/tensorflow,kamcpp/tensorflow,thjashin/tensorflow,kevin-coder/tensorflow-fork,AndreasMadsen/tensorflow,petewarden/tensorflow_makefile,unsiloai/syntaxnet-ops-hack,Bismarrck/tensorflow,aselle/tensorflow,jhaux/tensorflow,petewarden/tensorflow,freedomtan/tensorflow,anilmuthineni/tensorflow,pcm17/tensorflow,tensorflow/tensorflow-pywrap_tf_optimizer,horance-liu/tensorflow,MostafaGazar/tensorflow,JingJunYin/tensorflow,nikste/tensorflow,aam-at/tensorflow,MoamerEncsConcordiaCa/tensorflow,hehongliang/tensorflow,davidzchen/tensorflow,pcm17/tensorflow,jendap/tensorflow,awni/tensorflow,Kongsea/tensorflow,Bulochkin/tensorflow_pack,tensorflow/tensorflow,asadziach/tensorflow,pcm17/tensorflow,laosiaudi/tensorflow,johndpope/tensorflow,DavidNorman/tensorflow,Bismarrck/tensorflow,allenlavoie/tensorflow,dyoung418/tensorflow,zycdragonball/tensorflow,HKUST-SING/tensorflow,with-git/tensorflow,naturali/tensorflow,tiagofrepereira2012/tensorflow,juharris/tensorflow,benoitsteiner/tensorflow-xsmm,mixturemodel-flow/tensorflow,thesuperzapper/tensorflow,alshedivat/tensorflow,mortada/tensorflow,benoitsteiner/tensorflow-opencl,tensorflow/tensorflow,MycChiu/tensorflow,jhseu/tensorflow,markslwong/tensorflow,taknevski/tensorflow-xsmm,Bulochkin/tensorflow_pack,MostafaGazar/tensorflow,eaplatanios/tensorflow,seaotterman/tensorflow,petewarden/tensorflow,sandeepgupta2k4/tensorflow,memo/tensorflow,Kongsea/tensorflow,mixturemodel-flow/tensorflow,vrv/tensorflow,rabipanda/tensorflow,renyi533/tensorflow,alisidd/tensorflow,frreiss/tensorflow-fred,tiagofrepereira2012/tensorflow,frreiss/tensorflow-fred,guschmue/tensorflow,adamtiger/tensorflow,ArtsiomCh/tensorflow,markslwong/tensorflow,alsrgv/tensorflow,MoamerEncsConcordiaCa/tensorflow,ishay2b/tensorflow,memo/tensorflow,Xeralux/tensorflow,dendisuhubdy/tensorflow,dhalleine/tensorflow,scenarios/tensorflow,drpngx/tensorflow,yaroslavvb/tensorflow,ppwwyyxx/tensorflow,tensorflow/tensorflow-pywrap_saved_model,juharris/tensorflow,Xeralux/tensorflow,unsiloai/syntaxnet-ops-hack,guschmue/tensorflow,handroissuazo/tensorflow,TakayukiSakai/tensorflow,rabipanda/tensorflow,aselle/tensorflow,odejesush/tensorflow,vrv/tensorflow,jart/tensorflow,nolanliou/tensorflow,elingg/tensorflow,eaplatanios/tensorflow,tensorflow/tensorflow-pywrap_tf_optimizer,LUTAN/tensorflow,manjunaths/tensorflow,aselle/tensorflow,Kongsea/tensorflow,laszlocsomor/tensorflow,HaebinShin/tensorflow,code-sauce/tensorflow,codrut3/tensorflow,gautam1858/tensorflow,AndreasMadsen/tensorflow,ppries/tensorflow,cxxgtxy/tensorflow,horance-liu/tensorflow,alisidd/tensorflow,nburn42/tensorflow,xzturn/tensorflow,arborh/tensorflow,HKUST-SING/tensorflow,rdipietro/tensorflow,tensorflow/tensorflow-experimental_link_static_libraries_once,jalexvig/tensorflow,bowang/tensorflow,dancingdan/tensorflow,sandeepgupta2k4/tensorflow,benoitsteiner/tensorflow-xsmm,martinwicke/tensorflow,karllessard/tensorflow,HKUST-SING/tensorflow,EvenStrangest/tensorflow,aam-at/tensorflow,sandeepdsouza93/TensorFlow-15712,hehongliang/tensorflow,tensorflow/tensorflow,HKUST-SING/tensorflow,petewarden/tensorflow,Intel-Corporation/tensorflow,scenarios/tensorflow,Intel-Corporation/tensorflow,freedomtan/tensorflow,tensorflow/tensorflow-experimental_link_static_libraries_once,anilmuthineni/tensorflow,aam-at/tensorflow,dongjoon-hyun/tensorflow,caisq/tensorflow,cg31/tensorflow,girving/tensorflow,gojira/tensorflow,EvenStrangest/tensorflow,sjperkins/tensorflow,RapidApplicationDevelopment/tensorflow,benoitsteiner/tensorflow-xsmm,peterbraden/tensorflow,pavelchristof/gomoku-ai,paolodedios/tensorflow,pavelchristof/gomoku-ai,jalexvig/tensorflow,admcrae/tensorflow,kobejean/tensorflow,johndpope/tensorflow,mixturemodel-flow/tensorflow,dyoung418/tensorflow,a-doumoulakis/tensorflow,alsrgv/tensorflow,adit-chandra/tensorflow,chemelnucfin/tensorflow,gibiansky/tensorflow,tensorflow/tensorflow,freedomtan/tensorflow,rdipietro/tensorflow,meteorcloudy/tensorflow,laszlocsomor/tensorflow,caisq/tensorflow,ArtsiomCh/tensorflow,ghchinoy/tensorflow,allenlavoie/tensorflow,ppwwyyxx/tensorflow,dongjoon-hyun/tensorflow,andrewcmyers/tensorflow,nburn42/tensorflow,gunan/tensorflow,renyi533/tensorflow,arborh/tensorflow,Mazecreator/tensorflow,lukeiwanski/tensorflow,calebfoss/tensorflow,nolanliou/tensorflow,theflofly/tensorflow,neilhan/tensorflow,drpngx/tensorflow,suiyuan2009/tensorflow,MycChiu/tensorflow,LUTAN/tensorflow,cxxgtxy/tensorflow,gibiansky/tensorflow,odejesush/tensorflow,Kongsea/tensorflow,girving/tensorflow,juharris/tensorflow,jhaux/tensorflow,av8ramit/tensorflow,JingJunYin/tensorflow,alshedivat/tensorflow,haeusser/tensorflow,JingJunYin/tensorflow,jart/tensorflow,kchodorow/tensorflow,admcrae/tensorflow,gautam1858/tensorflow,MostafaGazar/tensorflow,petewarden/tensorflow_makefile,hehongliang/tensorflow,code-sauce/tensorflow,zasdfgbnm/tensorflow,juharris/tensorflow,gautam1858/tensorflow,adamtiger/tensorflow,chemelnucfin/tensorflow,lakshayg/tensorflow,ville-k/tensorflow,benoitsteiner/tensorflow,eaplatanios/tensorflow,sandeepgupta2k4/tensorflow,zycdragonball/tensorflow,mavenlin/tensorflow,meteorcloudy/tensorflow,krikru/tensorflow-opencl,alsrgv/tensorflow,adit-chandra/tensorflow,asadziach/tensorflow,dhalleine/tensorflow,kamcpp/tensorflow,johndpope/tensorflow,peterbraden/tensorflow,SnakeJenny/TensorFlow,peterbraden/tensorflow,eaplatanios/tensorflow,kevin-coder/tensorflow-fork,strint/tensorflow,av8ramit/tensorflow,panmari/tensorflow,chris-chris/tensorflow,meteorcloudy/tensorflow,SnakeJenny/TensorFlow,jbedorf/tensorflow,tillahoffmann/tensorflow,with-git/tensorflow,seanli9jan/tensorflow,sarvex/tensorflow,thjashin/tensorflow,ibmsoe/tensorflow,MostafaGazar/tensorflow,rabipanda/tensorflow,unsiloai/syntaxnet-ops-hack,ghchinoy/tensorflow,kobejean/tensorflow,asadziach/tensorflow,AnishShah/tensorflow,ageron/tensorflow,JingJunYin/tensorflow,ninotoshi/tensorflow,Mazecreator/tensorflow,yanchen036/tensorflow,calebfoss/tensorflow,alistairlow/tensorflow,calebfoss/tensorflow,yaroslavvb/tensorflow,panmari/tensorflow,wchan/tensorflow,sandeepgupta2k4/tensorflow,markslwong/tensorflow,manjunaths/tensorflow,davidzchen/tensorflow,Bulochkin/tensorflow_pack,yanchen036/tensorflow,theflofly/tensorflow,peterbraden/tensorflow,eaplatanios/tensorflow,gnieboer/tensorflow,alheinecke/tensorflow-xsmm,girving/tensorflow,gibiansky/tensorflow,Mazecreator/tensorflow,ArtsiomCh/tensorflow,tongwang01/tensorflow,eadgarchen/tensorflow,girving/tensorflow,sarvex/tensorflow,tiagofrepereira2012/tensorflow,jeffzheng1/tensorflow,ishay2b/tensorflow,yongtang/tensorflow,kobejean/tensorflow,frreiss/tensorflow-fred,chemelnucfin/tensorflow,ZhangXinNan/tensorflow,HaebinShin/tensorflow,drpngx/tensorflow,alisidd/tensorflow,DCSaunders/tensorflow,neilhan/tensorflow,xodus7/tensorflow,annarev/tensorflow,juharris/tensorflow,naturali/tensorflow,hsaputra/tensorflow,adamtiger/tensorflow,apark263/tensorflow,Bulochkin/tensorflow_pack,tensorflow/tensorflow-experimental_link_static_libraries_once,code-sauce/tensorflow,mdrumond/tensorflow,MostafaGazar/tensorflow,JVillella/tensorflow,ychfan/tensorflow,thesuperzapper/tensorflow,horance-liu/tensorflow,TakayukiSakai/tensorflow,pcm17/tensorflow,horance-liu/tensorflow,maciekcc/tensorflow,asimshankar/tensorflow,unsiloai/syntaxnet-ops-hack,kchodorow/tensorflow,av8ramit/tensorflow,laszlocsomor/tensorflow,markslwong/tensorflow,SnakeJenny/TensorFlow,asadziach/tensorflow,sandeepgupta2k4/tensorflow,maciekcc/tensorflow,tensorflow/tensorflow,jendap/tensorflow,kamcpp/tensorflow,strint/tensorflow,davidzchen/tensorflow,johndpope/tensorflow,Kongsea/tensorflow,mavenlin/tensorflow,tensorflow/tensorflow,dendisuhubdy/tensorflow,petewarden/tensorflow,tensorflow/tensorflow-pywrap_saved_model,mdrumond/tensorflow,ibmsoe/tensorflow,jhseu/tensorflow,tensorflow/skflow,jhseu/tensorflow,yaroslavvb/tensorflow,alivecor/tensorflow,gunan/tensorflow,TakayukiSakai/tensorflow,laosiaudi/tensorflow,kchodorow/tensorflow,dhalleine/tensorflow,scenarios/tensorflow,ran5515/DeepDecision,aselle/tensorflow,apark263/tensorflow,eadgarchen/tensorflow,AndreasMadsen/tensorflow,Bulochkin/tensorflow_pack,DCSaunders/tensorflow,awni/tensorflow,petewarden/tensorflow,petewarden/tensorflow,benoitsteiner/tensorflow-xsmm,alistairlow/tensorflow,lukeiwanski/tensorflow,brchiu/tensorflow,nanditav/15712-TensorFlow,asimshankar/tensorflow,caisq/tensorflow,Intel-tensorflow/tensorflow,aselle/tensorflow,benoitsteiner/tensorflow,ageron/tensorflow,mrry/tensorflow,panmari/tensorflow,laszlocsomor/tensorflow,nikste/tensorflow,scenarios/tensorflow,rabipanda/tensorflow,gibiansky/tensorflow,abhitopia/tensorflow,dendisuhubdy/tensorflow,aselle/tensorflow,jeffzheng1/tensorflow,freedomtan/tensorflow,pierreg/tensorflow,tornadozou/tensorflow,ivano666/tensorflow,suiyuan2009/tensorflow,zasdfgbnm/tensorflow,jwlawson/tensorflow,paolodedios/tensorflow,Mistobaan/tensorflow,chris-chris/tensorflow,llhe/tensorflow,wchan/tensorflow,manipopopo/tensorflow,rdipietro/tensorflow,petewarden/tensorflow,jart/tensorflow,theflofly/tensorflow,tillahoffmann/tensorflow,tensorflow/tensorflow,raymondxyang/tensorflow,guschmue/tensorflow,tensorflow/tensorflow-pywrap_tf_optimizer,allenlavoie/tensorflow,benoitsteiner/tensorflow,yongtang/tensorflow,DavidNorman/tensorflow,abhitopia/tensorflow,hsaputra/tensorflow,MoamerEncsConcordiaCa/tensorflow,yaroslavvb/tensorflow,theflofly/tensorflow,zycdragonball/tensorflow,xodus7/tensorflow,wangyum/tensorflow,markslwong/tensorflow,AndreasMadsen/tensorflow,yanchen036/tensorflow,memo/tensorflow,odejesush/tensorflow,MycChiu/tensorflow,brchiu/tensorflow,petewarden/tensorflow_makefile,aam-at/tensorflow,dansbecker/skflow,tntnatbry/tensorflow,llhe/tensorflow,sjperkins/tensorflow,alsrgv/tensorflow,ibmsoe/tensorflow,benoitsteiner/tensorflow,petewarden/tensorflow_makefile,rdipietro/tensorflow,manipopopo/tensorflow,gunan/tensorflow,asimshankar/tensorflow,kobejean/tensorflow,mortada/tensorflow,jart/tensorflow,aldian/tensorflow,jostep/tensorflow,thjashin/tensorflow,annarev/tensorflow,sandeepgupta2k4/tensorflow,sjperkins/tensorflow,Moriadry/tensorflow,freedomtan/tensorflow,snnn/tensorflow,Bismarrck/tensorflow,RapidApplicationDevelopment/tensorflow,andrewcmyers/tensorflow,wangyum/tensorflow,brchiu/tensorflow,DCSaunders/tensorflow,haeusser/tensorflow,eadgarchen/tensorflow,aselle/tensorflow,sandeepgupta2k4/tensorflow,renyi533/tensorflow,wchan/tensorflow,elingg/tensorflow,zycdragonball/tensorflow,EvenStrangest/tensorflow,nolanliou/tensorflow,martinwicke/tensorflow,caisq/tensorflow,DCSaunders/tensorflow,arborh/tensorflow,xzturn/tensorflow,thjashin/tensorflow,Bulochkin/tensorflow_pack,tntnatbry/tensorflow,av8ramit/tensorflow,tornadozou/tensorflow,Intel-tensorflow/tensorflow,arborh/tensorflow,wangyum/tensorflow,av8ramit/tensorflow,petewarden/tensorflow,snnn/tensorflow,cancan101/tensorflow,tensorflow/tensorflow-pywrap_tf_optimizer,elingg/tensorflow,taknevski/tensorflow-xsmm,arborh/tensorflow,gojira/tensorflow,benoitsteiner/tensorflow-xsmm,alheinecke/tensorflow-xsmm,llhe/tensorflow,dongjoon-hyun/tensorflow,thjashin/tensorflow,krikru/tensorflow-opencl,brchiu/tensorflow,aam-at/tensorflow,mixturemodel-flow/tensorflow,cg31/tensorflow,ageron/tensorflow,MycChiu/tensorflow,tensorflow/tensorflow-pywrap_tf_optimizer,nightjean/Deep-Learning,Mistobaan/tensorflow,ArtsiomCh/tensorflow,anilmuthineni/tensorflow,ville-k/tensorflow,gautam1858/tensorflow,Intel-Corporation/tensorflow,dendisuhubdy/tensorflow,alistairlow/tensorflow,dancingdan/tensorflow,sjperkins/tensorflow,MycChiu/tensorflow,yongtang/tensorflow,eerwitt/tensorflow,haeusser/tensorflow,mrry/tensorflow,horance-liu/tensorflow,tornadozou/tensorflow,tomasreimers/tensorflow-emscripten,kamcpp/tensorflow,karllessard/tensorflow,dongjoon-hyun/tensorflow,xodus7/tensorflow,codrut3/tensorflow,HaebinShin/tensorflow,hehongliang/tensorflow,strint/tensorflow,code-sauce/tensorflow,tornadozou/tensorflow,RapidApplicationDevelopment/tensorflow,annarev/tensorflow,RapidApplicationDevelopment/tensorflow,aldian/tensorflow,cg31/tensorflow,davidzchen/tensorflow,alshedivat/tensorflow,gunan/tensorflow,alivecor/tensorflow,pierreg/tensorflow,mdrumond/tensorflow,mortada/tensorflow,dhalleine/tensorflow,jendap/tensorflow,thesuperzapper/tensorflow,wangyum/tensorflow,AndreasMadsen/tensorflow,scenarios/tensorflow,alshedivat/tensorflow,RapidApplicationDevelopment/tensorflow,pcm17/tensorflow,ibab/tensorflow,nanditav/15712-TensorFlow,lukeiwanski/tensorflow,allenlavoie/tensorflow,andrewcmyers/tensorflow,jhseu/tensorflow,anand-c-goog/tensorflow,rabipanda/tensorflow,dongjoon-hyun/tensorflow,markslwong/tensorflow,andrewcmyers/tensorflow,mdrumond/tensorflow,HaebinShin/tensorflow,jhaux/tensorflow,alisidd/tensorflow
import numpy as np import tensorflow as tf from skflow import ops class DropoutTest(tf.test.TestCase): def test_dropout_float(self): - with self.test_session(): + with self.test_session() as session: x = tf.placeholder(tf.float32, [5, 5]) y = ops.dropout(x, 0.5) probs = tf.get_collection(ops.DROPOUTS) + session.run(tf.initialize_all_variables()) self.assertEqual(len(probs), 1) + self.assertEqual(session.run(probs[0]), 0.5) + + def test_dropout_tensor(self): + with self.test_session(): + x = tf.placeholder(tf.float32, [5, 5]) + y = tf.get_variable("prob", [], initializer=tf.constant_initializer(0.5)) + z = ops.dropout(x, y) + probs = tf.get_collection(ops.DROPOUTS) + self.assertEqual(probs, [y]) if __name__ == '__main__': tf.test.main()
Test for dropout probability be a tensor
## Code Before: import numpy as np import tensorflow as tf from skflow import ops class DropoutTest(tf.test.TestCase): def test_dropout_float(self): with self.test_session(): x = tf.placeholder(tf.float32, [5, 5]) y = ops.dropout(x, 0.5) probs = tf.get_collection(ops.DROPOUTS) self.assertEqual(len(probs), 1) if __name__ == '__main__': tf.test.main() ## Instruction: Test for dropout probability be a tensor ## Code After: import numpy as np import tensorflow as tf from skflow import ops class DropoutTest(tf.test.TestCase): def test_dropout_float(self): with self.test_session() as session: x = tf.placeholder(tf.float32, [5, 5]) y = ops.dropout(x, 0.5) probs = tf.get_collection(ops.DROPOUTS) session.run(tf.initialize_all_variables()) self.assertEqual(len(probs), 1) self.assertEqual(session.run(probs[0]), 0.5) def test_dropout_tensor(self): with self.test_session(): x = tf.placeholder(tf.float32, [5, 5]) y = tf.get_variable("prob", [], initializer=tf.constant_initializer(0.5)) z = ops.dropout(x, y) probs = tf.get_collection(ops.DROPOUTS) self.assertEqual(probs, [y]) if __name__ == '__main__': tf.test.main()
// ... existing code ... def test_dropout_float(self): with self.test_session() as session: x = tf.placeholder(tf.float32, [5, 5]) y = ops.dropout(x, 0.5) probs = tf.get_collection(ops.DROPOUTS) session.run(tf.initialize_all_variables()) self.assertEqual(len(probs), 1) self.assertEqual(session.run(probs[0]), 0.5) def test_dropout_tensor(self): with self.test_session(): x = tf.placeholder(tf.float32, [5, 5]) y = tf.get_variable("prob", [], initializer=tf.constant_initializer(0.5)) z = ops.dropout(x, y) probs = tf.get_collection(ops.DROPOUTS) self.assertEqual(probs, [y]) // ... rest of the code ...
b7b691d82accc012ee4308849a82ba8514e4a156
migrations/versions/20140430220209_4093ccb6d914.py
migrations/versions/20140430220209_4093ccb6d914.py
# revision identifiers, used by Alembic. revision = '4093ccb6d914' down_revision = None from alembic import op import sqlalchemy as sa from datetime import datetime def upgrade(): op.create_table('gallery', sa.Column('id', sa.Integer(), nullable=False), sa.Column('name', sa.Text(), nullable=False), sa.Column('folder', sa.types.VARCHAR(length=255), nullable=False), sa.Column('share_code', sa.Text(), nullable=False), sa.Column('modified', sa.DateTime(timezone=True), default=datetime.utcnow), sa.Column('created', sa.DateTime(timezone=True), nullable=True), sa.PrimaryKeyConstraint('id'), sa.UniqueConstraint('folder') ) op.create_table('user', sa.Column('id', sa.Integer(), nullable=False), sa.Column('name', sa.types.VARCHAR(length=255), nullable=False), sa.Column('password', sa.Text(), nullable=False), sa.Column('role', sa.Text(), nullable=False, server_default="user"), sa.PrimaryKeyConstraint('id'), sa.UniqueConstraint('name') ) def downgrade(): op.drop_table('user') op.drop_table('gallery')
# revision identifiers, used by Alembic. revision = '4093ccb6d914' down_revision = None from alembic import op import sqlalchemy as sa from datetime import datetime def upgrade(): op.create_table('gallery', sa.Column('id', sa.Integer(), nullable=False), sa.Column('name', sa.Text(), nullable=False), sa.Column('folder', sa.types.VARCHAR(length=255), nullable=False), sa.Column('share_code', sa.Text(), nullable=False), sa.Column('modified', sa.DateTime(timezone=True), default=datetime.utcnow), sa.Column('created', sa.DateTime(timezone=True), nullable=True), sa.PrimaryKeyConstraint('id'), sa.UniqueConstraint('folder') ) op.create_table('user', sa.Column('id', sa.Integer(), nullable=False), sa.Column('name', sa.types.VARCHAR(length=255), nullable=False), sa.Column('password', sa.Text(), nullable=False), sa.Column('role', sa.types.VARCHAR(length=255), nullable=False, server_default="user"), sa.PrimaryKeyConstraint('id'), sa.UniqueConstraint('name') ) def downgrade(): op.drop_table('user') op.drop_table('gallery')
Allow MySQL to set a default role
Allow MySQL to set a default role
Python
mit
taeram/ineffable,taeram/ineffable,taeram/ineffable
# revision identifiers, used by Alembic. revision = '4093ccb6d914' down_revision = None from alembic import op import sqlalchemy as sa from datetime import datetime def upgrade(): op.create_table('gallery', sa.Column('id', sa.Integer(), nullable=False), sa.Column('name', sa.Text(), nullable=False), sa.Column('folder', sa.types.VARCHAR(length=255), nullable=False), sa.Column('share_code', sa.Text(), nullable=False), sa.Column('modified', sa.DateTime(timezone=True), default=datetime.utcnow), sa.Column('created', sa.DateTime(timezone=True), nullable=True), sa.PrimaryKeyConstraint('id'), sa.UniqueConstraint('folder') ) op.create_table('user', sa.Column('id', sa.Integer(), nullable=False), sa.Column('name', sa.types.VARCHAR(length=255), nullable=False), sa.Column('password', sa.Text(), nullable=False), - sa.Column('role', sa.Text(), nullable=False, server_default="user"), + sa.Column('role', sa.types.VARCHAR(length=255), nullable=False, server_default="user"), sa.PrimaryKeyConstraint('id'), sa.UniqueConstraint('name') ) def downgrade(): op.drop_table('user') op.drop_table('gallery')
Allow MySQL to set a default role
## Code Before: # revision identifiers, used by Alembic. revision = '4093ccb6d914' down_revision = None from alembic import op import sqlalchemy as sa from datetime import datetime def upgrade(): op.create_table('gallery', sa.Column('id', sa.Integer(), nullable=False), sa.Column('name', sa.Text(), nullable=False), sa.Column('folder', sa.types.VARCHAR(length=255), nullable=False), sa.Column('share_code', sa.Text(), nullable=False), sa.Column('modified', sa.DateTime(timezone=True), default=datetime.utcnow), sa.Column('created', sa.DateTime(timezone=True), nullable=True), sa.PrimaryKeyConstraint('id'), sa.UniqueConstraint('folder') ) op.create_table('user', sa.Column('id', sa.Integer(), nullable=False), sa.Column('name', sa.types.VARCHAR(length=255), nullable=False), sa.Column('password', sa.Text(), nullable=False), sa.Column('role', sa.Text(), nullable=False, server_default="user"), sa.PrimaryKeyConstraint('id'), sa.UniqueConstraint('name') ) def downgrade(): op.drop_table('user') op.drop_table('gallery') ## Instruction: Allow MySQL to set a default role ## Code After: # revision identifiers, used by Alembic. revision = '4093ccb6d914' down_revision = None from alembic import op import sqlalchemy as sa from datetime import datetime def upgrade(): op.create_table('gallery', sa.Column('id', sa.Integer(), nullable=False), sa.Column('name', sa.Text(), nullable=False), sa.Column('folder', sa.types.VARCHAR(length=255), nullable=False), sa.Column('share_code', sa.Text(), nullable=False), sa.Column('modified', sa.DateTime(timezone=True), default=datetime.utcnow), sa.Column('created', sa.DateTime(timezone=True), nullable=True), sa.PrimaryKeyConstraint('id'), sa.UniqueConstraint('folder') ) op.create_table('user', sa.Column('id', sa.Integer(), nullable=False), sa.Column('name', sa.types.VARCHAR(length=255), nullable=False), sa.Column('password', sa.Text(), nullable=False), sa.Column('role', sa.types.VARCHAR(length=255), nullable=False, server_default="user"), sa.PrimaryKeyConstraint('id'), sa.UniqueConstraint('name') ) def downgrade(): op.drop_table('user') op.drop_table('gallery')
# ... existing code ... sa.Column('name', sa.types.VARCHAR(length=255), nullable=False), sa.Column('password', sa.Text(), nullable=False), sa.Column('role', sa.types.VARCHAR(length=255), nullable=False, server_default="user"), sa.PrimaryKeyConstraint('id'), sa.UniqueConstraint('name') # ... rest of the code ...
306cf5987c90d54d72037c19dd02f07be37cbb6f
make_mozilla/base/tests/decorators.py
make_mozilla/base/tests/decorators.py
from functools import wraps from nose.plugins.attrib import attr from nose.plugins.skip import SkipTest __all__ = ['wip'] def fail(message): raise AssertionError(message) def wip(f): @wraps(f) def run_test(*args, **kwargs): try: f(*args, **kwargs) except Exception as e: raise SkipTest("WIP test failed: " + str(e)) fail("test passed but marked as work in progress") return attr('wip')(run_test)
from functools import wraps from nose.plugins.attrib import attr from nose.plugins.skip import SkipTest import os __all__ = ['wip'] def fail(message): raise AssertionError(message) def wip(f): @wraps(f) def run_test(*args, **kwargs): try: f(*args, **kwargs) except Exception as e: raise SkipTest("WIP test failed: " + str(e)) fail("test passed but marked as work in progress") return attr('wip')(run_test) def integration(f): @wraps(f) def run_test(*args, **kwargs): integration_run = (os.getenv('INTEGRATION', None) is not None) if integration_run: f(*args, **kwargs) else: raise SkipTest("Skipping integration test") return attr('integration')(run_test)
Add integration test decorator to prevent certain tests running unless we really want them to.
Add integration test decorator to prevent certain tests running unless we really want them to.
Python
bsd-3-clause
mozilla/make.mozilla.org,mozilla/make.mozilla.org,mozilla/make.mozilla.org,mozilla/make.mozilla.org
from functools import wraps from nose.plugins.attrib import attr from nose.plugins.skip import SkipTest + import os __all__ = ['wip'] def fail(message): raise AssertionError(message) def wip(f): @wraps(f) def run_test(*args, **kwargs): try: f(*args, **kwargs) except Exception as e: raise SkipTest("WIP test failed: " + str(e)) fail("test passed but marked as work in progress") return attr('wip')(run_test) + def integration(f): + @wraps(f) + def run_test(*args, **kwargs): + integration_run = (os.getenv('INTEGRATION', None) is not None) + if integration_run: + f(*args, **kwargs) + else: + raise SkipTest("Skipping integration test") + return attr('integration')(run_test) +
Add integration test decorator to prevent certain tests running unless we really want them to.
## Code Before: from functools import wraps from nose.plugins.attrib import attr from nose.plugins.skip import SkipTest __all__ = ['wip'] def fail(message): raise AssertionError(message) def wip(f): @wraps(f) def run_test(*args, **kwargs): try: f(*args, **kwargs) except Exception as e: raise SkipTest("WIP test failed: " + str(e)) fail("test passed but marked as work in progress") return attr('wip')(run_test) ## Instruction: Add integration test decorator to prevent certain tests running unless we really want them to. ## Code After: from functools import wraps from nose.plugins.attrib import attr from nose.plugins.skip import SkipTest import os __all__ = ['wip'] def fail(message): raise AssertionError(message) def wip(f): @wraps(f) def run_test(*args, **kwargs): try: f(*args, **kwargs) except Exception as e: raise SkipTest("WIP test failed: " + str(e)) fail("test passed but marked as work in progress") return attr('wip')(run_test) def integration(f): @wraps(f) def run_test(*args, **kwargs): integration_run = (os.getenv('INTEGRATION', None) is not None) if integration_run: f(*args, **kwargs) else: raise SkipTest("Skipping integration test") return attr('integration')(run_test)
// ... existing code ... from nose.plugins.attrib import attr from nose.plugins.skip import SkipTest import os __all__ = ['wip'] // ... modified code ... return attr('wip')(run_test) def integration(f): @wraps(f) def run_test(*args, **kwargs): integration_run = (os.getenv('INTEGRATION', None) is not None) if integration_run: f(*args, **kwargs) else: raise SkipTest("Skipping integration test") return attr('integration')(run_test) // ... rest of the code ...
28a4f4ab9d6b7c3ea14d48c002273acfe05d7246
bumblebee/util.py
bumblebee/util.py
import shlex import exceptions import subprocess def bytefmt(num): for unit in [ "", "Ki", "Mi", "Gi" ]: if num < 1024.0: return "{:.2f}{}B".format(num, unit) num /= 1024.0 return "{:05.2f%}{}GiB".format(num) def durationfmt(duration): minutes, seconds = divmod(duration, 60) hours, minutes = divmod(minutes, 60) res = "{:02d}:{:02d}".format(minutes, seconds) if hours > 0: res = "{:02d}:{}".format(hours, res) return res def execute(cmd): args = shlex.split(cmd) p = subprocess.Popen(args, stdout=subprocess.PIPE, stderr=subprocess.STDOUT) out = p.communicate() if p.returncode != 0: raise exceptions.RuntimeError("{} exited with {}".format(cmd, p.returncode))
import shlex import subprocess try: from exceptions import RuntimeError except ImportError: # Python3 doesn't require this anymore pass def bytefmt(num): for unit in [ "", "Ki", "Mi", "Gi" ]: if num < 1024.0: return "{:.2f}{}B".format(num, unit) num /= 1024.0 return "{:05.2f%}{}GiB".format(num) def durationfmt(duration): minutes, seconds = divmod(duration, 60) hours, minutes = divmod(minutes, 60) res = "{:02d}:{:02d}".format(minutes, seconds) if hours > 0: res = "{:02d}:{}".format(hours, res) return res def execute(cmd): args = shlex.split(cmd) p = subprocess.Popen(args, stdout=subprocess.PIPE, stderr=subprocess.STDOUT) out = p.communicate() if p.returncode != 0: raise RuntimeError("{} exited with {}".format(cmd, p.returncode))
Fix import error for Python3
[core] Fix import error for Python3 Import exceptions module only for Python2. fixes #22
Python
mit
tobi-wan-kenobi/bumblebee-status,tobi-wan-kenobi/bumblebee-status
import shlex - import exceptions import subprocess + try: + from exceptions import RuntimeError + except ImportError: + # Python3 doesn't require this anymore + pass def bytefmt(num): for unit in [ "", "Ki", "Mi", "Gi" ]: if num < 1024.0: return "{:.2f}{}B".format(num, unit) num /= 1024.0 return "{:05.2f%}{}GiB".format(num) def durationfmt(duration): minutes, seconds = divmod(duration, 60) hours, minutes = divmod(minutes, 60) res = "{:02d}:{:02d}".format(minutes, seconds) if hours > 0: res = "{:02d}:{}".format(hours, res) return res def execute(cmd): args = shlex.split(cmd) p = subprocess.Popen(args, stdout=subprocess.PIPE, stderr=subprocess.STDOUT) out = p.communicate() if p.returncode != 0: - raise exceptions.RuntimeError("{} exited with {}".format(cmd, p.returncode)) + raise RuntimeError("{} exited with {}".format(cmd, p.returncode))
Fix import error for Python3
## Code Before: import shlex import exceptions import subprocess def bytefmt(num): for unit in [ "", "Ki", "Mi", "Gi" ]: if num < 1024.0: return "{:.2f}{}B".format(num, unit) num /= 1024.0 return "{:05.2f%}{}GiB".format(num) def durationfmt(duration): minutes, seconds = divmod(duration, 60) hours, minutes = divmod(minutes, 60) res = "{:02d}:{:02d}".format(minutes, seconds) if hours > 0: res = "{:02d}:{}".format(hours, res) return res def execute(cmd): args = shlex.split(cmd) p = subprocess.Popen(args, stdout=subprocess.PIPE, stderr=subprocess.STDOUT) out = p.communicate() if p.returncode != 0: raise exceptions.RuntimeError("{} exited with {}".format(cmd, p.returncode)) ## Instruction: Fix import error for Python3 ## Code After: import shlex import subprocess try: from exceptions import RuntimeError except ImportError: # Python3 doesn't require this anymore pass def bytefmt(num): for unit in [ "", "Ki", "Mi", "Gi" ]: if num < 1024.0: return "{:.2f}{}B".format(num, unit) num /= 1024.0 return "{:05.2f%}{}GiB".format(num) def durationfmt(duration): minutes, seconds = divmod(duration, 60) hours, minutes = divmod(minutes, 60) res = "{:02d}:{:02d}".format(minutes, seconds) if hours > 0: res = "{:02d}:{}".format(hours, res) return res def execute(cmd): args = shlex.split(cmd) p = subprocess.Popen(args, stdout=subprocess.PIPE, stderr=subprocess.STDOUT) out = p.communicate() if p.returncode != 0: raise RuntimeError("{} exited with {}".format(cmd, p.returncode))
... import shlex import subprocess try: from exceptions import RuntimeError except ImportError: # Python3 doesn't require this anymore pass def bytefmt(num): ... if p.returncode != 0: raise RuntimeError("{} exited with {}".format(cmd, p.returncode)) ...
0683645a2fb2323a9534d985005d843aada66040
anypytools/__init__.py
anypytools/__init__.py
"""AnyPyTools library.""" import sys import platform import logging from anypytools.abcutils import AnyPyProcess from anypytools.macroutils import AnyMacro from anypytools import macro_commands logger = logging.getLogger('abt.anypytools') logger.addHandler(logging.NullHandler()) __all__ = [ 'datautils', 'h5py_wrapper', 'AnyPyProcess', 'AnyMacro', 'macro_commands', 'print_versions', ] __version__ = '0.10.2' def print_versions(): """Print all the versions of software that AnyPyTools relies on.""" import numpy as np import scipy as sp print("-=" * 38) print("AnyPyTools version: %s" % __version__) print("NumPy version: %s" % np.__version__) print("SciPy version: %s" % sp.__version__) print("Python version: %s" % sys.version) (sysname, nodename, release, version, machine, processor) = \ platform.uname() print("Platform: %s-%s-%s (%s)" % (sysname, release, machine, version)) if sysname == "Linux": print("Linux dist: %s" % " ".join(platform.linux_distribution()[:-1])) if not processor: processor = "not recognized" print("Processor: %s" % processor) print("Byte-ordering: %s" % sys.byteorder) print("-=" * 38)
"""AnyPyTools library.""" import sys import platform import logging from anypytools.abcutils import AnyPyProcess, execute_anybodycon from anypytools.macroutils import AnyMacro from anypytools import macro_commands logger = logging.getLogger('abt.anypytools') logger.addHandler(logging.NullHandler()) __all__ = [ 'datautils', 'h5py_wrapper', 'AnyPyProcess', 'AnyMacro', 'macro_commands', 'print_versions', 'execute_anybodycon', ] __version__ = '0.10.2' def print_versions(): """Print all the versions of software that AnyPyTools relies on.""" import numpy as np import scipy as sp print("-=" * 38) print("AnyPyTools version: %s" % __version__) print("NumPy version: %s" % np.__version__) print("SciPy version: %s" % sp.__version__) print("Python version: %s" % sys.version) (sysname, nodename, release, version, machine, processor) = \ platform.uname() print("Platform: %s-%s-%s (%s)" % (sysname, release, machine, version)) if sysname == "Linux": print("Linux dist: %s" % " ".join(platform.linux_distribution()[:-1])) if not processor: processor = "not recognized" print("Processor: %s" % processor) print("Byte-ordering: %s" % sys.byteorder) print("-=" * 38)
Add execute_anybodycon to toplevel package
Add execute_anybodycon to toplevel package
Python
mit
AnyBody-Research-Group/AnyPyTools
"""AnyPyTools library.""" import sys import platform import logging - from anypytools.abcutils import AnyPyProcess + from anypytools.abcutils import AnyPyProcess, execute_anybodycon from anypytools.macroutils import AnyMacro from anypytools import macro_commands logger = logging.getLogger('abt.anypytools') logger.addHandler(logging.NullHandler()) __all__ = [ 'datautils', 'h5py_wrapper', 'AnyPyProcess', 'AnyMacro', 'macro_commands', - 'print_versions', + 'print_versions', 'execute_anybodycon', ] __version__ = '0.10.2' def print_versions(): """Print all the versions of software that AnyPyTools relies on.""" import numpy as np import scipy as sp print("-=" * 38) print("AnyPyTools version: %s" % __version__) print("NumPy version: %s" % np.__version__) print("SciPy version: %s" % sp.__version__) print("Python version: %s" % sys.version) (sysname, nodename, release, version, machine, processor) = \ platform.uname() print("Platform: %s-%s-%s (%s)" % (sysname, release, machine, version)) if sysname == "Linux": print("Linux dist: %s" % " ".join(platform.linux_distribution()[:-1])) if not processor: processor = "not recognized" print("Processor: %s" % processor) print("Byte-ordering: %s" % sys.byteorder) print("-=" * 38)
Add execute_anybodycon to toplevel package
## Code Before: """AnyPyTools library.""" import sys import platform import logging from anypytools.abcutils import AnyPyProcess from anypytools.macroutils import AnyMacro from anypytools import macro_commands logger = logging.getLogger('abt.anypytools') logger.addHandler(logging.NullHandler()) __all__ = [ 'datautils', 'h5py_wrapper', 'AnyPyProcess', 'AnyMacro', 'macro_commands', 'print_versions', ] __version__ = '0.10.2' def print_versions(): """Print all the versions of software that AnyPyTools relies on.""" import numpy as np import scipy as sp print("-=" * 38) print("AnyPyTools version: %s" % __version__) print("NumPy version: %s" % np.__version__) print("SciPy version: %s" % sp.__version__) print("Python version: %s" % sys.version) (sysname, nodename, release, version, machine, processor) = \ platform.uname() print("Platform: %s-%s-%s (%s)" % (sysname, release, machine, version)) if sysname == "Linux": print("Linux dist: %s" % " ".join(platform.linux_distribution()[:-1])) if not processor: processor = "not recognized" print("Processor: %s" % processor) print("Byte-ordering: %s" % sys.byteorder) print("-=" * 38) ## Instruction: Add execute_anybodycon to toplevel package ## Code After: """AnyPyTools library.""" import sys import platform import logging from anypytools.abcutils import AnyPyProcess, execute_anybodycon from anypytools.macroutils import AnyMacro from anypytools import macro_commands logger = logging.getLogger('abt.anypytools') logger.addHandler(logging.NullHandler()) __all__ = [ 'datautils', 'h5py_wrapper', 'AnyPyProcess', 'AnyMacro', 'macro_commands', 'print_versions', 'execute_anybodycon', ] __version__ = '0.10.2' def print_versions(): """Print all the versions of software that AnyPyTools relies on.""" import numpy as np import scipy as sp print("-=" * 38) print("AnyPyTools version: %s" % __version__) print("NumPy version: %s" % np.__version__) print("SciPy version: %s" % sp.__version__) print("Python version: %s" % sys.version) (sysname, nodename, release, version, machine, processor) = \ platform.uname() print("Platform: %s-%s-%s (%s)" % (sysname, release, machine, version)) if sysname == "Linux": print("Linux dist: %s" % " ".join(platform.linux_distribution()[:-1])) if not processor: processor = "not recognized" print("Processor: %s" % processor) print("Byte-ordering: %s" % sys.byteorder) print("-=" * 38)
// ... existing code ... import logging from anypytools.abcutils import AnyPyProcess, execute_anybodycon from anypytools.macroutils import AnyMacro from anypytools import macro_commands // ... modified code ... __all__ = [ 'datautils', 'h5py_wrapper', 'AnyPyProcess', 'AnyMacro', 'macro_commands', 'print_versions', 'execute_anybodycon', ] // ... rest of the code ...
6f5be9af15898f089c3ee83ca1f05fbd4570fcfa
src/cms/apps/news/models.py
src/cms/apps/news/models.py
"""Models used by the news publication application.""" from django.db import models from cms.apps.pages.models import Page, PageBase, PageField, HtmlField from cms.apps.news.content import NewsFeed, NewsArticle class Article(PageBase): """A news article.""" news_feed = PageField(Page, "newsfeed") content = HtmlField(blank=True, null=True) summary = models.TextField(blank=True, null=True, help_text="A short summary of this article. This will be used on news pages and RSS feeds. If not specified, then a summarized version of the content will be used.") is_featured = models.BooleanField("featured", default=False, help_text="Featured articles will remain at the top of any news feeds.") def get_absolute_url(self): """Returns the absolute URL of the article.""" return self.parent.content.reverse("article_detail", self.publication_date.year, self.publication_date.month, self.url_title, "") class Meta: verbose_name = "news article" Page.register_content(NewsFeed)
"""Models used by the news publication application.""" from django.db import models from cms.apps.pages.models import Page, PageBase, PageField, HtmlField from cms.apps.news.content import NewsFeed, NewsArticle class Article(PageBase): """A news article.""" news_feed = PageField(Page, "newsfeed") content = HtmlField(blank=True, null=True) summary = models.TextField(blank=True, null=True, help_text="A short summary of this article. This will be used on news pages and RSS feeds. If not specified, then a summarized version of the content will be used.") is_featured = models.BooleanField("featured", default=False, help_text="Featured articles will remain at the top of any news feeds.") def get_absolute_url(self): """Returns the absolute URL of the article.""" return self.parent.content.reverse("article_detail", self.publication_date.year, self.publication_date.month, self.url_title, "") class Meta: verbose_name = "news article" unique_together = (("news_feed", "url_title",),) Page.register_content(NewsFeed)
Set unique together on news article.
Set unique together on news article.
Python
bsd-3-clause
lewiscollard/cms,etianen/cms,etianen/cms,danielsamuels/cms,jamesfoley/cms,lewiscollard/cms,jamesfoley/cms,dan-gamble/cms,lewiscollard/cms,jamesfoley/cms,dan-gamble/cms,dan-gamble/cms,jamesfoley/cms,danielsamuels/cms,etianen/cms,danielsamuels/cms
"""Models used by the news publication application.""" from django.db import models from cms.apps.pages.models import Page, PageBase, PageField, HtmlField from cms.apps.news.content import NewsFeed, NewsArticle class Article(PageBase): """A news article.""" news_feed = PageField(Page, "newsfeed") content = HtmlField(blank=True, null=True) summary = models.TextField(blank=True, null=True, help_text="A short summary of this article. This will be used on news pages and RSS feeds. If not specified, then a summarized version of the content will be used.") is_featured = models.BooleanField("featured", default=False, help_text="Featured articles will remain at the top of any news feeds.") def get_absolute_url(self): """Returns the absolute URL of the article.""" return self.parent.content.reverse("article_detail", self.publication_date.year, self.publication_date.month, self.url_title, "") class Meta: verbose_name = "news article" + unique_together = (("news_feed", "url_title",),) - - Page.register_content(NewsFeed)
Set unique together on news article.
## Code Before: """Models used by the news publication application.""" from django.db import models from cms.apps.pages.models import Page, PageBase, PageField, HtmlField from cms.apps.news.content import NewsFeed, NewsArticle class Article(PageBase): """A news article.""" news_feed = PageField(Page, "newsfeed") content = HtmlField(blank=True, null=True) summary = models.TextField(blank=True, null=True, help_text="A short summary of this article. This will be used on news pages and RSS feeds. If not specified, then a summarized version of the content will be used.") is_featured = models.BooleanField("featured", default=False, help_text="Featured articles will remain at the top of any news feeds.") def get_absolute_url(self): """Returns the absolute URL of the article.""" return self.parent.content.reverse("article_detail", self.publication_date.year, self.publication_date.month, self.url_title, "") class Meta: verbose_name = "news article" Page.register_content(NewsFeed) ## Instruction: Set unique together on news article. ## Code After: """Models used by the news publication application.""" from django.db import models from cms.apps.pages.models import Page, PageBase, PageField, HtmlField from cms.apps.news.content import NewsFeed, NewsArticle class Article(PageBase): """A news article.""" news_feed = PageField(Page, "newsfeed") content = HtmlField(blank=True, null=True) summary = models.TextField(blank=True, null=True, help_text="A short summary of this article. This will be used on news pages and RSS feeds. If not specified, then a summarized version of the content will be used.") is_featured = models.BooleanField("featured", default=False, help_text="Featured articles will remain at the top of any news feeds.") def get_absolute_url(self): """Returns the absolute URL of the article.""" return self.parent.content.reverse("article_detail", self.publication_date.year, self.publication_date.month, self.url_title, "") class Meta: verbose_name = "news article" unique_together = (("news_feed", "url_title",),) Page.register_content(NewsFeed)
// ... existing code ... class Meta: verbose_name = "news article" unique_together = (("news_feed", "url_title",),) // ... rest of the code ...
3990e3aa64cff288def07ee36e24026cc15282c0
taiga/projects/issues/serializers.py
taiga/projects/issues/serializers.py
from rest_framework import serializers from taiga.base.serializers import PickleField, NeighborsSerializerMixin from . import models class IssueSerializer(serializers.ModelSerializer): tags = PickleField(required=False) comment = serializers.SerializerMethodField("get_comment") is_closed = serializers.Field(source="is_closed") class Meta: model = models.Issue def get_comment(self, obj): return "" class IssueNeighborsSerializer(NeighborsSerializerMixin, IssueSerializer): def serialize_neighbor(self, neighbor): return NeighborIssueSerializer(neighbor).data class NeighborIssueSerializer(serializers.ModelSerializer): class Meta: model = models.Issue fields = ("id", "ref", "subject") depth = 0
from rest_framework import serializers from taiga.base.serializers import PickleField, NeighborsSerializerMixin from . import models class IssueSerializer(serializers.ModelSerializer): tags = PickleField(required=False) is_closed = serializers.Field(source="is_closed") class Meta: model = models.Issue class IssueNeighborsSerializer(NeighborsSerializerMixin, IssueSerializer): def serialize_neighbor(self, neighbor): return NeighborIssueSerializer(neighbor).data class NeighborIssueSerializer(serializers.ModelSerializer): class Meta: model = models.Issue fields = ("id", "ref", "subject") depth = 0
Remove unnecessary field from IssueSerializer
Remove unnecessary field from IssueSerializer
Python
agpl-3.0
forging2012/taiga-back,EvgeneOskin/taiga-back,xdevelsistemas/taiga-back-community,seanchen/taiga-back,bdang2012/taiga-back-casting,Rademade/taiga-back,crr0004/taiga-back,dayatz/taiga-back,rajiteh/taiga-back,dycodedev/taiga-back,crr0004/taiga-back,obimod/taiga-back,Zaneh-/bearded-tribble-back,seanchen/taiga-back,gauravjns/taiga-back,joshisa/taiga-back,19kestier/taiga-back,jeffdwyatt/taiga-back,taigaio/taiga-back,WALR/taiga-back,joshisa/taiga-back,astronaut1712/taiga-back,taigaio/taiga-back,coopsource/taiga-back,gam-phon/taiga-back,Rademade/taiga-back,obimod/taiga-back,obimod/taiga-back,CMLL/taiga-back,frt-arch/taiga-back,dycodedev/taiga-back,bdang2012/taiga-back-casting,Tigerwhit4/taiga-back,19kestier/taiga-back,EvgeneOskin/taiga-back,EvgeneOskin/taiga-back,astagi/taiga-back,bdang2012/taiga-back-casting,Zaneh-/bearded-tribble-back,dayatz/taiga-back,CoolCloud/taiga-back,astronaut1712/taiga-back,jeffdwyatt/taiga-back,crr0004/taiga-back,WALR/taiga-back,gam-phon/taiga-back,CMLL/taiga-back,seanchen/taiga-back,astagi/taiga-back,gauravjns/taiga-back,gam-phon/taiga-back,WALR/taiga-back,jeffdwyatt/taiga-back,Tigerwhit4/taiga-back,Zaneh-/bearded-tribble-back,seanchen/taiga-back,xdevelsistemas/taiga-back-community,coopsource/taiga-back,astagi/taiga-back,EvgeneOskin/taiga-back,obimod/taiga-back,gam-phon/taiga-back,coopsource/taiga-back,CoolCloud/taiga-back,rajiteh/taiga-back,dycodedev/taiga-back,bdang2012/taiga-back-casting,19kestier/taiga-back,astronaut1712/taiga-back,forging2012/taiga-back,CMLL/taiga-back,frt-arch/taiga-back,astagi/taiga-back,WALR/taiga-back,forging2012/taiga-back,rajiteh/taiga-back,frt-arch/taiga-back,Rademade/taiga-back,xdevelsistemas/taiga-back-community,taigaio/taiga-back,joshisa/taiga-back,gauravjns/taiga-back,Rademade/taiga-back,crr0004/taiga-back,forging2012/taiga-back,joshisa/taiga-back,CMLL/taiga-back,dycodedev/taiga-back,coopsource/taiga-back,CoolCloud/taiga-back,Rademade/taiga-back,astronaut1712/taiga-back,jeffdwyatt/taiga-back,CoolCloud/taiga-back,gauravjns/taiga-back,rajiteh/taiga-back,dayatz/taiga-back,Tigerwhit4/taiga-back,Tigerwhit4/taiga-back
from rest_framework import serializers from taiga.base.serializers import PickleField, NeighborsSerializerMixin from . import models class IssueSerializer(serializers.ModelSerializer): tags = PickleField(required=False) - comment = serializers.SerializerMethodField("get_comment") is_closed = serializers.Field(source="is_closed") class Meta: model = models.Issue - - def get_comment(self, obj): - return "" class IssueNeighborsSerializer(NeighborsSerializerMixin, IssueSerializer): def serialize_neighbor(self, neighbor): return NeighborIssueSerializer(neighbor).data class NeighborIssueSerializer(serializers.ModelSerializer): class Meta: model = models.Issue fields = ("id", "ref", "subject") depth = 0
Remove unnecessary field from IssueSerializer
## Code Before: from rest_framework import serializers from taiga.base.serializers import PickleField, NeighborsSerializerMixin from . import models class IssueSerializer(serializers.ModelSerializer): tags = PickleField(required=False) comment = serializers.SerializerMethodField("get_comment") is_closed = serializers.Field(source="is_closed") class Meta: model = models.Issue def get_comment(self, obj): return "" class IssueNeighborsSerializer(NeighborsSerializerMixin, IssueSerializer): def serialize_neighbor(self, neighbor): return NeighborIssueSerializer(neighbor).data class NeighborIssueSerializer(serializers.ModelSerializer): class Meta: model = models.Issue fields = ("id", "ref", "subject") depth = 0 ## Instruction: Remove unnecessary field from IssueSerializer ## Code After: from rest_framework import serializers from taiga.base.serializers import PickleField, NeighborsSerializerMixin from . import models class IssueSerializer(serializers.ModelSerializer): tags = PickleField(required=False) is_closed = serializers.Field(source="is_closed") class Meta: model = models.Issue class IssueNeighborsSerializer(NeighborsSerializerMixin, IssueSerializer): def serialize_neighbor(self, neighbor): return NeighborIssueSerializer(neighbor).data class NeighborIssueSerializer(serializers.ModelSerializer): class Meta: model = models.Issue fields = ("id", "ref", "subject") depth = 0
... class IssueSerializer(serializers.ModelSerializer): tags = PickleField(required=False) is_closed = serializers.Field(source="is_closed") ... class Meta: model = models.Issue ...
76551f7a05506a872ec6535eb3263710650ea8ce
glue/core/data_factories/__init__.py
glue/core/data_factories/__init__.py
from .helpers import * from .gridded import * from .pandas import * from .excel import * from .image import * from .dendrogram import * from .tables import *
from .helpers import * from .gridded import * from .pandas import * from .excel import * from .image import * from .tables import * from .dendrogram import *
Order of import matters for disambiguation, but this should be fixed later to avoid this.
Order of import matters for disambiguation, but this should be fixed later to avoid this.
Python
bsd-3-clause
saimn/glue,stscieisenhamer/glue,stscieisenhamer/glue,JudoWill/glue,saimn/glue,JudoWill/glue
from .helpers import * from .gridded import * from .pandas import * from .excel import * from .image import * + from .tables import * from .dendrogram import * - from .tables import *
Order of import matters for disambiguation, but this should be fixed later to avoid this.
## Code Before: from .helpers import * from .gridded import * from .pandas import * from .excel import * from .image import * from .dendrogram import * from .tables import * ## Instruction: Order of import matters for disambiguation, but this should be fixed later to avoid this. ## Code After: from .helpers import * from .gridded import * from .pandas import * from .excel import * from .image import * from .tables import * from .dendrogram import *
... from .excel import * from .image import * from .tables import * from .dendrogram import * ...
f57c9643a32cca012fdccac40899c6de38e35af9
ass/ets/__init__.py
ass/ets/__init__.py
from bundles import Environment, Assets, Bundle, Manifest import filters as f from options import Option, Options, Undefined, dict_getter
from bundles import Environment, Assets, Bundle, Manifest import filters as f from options import Option, Options, Undefined, dict_getter from pipeable import worker
Make @worker available for import on ass.ets.
Make @worker available for import on ass.ets.
Python
bsd-2-clause
kaste/ass.ets,kaste/ass.ets
from bundles import Environment, Assets, Bundle, Manifest import filters as f from options import Option, Options, Undefined, dict_getter + from pipeable import worker
Make @worker available for import on ass.ets.
## Code Before: from bundles import Environment, Assets, Bundle, Manifest import filters as f from options import Option, Options, Undefined, dict_getter ## Instruction: Make @worker available for import on ass.ets. ## Code After: from bundles import Environment, Assets, Bundle, Manifest import filters as f from options import Option, Options, Undefined, dict_getter from pipeable import worker
// ... existing code ... import filters as f from options import Option, Options, Undefined, dict_getter from pipeable import worker // ... rest of the code ...
6dd04ed490c49c85bf91db2cb0bf2bed82b5967b
fasttsne/__init__.py
fasttsne/__init__.py
import scipy.linalg as la import numpy as np from py_bh_tsne import _TSNE as TSNE def fast_tsne(data, pca_d=50, d=2, perplexity=30., theta=0.5): """ Run Barnes-Hut T-SNE on _data_. @param data The data. @param pca_d The dimensionality of data is reduced via PCA to this dimensionality. @param d The embedding dimensionality. Must be fixed to 2. @param perplexity The perplexity controls the effective number of neighbors. @param theta If set to 0, exact t-SNE is run, which takes very long for dataset > 5000 samples. """ N, _ = data.shape # inplace!! data -= data.mean(axis=0) # working with covariance + (svd on cov.) is # much faster than svd on data directly. cov = np.dot(data.T, data)/N u, s, v = la.svd(cov, full_matrices=False) u = u[:,0:pca_d] X = np.dot(data, u) tsne = TSNE() Y = tsne.run(X, N, pca_d, d, perplexity, theta) return Y
import scipy.linalg as la import numpy as np from fasttsne import _TSNE as TSNE def fast_tsne(data, pca_d=None, d=2, perplexity=30., theta=0.5): """ Run Barnes-Hut T-SNE on _data_. @param data The data. @param pca_d The dimensionality of data is reduced via PCA to this dimensionality. @param d The embedding dimensionality. Must be fixed to 2. @param perplexity The perplexity controls the effective number of neighbors. @param theta If set to 0, exact t-SNE is run, which takes very long for dataset > 5000 samples. """ N, _ = data.shape # inplace!! if pca_d is None: X = data else: # do PCA data -= data.mean(axis=0) # working with covariance + (svd on cov.) is # much faster than svd on data directly. cov = np.dot(data.T, data)/N u, s, v = la.svd(cov, full_matrices=False) u = u[:,0:pca_d] X = np.dot(data, u) tsne = TSNE() Y = tsne.run(X, N, X.shape[1], d, perplexity, theta) return Y
FIX (from Justin Bayer): avoid memory segfault when pca_d is choosen too big.
FIX (from Justin Bayer): avoid memory segfault when pca_d is choosen too big.
Python
bsd-3-clause
pryvkin10x/tsne,douglasbagnall/py_bh_tsne,douglasbagnall/py_bh_tsne,pryvkin10x/tsne,pryvkin10x/tsne
import scipy.linalg as la import numpy as np - from py_bh_tsne import _TSNE as TSNE + from fasttsne import _TSNE as TSNE - def fast_tsne(data, pca_d=50, d=2, perplexity=30., theta=0.5): + def fast_tsne(data, pca_d=None, d=2, perplexity=30., theta=0.5): """ Run Barnes-Hut T-SNE on _data_. @param data The data. @param pca_d The dimensionality of data is reduced via PCA to this dimensionality. @param d The embedding dimensionality. Must be fixed to 2. @param perplexity The perplexity controls the effective number of neighbors. @param theta If set to 0, exact t-SNE is run, which takes very long for dataset > 5000 samples. """ N, _ = data.shape # inplace!! + + if pca_d is None: + X = data + else: + # do PCA - data -= data.mean(axis=0) + data -= data.mean(axis=0) - # working with covariance + (svd on cov.) is + # working with covariance + (svd on cov.) is - # much faster than svd on data directly. + # much faster than svd on data directly. - cov = np.dot(data.T, data)/N + cov = np.dot(data.T, data)/N - u, s, v = la.svd(cov, full_matrices=False) + u, s, v = la.svd(cov, full_matrices=False) - u = u[:,0:pca_d] + u = u[:,0:pca_d] - X = np.dot(data, u) + X = np.dot(data, u) tsne = TSNE() - Y = tsne.run(X, N, pca_d, d, perplexity, theta) + Y = tsne.run(X, N, X.shape[1], d, perplexity, theta) return Y
FIX (from Justin Bayer): avoid memory segfault when pca_d is choosen too big.
## Code Before: import scipy.linalg as la import numpy as np from py_bh_tsne import _TSNE as TSNE def fast_tsne(data, pca_d=50, d=2, perplexity=30., theta=0.5): """ Run Barnes-Hut T-SNE on _data_. @param data The data. @param pca_d The dimensionality of data is reduced via PCA to this dimensionality. @param d The embedding dimensionality. Must be fixed to 2. @param perplexity The perplexity controls the effective number of neighbors. @param theta If set to 0, exact t-SNE is run, which takes very long for dataset > 5000 samples. """ N, _ = data.shape # inplace!! data -= data.mean(axis=0) # working with covariance + (svd on cov.) is # much faster than svd on data directly. cov = np.dot(data.T, data)/N u, s, v = la.svd(cov, full_matrices=False) u = u[:,0:pca_d] X = np.dot(data, u) tsne = TSNE() Y = tsne.run(X, N, pca_d, d, perplexity, theta) return Y ## Instruction: FIX (from Justin Bayer): avoid memory segfault when pca_d is choosen too big. ## Code After: import scipy.linalg as la import numpy as np from fasttsne import _TSNE as TSNE def fast_tsne(data, pca_d=None, d=2, perplexity=30., theta=0.5): """ Run Barnes-Hut T-SNE on _data_. @param data The data. @param pca_d The dimensionality of data is reduced via PCA to this dimensionality. @param d The embedding dimensionality. Must be fixed to 2. @param perplexity The perplexity controls the effective number of neighbors. @param theta If set to 0, exact t-SNE is run, which takes very long for dataset > 5000 samples. """ N, _ = data.shape # inplace!! if pca_d is None: X = data else: # do PCA data -= data.mean(axis=0) # working with covariance + (svd on cov.) is # much faster than svd on data directly. cov = np.dot(data.T, data)/N u, s, v = la.svd(cov, full_matrices=False) u = u[:,0:pca_d] X = np.dot(data, u) tsne = TSNE() Y = tsne.run(X, N, X.shape[1], d, perplexity, theta) return Y
# ... existing code ... from fasttsne import _TSNE as TSNE def fast_tsne(data, pca_d=None, d=2, perplexity=30., theta=0.5): """ Run Barnes-Hut T-SNE on _data_. # ... modified code ... # inplace!! if pca_d is None: X = data else: # do PCA data -= data.mean(axis=0) # working with covariance + (svd on cov.) is # much faster than svd on data directly. cov = np.dot(data.T, data)/N u, s, v = la.svd(cov, full_matrices=False) u = u[:,0:pca_d] X = np.dot(data, u) tsne = TSNE() Y = tsne.run(X, N, X.shape[1], d, perplexity, theta) return Y # ... rest of the code ...
94d47cfc6db684beda275f8658660a3bd92b319d
src/syft/grid/client/request_api/user_api.py
src/syft/grid/client/request_api/user_api.py
from typing import Any from typing import Dict # third party from pandas import DataFrame # syft relative from ...messages.user_messages import CreateUserMessage from ...messages.user_messages import DeleteUserMessage from ...messages.user_messages import GetUserMessage from ...messages.user_messages import GetUsersMessage from ...messages.user_messages import UpdateUserMessage from .request_api import GridRequestAPI class UserRequestAPI(GridRequestAPI): response_key = "user" def __init__(self, send): super().__init__( create_msg=CreateUserMessage, get_msg=GetUserMessage, get_all_msg=GetUsersMessage, update_msg=UpdateUserMessage, delete_msg=DeleteUserMessage, send=send, response_key=UserRequestAPI.response_key, ) def __getitem__(self, key): return self.get(user_id=key) def __delitem__(self, key): self.delete(user_id=key)
from typing import Any from typing import Callable # syft relative from ...messages.user_messages import CreateUserMessage from ...messages.user_messages import DeleteUserMessage from ...messages.user_messages import GetUserMessage from ...messages.user_messages import GetUsersMessage from ...messages.user_messages import UpdateUserMessage from .request_api import GridRequestAPI class UserRequestAPI(GridRequestAPI): response_key = "user" def __init__(self, send: Callable): super().__init__( create_msg=CreateUserMessage, get_msg=GetUserMessage, get_all_msg=GetUsersMessage, update_msg=UpdateUserMessage, delete_msg=DeleteUserMessage, send=send, response_key=UserRequestAPI.response_key, ) def __getitem__(self, key: int) -> Any: return self.get(user_id=key) def __delitem__(self, key: int) -> None: self.delete(user_id=key)
Update User API - ADD type hints - Remove unused imports
Update User API - ADD type hints - Remove unused imports
Python
apache-2.0
OpenMined/PySyft,OpenMined/PySyft,OpenMined/PySyft,OpenMined/PySyft
from typing import Any - from typing import Dict + from typing import Callable - - # third party - from pandas import DataFrame # syft relative from ...messages.user_messages import CreateUserMessage from ...messages.user_messages import DeleteUserMessage from ...messages.user_messages import GetUserMessage from ...messages.user_messages import GetUsersMessage from ...messages.user_messages import UpdateUserMessage from .request_api import GridRequestAPI class UserRequestAPI(GridRequestAPI): response_key = "user" - def __init__(self, send): + def __init__(self, send: Callable): super().__init__( create_msg=CreateUserMessage, get_msg=GetUserMessage, get_all_msg=GetUsersMessage, update_msg=UpdateUserMessage, delete_msg=DeleteUserMessage, send=send, response_key=UserRequestAPI.response_key, ) - def __getitem__(self, key): + def __getitem__(self, key: int) -> Any: return self.get(user_id=key) - def __delitem__(self, key): + def __delitem__(self, key: int) -> None: self.delete(user_id=key)
Update User API - ADD type hints - Remove unused imports
## Code Before: from typing import Any from typing import Dict # third party from pandas import DataFrame # syft relative from ...messages.user_messages import CreateUserMessage from ...messages.user_messages import DeleteUserMessage from ...messages.user_messages import GetUserMessage from ...messages.user_messages import GetUsersMessage from ...messages.user_messages import UpdateUserMessage from .request_api import GridRequestAPI class UserRequestAPI(GridRequestAPI): response_key = "user" def __init__(self, send): super().__init__( create_msg=CreateUserMessage, get_msg=GetUserMessage, get_all_msg=GetUsersMessage, update_msg=UpdateUserMessage, delete_msg=DeleteUserMessage, send=send, response_key=UserRequestAPI.response_key, ) def __getitem__(self, key): return self.get(user_id=key) def __delitem__(self, key): self.delete(user_id=key) ## Instruction: Update User API - ADD type hints - Remove unused imports ## Code After: from typing import Any from typing import Callable # syft relative from ...messages.user_messages import CreateUserMessage from ...messages.user_messages import DeleteUserMessage from ...messages.user_messages import GetUserMessage from ...messages.user_messages import GetUsersMessage from ...messages.user_messages import UpdateUserMessage from .request_api import GridRequestAPI class UserRequestAPI(GridRequestAPI): response_key = "user" def __init__(self, send: Callable): super().__init__( create_msg=CreateUserMessage, get_msg=GetUserMessage, get_all_msg=GetUsersMessage, update_msg=UpdateUserMessage, delete_msg=DeleteUserMessage, send=send, response_key=UserRequestAPI.response_key, ) def __getitem__(self, key: int) -> Any: return self.get(user_id=key) def __delitem__(self, key: int) -> None: self.delete(user_id=key)
// ... existing code ... from typing import Any from typing import Callable # syft relative // ... modified code ... response_key = "user" def __init__(self, send: Callable): super().__init__( create_msg=CreateUserMessage, ... ) def __getitem__(self, key: int) -> Any: return self.get(user_id=key) def __delitem__(self, key: int) -> None: self.delete(user_id=key) // ... rest of the code ...
ee9b6b1640745bb7b757f1ec8603b19d4f678fb8
core/observables/file.py
core/observables/file.py
from __future__ import unicode_literals from mongoengine import * from core.observables import Observable from core.observables import Hash class File(Observable): value = StringField(verbose_name="SHA256 hash") mime_type = StringField(verbose_name="MIME type") hashes = DictField(verbose_name="Hashes") body = ReferenceField("AttachedFile") filenames = ListField(StringField(), verbose_name="Filenames") DISPLAY_FIELDS = Observable.DISPLAY_FIELDS + [("mime_type", "MIME Type")] @staticmethod def check_type(txt): return True def info(self): i = Observable.info(self) i['mime_type'] = self.mime_type i['hashes'] = self.hashes return i
from __future__ import unicode_literals from flask import url_for from flask_mongoengine.wtf import model_form from mongoengine import * from core.observables import Observable from core.database import StringListField class File(Observable): value = StringField(verbose_name="Value") mime_type = StringField(verbose_name="MIME type") hashes = DictField(verbose_name="Hashes") body = ReferenceField("AttachedFile") filenames = ListField(StringField(), verbose_name="Filenames") DISPLAY_FIELDS = Observable.DISPLAY_FIELDS + [("mime_type", "MIME Type")] exclude_fields = Observable.exclude_fields + ['hashes', 'body'] @classmethod def get_form(klass): form = model_form(klass, exclude=klass.exclude_fields) form.filenames = StringListField("Filenames") return form @staticmethod def check_type(txt): return True def info(self): i = Observable.info(self) i['mime_type'] = self.mime_type i['hashes'] = self.hashes return i
Clean up File edit view
Clean up File edit view
Python
apache-2.0
yeti-platform/yeti,yeti-platform/yeti,yeti-platform/yeti,yeti-platform/yeti
from __future__ import unicode_literals + from flask import url_for + from flask_mongoengine.wtf import model_form from mongoengine import * + from core.observables import Observable - from core.observables import Hash + from core.database import StringListField class File(Observable): - value = StringField(verbose_name="SHA256 hash") + value = StringField(verbose_name="Value") - mime_type = StringField(verbose_name="MIME type") hashes = DictField(verbose_name="Hashes") body = ReferenceField("AttachedFile") filenames = ListField(StringField(), verbose_name="Filenames") DISPLAY_FIELDS = Observable.DISPLAY_FIELDS + [("mime_type", "MIME Type")] + exclude_fields = Observable.exclude_fields + ['hashes', 'body'] + + @classmethod + def get_form(klass): + form = model_form(klass, exclude=klass.exclude_fields) + form.filenames = StringListField("Filenames") + return form @staticmethod def check_type(txt): return True def info(self): i = Observable.info(self) i['mime_type'] = self.mime_type i['hashes'] = self.hashes return i
Clean up File edit view
## Code Before: from __future__ import unicode_literals from mongoengine import * from core.observables import Observable from core.observables import Hash class File(Observable): value = StringField(verbose_name="SHA256 hash") mime_type = StringField(verbose_name="MIME type") hashes = DictField(verbose_name="Hashes") body = ReferenceField("AttachedFile") filenames = ListField(StringField(), verbose_name="Filenames") DISPLAY_FIELDS = Observable.DISPLAY_FIELDS + [("mime_type", "MIME Type")] @staticmethod def check_type(txt): return True def info(self): i = Observable.info(self) i['mime_type'] = self.mime_type i['hashes'] = self.hashes return i ## Instruction: Clean up File edit view ## Code After: from __future__ import unicode_literals from flask import url_for from flask_mongoengine.wtf import model_form from mongoengine import * from core.observables import Observable from core.database import StringListField class File(Observable): value = StringField(verbose_name="Value") mime_type = StringField(verbose_name="MIME type") hashes = DictField(verbose_name="Hashes") body = ReferenceField("AttachedFile") filenames = ListField(StringField(), verbose_name="Filenames") DISPLAY_FIELDS = Observable.DISPLAY_FIELDS + [("mime_type", "MIME Type")] exclude_fields = Observable.exclude_fields + ['hashes', 'body'] @classmethod def get_form(klass): form = model_form(klass, exclude=klass.exclude_fields) form.filenames = StringListField("Filenames") return form @staticmethod def check_type(txt): return True def info(self): i = Observable.info(self) i['mime_type'] = self.mime_type i['hashes'] = self.hashes return i
# ... existing code ... from __future__ import unicode_literals from flask import url_for from flask_mongoengine.wtf import model_form from mongoengine import * from core.observables import Observable from core.database import StringListField # ... modified code ... class File(Observable): value = StringField(verbose_name="Value") mime_type = StringField(verbose_name="MIME type") hashes = DictField(verbose_name="Hashes") ... DISPLAY_FIELDS = Observable.DISPLAY_FIELDS + [("mime_type", "MIME Type")] exclude_fields = Observable.exclude_fields + ['hashes', 'body'] @classmethod def get_form(klass): form = model_form(klass, exclude=klass.exclude_fields) form.filenames = StringListField("Filenames") return form @staticmethod # ... rest of the code ...
d3a203725d13a7abef091f0070f90826d3225dbc
settings_travis.py
settings_travis.py
import ssl LDAP_SERVER = 'ldap.rserver.de' LDAP_PORT = 3389 LDAP_SSL_PORT = 6636 LDAP_REQUIRE_CERT = ssl.CERT_NONE
import ssl LDAP_SERVER = 'ldap.rserver.de' LDAP_PORT = 3389 LDAP_SSL_PORT = 6636 LDAP_REQUIRE_CERT = ssl.CERT_NONE LDAP_TLS_VERSION = ssl.PROTOCOL_TLSv1
Fix travis unit test for python 3.3
Fix travis unit test for python 3.3
Python
bsd-2-clause
rroemhild/flask-ldapconn
import ssl LDAP_SERVER = 'ldap.rserver.de' LDAP_PORT = 3389 LDAP_SSL_PORT = 6636 LDAP_REQUIRE_CERT = ssl.CERT_NONE + LDAP_TLS_VERSION = ssl.PROTOCOL_TLSv1
Fix travis unit test for python 3.3
## Code Before: import ssl LDAP_SERVER = 'ldap.rserver.de' LDAP_PORT = 3389 LDAP_SSL_PORT = 6636 LDAP_REQUIRE_CERT = ssl.CERT_NONE ## Instruction: Fix travis unit test for python 3.3 ## Code After: import ssl LDAP_SERVER = 'ldap.rserver.de' LDAP_PORT = 3389 LDAP_SSL_PORT = 6636 LDAP_REQUIRE_CERT = ssl.CERT_NONE LDAP_TLS_VERSION = ssl.PROTOCOL_TLSv1
# ... existing code ... LDAP_SSL_PORT = 6636 LDAP_REQUIRE_CERT = ssl.CERT_NONE LDAP_TLS_VERSION = ssl.PROTOCOL_TLSv1 # ... rest of the code ...
e880522e226b593be2985cdc85cb0ebd87e53a98
astral/models/tests/factories.py
astral/models/tests/factories.py
import factory import faker import random import uuid from astral.models.stream import Stream from astral.models.node import Node from astral.models.ticket import Ticket ELIXIR_CREATION = lambda class_to_create, **kwargs: class_to_create(**kwargs) factory.Factory.set_creation_function(ELIXIR_CREATION) class StreamFactory(factory.Factory): id = factory.Sequence(lambda n: int(n) + 1) name = factory.LazyAttribute(lambda a: ' '.join(faker.lorem.words())) class NodeFactory(factory.Factory): ip_address = factory.LazyAttribute(lambda a: faker.internet.ip_address()) uuid = factory.LazyAttribute(lambda a: random.randrange(1000, 1000000)) port = factory.LazyAttribute(lambda a: random.randrange(1000, 10000)) class ThisNodeFactory(factory.Factory): FACTORY_FOR = Node ip_address = '127.0.0.1' uuid = factory.LazyAttribute(lambda a: uuid.getnode()) port = factory.LazyAttribute(lambda a: random.randrange(1000, 10000)) class TicketFactory(factory.Factory): source = factory.LazyAttribute(lambda a: NodeFactory()) destination = factory.LazyAttribute(lambda a: NodeFactory()) stream = factory.LazyAttribute(lambda a: StreamFactory())
import factory import faker import random import uuid from astral.models.stream import Stream from astral.models.node import Node from astral.models.ticket import Ticket ELIXIR_CREATION = lambda class_to_create, **kwargs: class_to_create(**kwargs) factory.Factory.set_creation_function(ELIXIR_CREATION) class StreamFactory(factory.Factory): id = factory.Sequence(lambda n: int(n) + 1) name = factory.LazyAttribute(lambda a: ' '.join(faker.lorem.words())) source = factory.LazyAttribute(lambda a: NodeFactory()) class NodeFactory(factory.Factory): ip_address = factory.LazyAttribute(lambda a: faker.internet.ip_address()) uuid = factory.LazyAttribute(lambda a: random.randrange(1000, 1000000)) port = factory.LazyAttribute(lambda a: random.randrange(1000, 10000)) class ThisNodeFactory(factory.Factory): FACTORY_FOR = Node ip_address = '127.0.0.1' uuid = factory.LazyAttribute(lambda a: uuid.getnode()) port = factory.LazyAttribute(lambda a: random.randrange(1000, 10000)) class TicketFactory(factory.Factory): source = factory.LazyAttribute(lambda a: NodeFactory()) destination = factory.LazyAttribute(lambda a: NodeFactory()) stream = factory.LazyAttribute(lambda a: StreamFactory())
Make sure streams always have a source.
Make sure streams always have a source.
Python
mit
peplin/astral
import factory import faker import random import uuid from astral.models.stream import Stream from astral.models.node import Node from astral.models.ticket import Ticket ELIXIR_CREATION = lambda class_to_create, **kwargs: class_to_create(**kwargs) factory.Factory.set_creation_function(ELIXIR_CREATION) class StreamFactory(factory.Factory): id = factory.Sequence(lambda n: int(n) + 1) name = factory.LazyAttribute(lambda a: ' '.join(faker.lorem.words())) + source = factory.LazyAttribute(lambda a: NodeFactory()) class NodeFactory(factory.Factory): ip_address = factory.LazyAttribute(lambda a: faker.internet.ip_address()) uuid = factory.LazyAttribute(lambda a: random.randrange(1000, 1000000)) port = factory.LazyAttribute(lambda a: random.randrange(1000, 10000)) class ThisNodeFactory(factory.Factory): FACTORY_FOR = Node ip_address = '127.0.0.1' uuid = factory.LazyAttribute(lambda a: uuid.getnode()) port = factory.LazyAttribute(lambda a: random.randrange(1000, 10000)) class TicketFactory(factory.Factory): source = factory.LazyAttribute(lambda a: NodeFactory()) destination = factory.LazyAttribute(lambda a: NodeFactory()) stream = factory.LazyAttribute(lambda a: StreamFactory())
Make sure streams always have a source.
## Code Before: import factory import faker import random import uuid from astral.models.stream import Stream from astral.models.node import Node from astral.models.ticket import Ticket ELIXIR_CREATION = lambda class_to_create, **kwargs: class_to_create(**kwargs) factory.Factory.set_creation_function(ELIXIR_CREATION) class StreamFactory(factory.Factory): id = factory.Sequence(lambda n: int(n) + 1) name = factory.LazyAttribute(lambda a: ' '.join(faker.lorem.words())) class NodeFactory(factory.Factory): ip_address = factory.LazyAttribute(lambda a: faker.internet.ip_address()) uuid = factory.LazyAttribute(lambda a: random.randrange(1000, 1000000)) port = factory.LazyAttribute(lambda a: random.randrange(1000, 10000)) class ThisNodeFactory(factory.Factory): FACTORY_FOR = Node ip_address = '127.0.0.1' uuid = factory.LazyAttribute(lambda a: uuid.getnode()) port = factory.LazyAttribute(lambda a: random.randrange(1000, 10000)) class TicketFactory(factory.Factory): source = factory.LazyAttribute(lambda a: NodeFactory()) destination = factory.LazyAttribute(lambda a: NodeFactory()) stream = factory.LazyAttribute(lambda a: StreamFactory()) ## Instruction: Make sure streams always have a source. ## Code After: import factory import faker import random import uuid from astral.models.stream import Stream from astral.models.node import Node from astral.models.ticket import Ticket ELIXIR_CREATION = lambda class_to_create, **kwargs: class_to_create(**kwargs) factory.Factory.set_creation_function(ELIXIR_CREATION) class StreamFactory(factory.Factory): id = factory.Sequence(lambda n: int(n) + 1) name = factory.LazyAttribute(lambda a: ' '.join(faker.lorem.words())) source = factory.LazyAttribute(lambda a: NodeFactory()) class NodeFactory(factory.Factory): ip_address = factory.LazyAttribute(lambda a: faker.internet.ip_address()) uuid = factory.LazyAttribute(lambda a: random.randrange(1000, 1000000)) port = factory.LazyAttribute(lambda a: random.randrange(1000, 10000)) class ThisNodeFactory(factory.Factory): FACTORY_FOR = Node ip_address = '127.0.0.1' uuid = factory.LazyAttribute(lambda a: uuid.getnode()) port = factory.LazyAttribute(lambda a: random.randrange(1000, 10000)) class TicketFactory(factory.Factory): source = factory.LazyAttribute(lambda a: NodeFactory()) destination = factory.LazyAttribute(lambda a: NodeFactory()) stream = factory.LazyAttribute(lambda a: StreamFactory())
... id = factory.Sequence(lambda n: int(n) + 1) name = factory.LazyAttribute(lambda a: ' '.join(faker.lorem.words())) source = factory.LazyAttribute(lambda a: NodeFactory()) ...
a589aa63f250a347ab24b7309e65ef25c7281437
src/sentry/utils/imports.py
src/sentry/utils/imports.py
from __future__ import absolute_import import pkgutil import six class ModuleProxyCache(dict): def __missing__(self, key): if '.' not in key: return __import__(key) module_name, class_name = key.rsplit('.', 1) module = __import__(module_name, {}, {}, [class_name]) handler = getattr(module, class_name) # We cache a NoneType for missing imports to avoid repeated lookups self[key] = handler return handler _cache = ModuleProxyCache() def import_string(path): """ Path must be module.path.ClassName >>> cls = import_string('sentry.models.Group') """ result = _cache[path] return result def import_submodules(context, root_module, path): """ Import all submodules and register them in the ``context`` namespace. >>> import_submodules(locals(), __name__, __path__) """ for loader, module_name, is_pkg in pkgutil.walk_packages(path, root_module + '.'): module = loader.find_module(module_name).load_module(module_name) for k, v in six.iteritems(vars(module)): if not k.startswith('_'): context[k] = v context[module_name] = module
from __future__ import absolute_import import pkgutil import six class ModuleProxyCache(dict): def __missing__(self, key): if '.' not in key: return __import__(key) module_name, class_name = key.rsplit('.', 1) module = __import__(module_name, {}, {}, [class_name]) handler = getattr(module, class_name) # We cache a NoneType for missing imports to avoid repeated lookups self[key] = handler return handler _cache = ModuleProxyCache() def import_string(path): """ Path must be module.path.ClassName >>> cls = import_string('sentry.models.Group') """ result = _cache[path] return result def import_submodules(context, root_module, path): """ Import all submodules and register them in the ``context`` namespace. >>> import_submodules(locals(), __name__, __path__) """ for loader, module_name, is_pkg in pkgutil.walk_packages(path, root_module + '.'): # this causes a Runtime error with model conflicts # module = loader.find_module(module_name).load_module(module_name) module = __import__(module_name, globals(), locals(), ['__name__']) for k, v in six.iteritems(vars(module)): if not k.startswith('_'): context[k] = v context[module_name] = module
Correct import behavior to prevent Runtime error
Correct import behavior to prevent Runtime error
Python
bsd-3-clause
gencer/sentry,jean/sentry,fotinakis/sentry,gencer/sentry,looker/sentry,BuildingLink/sentry,gencer/sentry,jean/sentry,JackDanger/sentry,fotinakis/sentry,BuildingLink/sentry,beeftornado/sentry,zenefits/sentry,beeftornado/sentry,looker/sentry,ifduyue/sentry,ifduyue/sentry,JamesMura/sentry,ifduyue/sentry,JamesMura/sentry,BuildingLink/sentry,zenefits/sentry,fotinakis/sentry,ifduyue/sentry,zenefits/sentry,zenefits/sentry,beeftornado/sentry,BuildingLink/sentry,JamesMura/sentry,jean/sentry,gencer/sentry,ifduyue/sentry,jean/sentry,alexm92/sentry,mvaled/sentry,gencer/sentry,alexm92/sentry,looker/sentry,JackDanger/sentry,mvaled/sentry,mvaled/sentry,BuildingLink/sentry,JamesMura/sentry,JackDanger/sentry,fotinakis/sentry,looker/sentry,zenefits/sentry,mvaled/sentry,mvaled/sentry,jean/sentry,JamesMura/sentry,alexm92/sentry,looker/sentry,mvaled/sentry
from __future__ import absolute_import import pkgutil import six class ModuleProxyCache(dict): def __missing__(self, key): if '.' not in key: return __import__(key) module_name, class_name = key.rsplit('.', 1) module = __import__(module_name, {}, {}, [class_name]) handler = getattr(module, class_name) # We cache a NoneType for missing imports to avoid repeated lookups self[key] = handler return handler _cache = ModuleProxyCache() def import_string(path): """ Path must be module.path.ClassName >>> cls = import_string('sentry.models.Group') """ result = _cache[path] return result def import_submodules(context, root_module, path): """ Import all submodules and register them in the ``context`` namespace. >>> import_submodules(locals(), __name__, __path__) """ for loader, module_name, is_pkg in pkgutil.walk_packages(path, root_module + '.'): + # this causes a Runtime error with model conflicts - module = loader.find_module(module_name).load_module(module_name) + # module = loader.find_module(module_name).load_module(module_name) + module = __import__(module_name, globals(), locals(), ['__name__']) for k, v in six.iteritems(vars(module)): if not k.startswith('_'): context[k] = v context[module_name] = module
Correct import behavior to prevent Runtime error
## Code Before: from __future__ import absolute_import import pkgutil import six class ModuleProxyCache(dict): def __missing__(self, key): if '.' not in key: return __import__(key) module_name, class_name = key.rsplit('.', 1) module = __import__(module_name, {}, {}, [class_name]) handler = getattr(module, class_name) # We cache a NoneType for missing imports to avoid repeated lookups self[key] = handler return handler _cache = ModuleProxyCache() def import_string(path): """ Path must be module.path.ClassName >>> cls = import_string('sentry.models.Group') """ result = _cache[path] return result def import_submodules(context, root_module, path): """ Import all submodules and register them in the ``context`` namespace. >>> import_submodules(locals(), __name__, __path__) """ for loader, module_name, is_pkg in pkgutil.walk_packages(path, root_module + '.'): module = loader.find_module(module_name).load_module(module_name) for k, v in six.iteritems(vars(module)): if not k.startswith('_'): context[k] = v context[module_name] = module ## Instruction: Correct import behavior to prevent Runtime error ## Code After: from __future__ import absolute_import import pkgutil import six class ModuleProxyCache(dict): def __missing__(self, key): if '.' not in key: return __import__(key) module_name, class_name = key.rsplit('.', 1) module = __import__(module_name, {}, {}, [class_name]) handler = getattr(module, class_name) # We cache a NoneType for missing imports to avoid repeated lookups self[key] = handler return handler _cache = ModuleProxyCache() def import_string(path): """ Path must be module.path.ClassName >>> cls = import_string('sentry.models.Group') """ result = _cache[path] return result def import_submodules(context, root_module, path): """ Import all submodules and register them in the ``context`` namespace. >>> import_submodules(locals(), __name__, __path__) """ for loader, module_name, is_pkg in pkgutil.walk_packages(path, root_module + '.'): # this causes a Runtime error with model conflicts # module = loader.find_module(module_name).load_module(module_name) module = __import__(module_name, globals(), locals(), ['__name__']) for k, v in six.iteritems(vars(module)): if not k.startswith('_'): context[k] = v context[module_name] = module
# ... existing code ... """ for loader, module_name, is_pkg in pkgutil.walk_packages(path, root_module + '.'): # this causes a Runtime error with model conflicts # module = loader.find_module(module_name).load_module(module_name) module = __import__(module_name, globals(), locals(), ['__name__']) for k, v in six.iteritems(vars(module)): if not k.startswith('_'): # ... rest of the code ...
5a6b19f956dfde65a1d8316fd4bebe4697846e45
connman_dispatcher/detect.py
connman_dispatcher/detect.py
import glib import dbus from dbus.mainloop.glib import DBusGMainLoop from pyee import EventEmitter import logbook logger = logbook.Logger('connman-dispatcher') __all__ = ['detector'] def property_changed(_, message): if message.get_member() == "PropertyChanged": _, state = message.get_args_list() if state == 'online' and not detector.is_online: logger.info('network state change: online' ) detector.emit('up') detector.is_online = True elif state == 'idle': logger.info('network state change: offline' ) detector.emit('down') detector.is_online = False detector = EventEmitter() detector.is_online = is_online() DBusGMainLoop(set_as_default=True) bus = dbus.SystemBus() bus.add_match_string_non_blocking("interface='net.connman.Manager'") bus.add_message_filter(property_changed) manager = dbus.Interface(bus.get_object('net.connman', "/"), 'net.connman.Manager') def is_online(): properties = manager.GetProperties() if properties['State'] == 'online': return True return False def run(): mainloop = glib.MainLoop() mainloop.run() detector.run = run detector.is_online = is_online
import glib import dbus from dbus.mainloop.glib import DBusGMainLoop from pyee import EventEmitter import logbook logger = logbook.Logger('connman-dispatcher') __all__ = ['detector'] def property_changed(_, message): if message.get_member() == "PropertyChanged": _, state = message.get_args_list() if state == 'online' and detector.state == 'offline': logger.info('network state change: online' ) detector.emit('up') detector.state = 'online' elif state == 'idle': logger.info('network state change: offline' ) detector.emit('down') detector.state = 'online' detector = EventEmitter() DBusGMainLoop(set_as_default=True) bus = dbus.SystemBus() bus.add_match_string_non_blocking("interface='net.connman.Manager'") bus.add_message_filter(property_changed) manager = dbus.Interface(bus.get_object('net.connman', "/"), 'net.connman.Manager') def is_online(): properties = manager.GetProperties() if properties['State'] == 'online': return True return False def run(): mainloop = glib.MainLoop() mainloop.run() detector.run = run detector.is_online = is_online detector.state = 'online' if is_online() else 'offline'
Use .state instead of .is_online to keep internal state
Use .state instead of .is_online to keep internal state
Python
isc
a-sk/connman-dispatcher
import glib import dbus from dbus.mainloop.glib import DBusGMainLoop from pyee import EventEmitter import logbook logger = logbook.Logger('connman-dispatcher') __all__ = ['detector'] def property_changed(_, message): if message.get_member() == "PropertyChanged": _, state = message.get_args_list() - if state == 'online' and not detector.is_online: + if state == 'online' and detector.state == 'offline': logger.info('network state change: online' ) detector.emit('up') - detector.is_online = True + detector.state = 'online' elif state == 'idle': logger.info('network state change: offline' ) detector.emit('down') - detector.is_online = False + detector.state = 'online' detector = EventEmitter() - detector.is_online = is_online() DBusGMainLoop(set_as_default=True) bus = dbus.SystemBus() bus.add_match_string_non_blocking("interface='net.connman.Manager'") bus.add_message_filter(property_changed) manager = dbus.Interface(bus.get_object('net.connman', "/"), 'net.connman.Manager') def is_online(): properties = manager.GetProperties() if properties['State'] == 'online': return True return False def run(): mainloop = glib.MainLoop() mainloop.run() detector.run = run detector.is_online = is_online + detector.state = 'online' if is_online() else 'offline'
Use .state instead of .is_online to keep internal state
## Code Before: import glib import dbus from dbus.mainloop.glib import DBusGMainLoop from pyee import EventEmitter import logbook logger = logbook.Logger('connman-dispatcher') __all__ = ['detector'] def property_changed(_, message): if message.get_member() == "PropertyChanged": _, state = message.get_args_list() if state == 'online' and not detector.is_online: logger.info('network state change: online' ) detector.emit('up') detector.is_online = True elif state == 'idle': logger.info('network state change: offline' ) detector.emit('down') detector.is_online = False detector = EventEmitter() detector.is_online = is_online() DBusGMainLoop(set_as_default=True) bus = dbus.SystemBus() bus.add_match_string_non_blocking("interface='net.connman.Manager'") bus.add_message_filter(property_changed) manager = dbus.Interface(bus.get_object('net.connman', "/"), 'net.connman.Manager') def is_online(): properties = manager.GetProperties() if properties['State'] == 'online': return True return False def run(): mainloop = glib.MainLoop() mainloop.run() detector.run = run detector.is_online = is_online ## Instruction: Use .state instead of .is_online to keep internal state ## Code After: import glib import dbus from dbus.mainloop.glib import DBusGMainLoop from pyee import EventEmitter import logbook logger = logbook.Logger('connman-dispatcher') __all__ = ['detector'] def property_changed(_, message): if message.get_member() == "PropertyChanged": _, state = message.get_args_list() if state == 'online' and detector.state == 'offline': logger.info('network state change: online' ) detector.emit('up') detector.state = 'online' elif state == 'idle': logger.info('network state change: offline' ) detector.emit('down') detector.state = 'online' detector = EventEmitter() DBusGMainLoop(set_as_default=True) bus = dbus.SystemBus() bus.add_match_string_non_blocking("interface='net.connman.Manager'") bus.add_message_filter(property_changed) manager = dbus.Interface(bus.get_object('net.connman', "/"), 'net.connman.Manager') def is_online(): properties = manager.GetProperties() if properties['State'] == 'online': return True return False def run(): mainloop = glib.MainLoop() mainloop.run() detector.run = run detector.is_online = is_online detector.state = 'online' if is_online() else 'offline'
... if message.get_member() == "PropertyChanged": _, state = message.get_args_list() if state == 'online' and detector.state == 'offline': logger.info('network state change: online' ) detector.emit('up') detector.state = 'online' elif state == 'idle': logger.info('network state change: offline' ) detector.emit('down') detector.state = 'online' detector = EventEmitter() DBusGMainLoop(set_as_default=True) ... detector.run = run detector.is_online = is_online detector.state = 'online' if is_online() else 'offline' ...
3747158af790a38ccfce217426ee5261877e9f0e
project/api/management/commands/seed_database.py
project/api/management/commands/seed_database.py
from django.core.management.base import BaseCommand from api.factories import ( InternationalFactory, ) class Command(BaseCommand): help = "Command to seed database." def handle(self, *args, **options): InternationalFactory()
from django.core.management.base import BaseCommand class Command(BaseCommand): help = "Command to seed database." from api.factories import ( InternationalFactory, ) def handle(self, *args, **options): self.InternationalFactory()
Fix seeding in management command
Fix seeding in management command
Python
bsd-2-clause
barberscore/barberscore-api,barberscore/barberscore-api,barberscore/barberscore-api,barberscore/barberscore-api,dbinetti/barberscore-django,dbinetti/barberscore,dbinetti/barberscore-django,dbinetti/barberscore
from django.core.management.base import BaseCommand - - - from api.factories import ( - InternationalFactory, - ) class Command(BaseCommand): help = "Command to seed database." + from api.factories import ( + InternationalFactory, + ) + def handle(self, *args, **options): - InternationalFactory() + self.InternationalFactory()
Fix seeding in management command
## Code Before: from django.core.management.base import BaseCommand from api.factories import ( InternationalFactory, ) class Command(BaseCommand): help = "Command to seed database." def handle(self, *args, **options): InternationalFactory() ## Instruction: Fix seeding in management command ## Code After: from django.core.management.base import BaseCommand class Command(BaseCommand): help = "Command to seed database." from api.factories import ( InternationalFactory, ) def handle(self, *args, **options): self.InternationalFactory()
# ... existing code ... from django.core.management.base import BaseCommand # ... modified code ... help = "Command to seed database." from api.factories import ( InternationalFactory, ) def handle(self, *args, **options): self.InternationalFactory() # ... rest of the code ...
d883a0fd09a42ff84ebb2ccf331692167370444b
ESLog/esloghandler.py
ESLog/esloghandler.py
from datetime import datetime import logging import os import json import urllib.request import urllib.parse class ESLogHandler(logging.Handler): def __init__(self, url, index=None, doc_type="log", level=logging.NOTSET): logging.Handler.__init__(self, level=level) self.url = urllib.parse.urlparse(url) print(self.url) # end __init__ def emit(self, record): # Break the record down to a dictionary message = dict() message["timestamp"] = datetime.now().isoformat() message["level"] = record.levelname message["name"] = record.name message["lineno"] = record.lineno message["message"] = record.msg json_message = json.dumps(message) json_message_bytes = json_message.encode("utf8") urllib.request.urlopen(self.url, data=json_message_bytes) # end emit # end ESLogHandler
from datetime import datetime import logging import os import json import urllib.request import urllib.parse class ESLogHandler(logging.Handler): def __init__(self, url, index=None, doc_type="log", level=logging.NOTSET): logging.Handler.__init__(self, level=level) # Parse the url self.url = urllib.parse.urlparse(url) # If no scheme is given, set it to http if not self.url.scheme: self.url.scheme = "http" # If a scheme is given but it is not http, raise an exception elif self.url.scheme != "http": raise ValueError("Only HTTP is supported.") # If no port is given default to 9200 if not self.url.port: self.url.port = "9200" # If no path is given or it is only a / use thi index and doc_type to construct one. if not self.url.path or self.url.path == "/": # an index is mandatory for Elasticsearch, doc_type too but it defaults to log if not index: raise ValueError("Elasticsearch index cannot be ommitted.") else: self.url.path = os.path.join("/", index, doc_type) # end __init__ def emit(self, record): # Break the record down to a dictionary message = dict() message["timestamp"] = datetime.now().isoformat() message["level"] = record.levelname message["name"] = record.name message["lineno"] = record.lineno message["message"] = record.msg json_message = json.dumps(message) json_message_bytes = json_message.encode("utf8") urllib.request.urlopen(self.url, data=json_message_bytes) # end emit # end ESLogHandler
Revert "trying to simplefy __init__"
Revert "trying to simplefy __init__" This reverts commit f2e3887bcd53b1c35af2d9dfe2363ea9e2a407f5.
Python
mit
Rio/ESLog
from datetime import datetime import logging import os import json import urllib.request import urllib.parse class ESLogHandler(logging.Handler): def __init__(self, url, index=None, doc_type="log", level=logging.NOTSET): logging.Handler.__init__(self, level=level) - + + # Parse the url self.url = urllib.parse.urlparse(url) - - print(self.url) + + # If no scheme is given, set it to http + if not self.url.scheme: + self.url.scheme = "http" + + # If a scheme is given but it is not http, raise an exception + elif self.url.scheme != "http": + raise ValueError("Only HTTP is supported.") + + # If no port is given default to 9200 + if not self.url.port: + self.url.port = "9200" + + # If no path is given or it is only a / use thi index and doc_type to construct one. + if not self.url.path or self.url.path == "/": + # an index is mandatory for Elasticsearch, doc_type too but it defaults to log + if not index: + raise ValueError("Elasticsearch index cannot be ommitted.") + + else: + self.url.path = os.path.join("/", index, doc_type) # end __init__ def emit(self, record): # Break the record down to a dictionary message = dict() message["timestamp"] = datetime.now().isoformat() message["level"] = record.levelname message["name"] = record.name message["lineno"] = record.lineno message["message"] = record.msg json_message = json.dumps(message) json_message_bytes = json_message.encode("utf8") urllib.request.urlopen(self.url, data=json_message_bytes) # end emit # end ESLogHandler
Revert "trying to simplefy __init__"
## Code Before: from datetime import datetime import logging import os import json import urllib.request import urllib.parse class ESLogHandler(logging.Handler): def __init__(self, url, index=None, doc_type="log", level=logging.NOTSET): logging.Handler.__init__(self, level=level) self.url = urllib.parse.urlparse(url) print(self.url) # end __init__ def emit(self, record): # Break the record down to a dictionary message = dict() message["timestamp"] = datetime.now().isoformat() message["level"] = record.levelname message["name"] = record.name message["lineno"] = record.lineno message["message"] = record.msg json_message = json.dumps(message) json_message_bytes = json_message.encode("utf8") urllib.request.urlopen(self.url, data=json_message_bytes) # end emit # end ESLogHandler ## Instruction: Revert "trying to simplefy __init__" ## Code After: from datetime import datetime import logging import os import json import urllib.request import urllib.parse class ESLogHandler(logging.Handler): def __init__(self, url, index=None, doc_type="log", level=logging.NOTSET): logging.Handler.__init__(self, level=level) # Parse the url self.url = urllib.parse.urlparse(url) # If no scheme is given, set it to http if not self.url.scheme: self.url.scheme = "http" # If a scheme is given but it is not http, raise an exception elif self.url.scheme != "http": raise ValueError("Only HTTP is supported.") # If no port is given default to 9200 if not self.url.port: self.url.port = "9200" # If no path is given or it is only a / use thi index and doc_type to construct one. if not self.url.path or self.url.path == "/": # an index is mandatory for Elasticsearch, doc_type too but it defaults to log if not index: raise ValueError("Elasticsearch index cannot be ommitted.") else: self.url.path = os.path.join("/", index, doc_type) # end __init__ def emit(self, record): # Break the record down to a dictionary message = dict() message["timestamp"] = datetime.now().isoformat() message["level"] = record.levelname message["name"] = record.name message["lineno"] = record.lineno message["message"] = record.msg json_message = json.dumps(message) json_message_bytes = json_message.encode("utf8") urllib.request.urlopen(self.url, data=json_message_bytes) # end emit # end ESLogHandler
// ... existing code ... def __init__(self, url, index=None, doc_type="log", level=logging.NOTSET): logging.Handler.__init__(self, level=level) # Parse the url self.url = urllib.parse.urlparse(url) # If no scheme is given, set it to http if not self.url.scheme: self.url.scheme = "http" # If a scheme is given but it is not http, raise an exception elif self.url.scheme != "http": raise ValueError("Only HTTP is supported.") # If no port is given default to 9200 if not self.url.port: self.url.port = "9200" # If no path is given or it is only a / use thi index and doc_type to construct one. if not self.url.path or self.url.path == "/": # an index is mandatory for Elasticsearch, doc_type too but it defaults to log if not index: raise ValueError("Elasticsearch index cannot be ommitted.") else: self.url.path = os.path.join("/", index, doc_type) # end __init__ // ... rest of the code ...
191d73fb6d30b691da8d9c55bfd36f055aea19d5
backend/pokehelper.py
backend/pokehelper.py
import json import os class Pokehelper(object): def __init__(self): basepath = os.path.dirname(__file__) filepath = os.path.abspath(os.path.join(basepath, 'data/pokemon.json' )) with open(filepath) as pokejson: self.pokelist = json.load(pokejson) ### ### LIST STARTS AT 0, EVERY PokeNumber needs a -1 ### def get_pokefamily(self, pokemon_number): return self.pokelist[pokemon_number-1]['family'] def get_evolution_name(self, pokemon_number): # NOT a safe method to use, just for testing purposes return self.pokelist[pokemon_number]['name'] def get_pokename(self, poke_number): return self.pokelist[poke_number-1]['name'] def get_base_attack(self, poke_number): return self.pokelist[poke_number-1]['stats']['attack'] def get_base_defense(self, poke_number): return self.pokelist[poke_number-1]['stats']['defense']
import json import os emptymon = {'moves1': [], 'family': 1, 'name': 'not-in-database', 'moves2': [], 'type2': 'nil', 'id': -1, 'candy': -1, 'type1': 'nil', 'stats': {'stamina': -1, 'attack': -1, 'defense': -1}} class Pokehelper(object): def __init__(self): basepath = os.path.dirname(__file__) filepath = os.path.abspath(os.path.join(basepath, 'data/pokemon.json')) with open(filepath) as pokejson: self.pokelist = json.load(pokejson) ### # LIST STARTS AT 0, EVERY PokeNumber needs a -1 ### def get_pokefamily(self, poke_number): if (poke_number > 151): return emptymon['family'] return self.pokelist[poke_number-1]['family'] def get_evolution_name(self, poke_number): # NOT a safe method to use, just for testing purposes if (poke_number > 151): return emptymon['name'] return self.pokelist[poke_number]['name'] def get_pokename(self, poke_number): if (poke_number > 151): return emptymon['name'] return self.pokelist[poke_number-1]['name'] def get_base_attack(self, poke_number): if (poke_number > 151): return emptymon['stats']['attack'] return self.pokelist[poke_number-1]['stats']['attack'] def get_base_defense(self, poke_number): if (poke_number > 151): return emptymon['stats']['defense'] return self.pokelist[poke_number-1]['stats']['defense']
Add fallback if pokemon_id > 151
Add fallback if pokemon_id > 151
Python
mit
Phaetec/pogo-cruncher,Phaetec/pogo-cruncher,Phaetec/pogo-cruncher
import json import os + + emptymon = {'moves1': [], 'family': 1, 'name': 'not-in-database', 'moves2': [], 'type2': 'nil', 'id': -1, 'candy': -1, 'type1': 'nil', 'stats': {'stamina': -1, 'attack': -1, 'defense': -1}} + class Pokehelper(object): def __init__(self): basepath = os.path.dirname(__file__) - filepath = os.path.abspath(os.path.join(basepath, 'data/pokemon.json' )) + filepath = os.path.abspath(os.path.join(basepath, 'data/pokemon.json')) with open(filepath) as pokejson: self.pokelist = json.load(pokejson) - ### + ### - ### LIST STARTS AT 0, EVERY PokeNumber needs a -1 + # LIST STARTS AT 0, EVERY PokeNumber needs a -1 - ### + ### - def get_pokefamily(self, pokemon_number): + def get_pokefamily(self, poke_number): + if (poke_number > 151): + return emptymon['family'] - return self.pokelist[pokemon_number-1]['family'] + return self.pokelist[poke_number-1]['family'] - def get_evolution_name(self, pokemon_number): + def get_evolution_name(self, poke_number): # NOT a safe method to use, just for testing purposes + if (poke_number > 151): + return emptymon['name'] - return self.pokelist[pokemon_number]['name'] + return self.pokelist[poke_number]['name'] def get_pokename(self, poke_number): + if (poke_number > 151): + return emptymon['name'] return self.pokelist[poke_number-1]['name'] def get_base_attack(self, poke_number): + if (poke_number > 151): + return emptymon['stats']['attack'] return self.pokelist[poke_number-1]['stats']['attack'] def get_base_defense(self, poke_number): + if (poke_number > 151): + return emptymon['stats']['defense'] return self.pokelist[poke_number-1]['stats']['defense']
Add fallback if pokemon_id > 151
## Code Before: import json import os class Pokehelper(object): def __init__(self): basepath = os.path.dirname(__file__) filepath = os.path.abspath(os.path.join(basepath, 'data/pokemon.json' )) with open(filepath) as pokejson: self.pokelist = json.load(pokejson) ### ### LIST STARTS AT 0, EVERY PokeNumber needs a -1 ### def get_pokefamily(self, pokemon_number): return self.pokelist[pokemon_number-1]['family'] def get_evolution_name(self, pokemon_number): # NOT a safe method to use, just for testing purposes return self.pokelist[pokemon_number]['name'] def get_pokename(self, poke_number): return self.pokelist[poke_number-1]['name'] def get_base_attack(self, poke_number): return self.pokelist[poke_number-1]['stats']['attack'] def get_base_defense(self, poke_number): return self.pokelist[poke_number-1]['stats']['defense'] ## Instruction: Add fallback if pokemon_id > 151 ## Code After: import json import os emptymon = {'moves1': [], 'family': 1, 'name': 'not-in-database', 'moves2': [], 'type2': 'nil', 'id': -1, 'candy': -1, 'type1': 'nil', 'stats': {'stamina': -1, 'attack': -1, 'defense': -1}} class Pokehelper(object): def __init__(self): basepath = os.path.dirname(__file__) filepath = os.path.abspath(os.path.join(basepath, 'data/pokemon.json')) with open(filepath) as pokejson: self.pokelist = json.load(pokejson) ### # LIST STARTS AT 0, EVERY PokeNumber needs a -1 ### def get_pokefamily(self, poke_number): if (poke_number > 151): return emptymon['family'] return self.pokelist[poke_number-1]['family'] def get_evolution_name(self, poke_number): # NOT a safe method to use, just for testing purposes if (poke_number > 151): return emptymon['name'] return self.pokelist[poke_number]['name'] def get_pokename(self, poke_number): if (poke_number > 151): return emptymon['name'] return self.pokelist[poke_number-1]['name'] def get_base_attack(self, poke_number): if (poke_number > 151): return emptymon['stats']['attack'] return self.pokelist[poke_number-1]['stats']['attack'] def get_base_defense(self, poke_number): if (poke_number > 151): return emptymon['stats']['defense'] return self.pokelist[poke_number-1]['stats']['defense']
# ... existing code ... import json import os emptymon = {'moves1': [], 'family': 1, 'name': 'not-in-database', 'moves2': [], 'type2': 'nil', 'id': -1, 'candy': -1, 'type1': 'nil', 'stats': {'stamina': -1, 'attack': -1, 'defense': -1}} class Pokehelper(object): # ... modified code ... basepath = os.path.dirname(__file__) filepath = os.path.abspath(os.path.join(basepath, 'data/pokemon.json')) with open(filepath) as pokejson: self.pokelist = json.load(pokejson) ### # LIST STARTS AT 0, EVERY PokeNumber needs a -1 ### def get_pokefamily(self, poke_number): if (poke_number > 151): return emptymon['family'] return self.pokelist[poke_number-1]['family'] def get_evolution_name(self, poke_number): # NOT a safe method to use, just for testing purposes if (poke_number > 151): return emptymon['name'] return self.pokelist[poke_number]['name'] def get_pokename(self, poke_number): if (poke_number > 151): return emptymon['name'] return self.pokelist[poke_number-1]['name'] def get_base_attack(self, poke_number): if (poke_number > 151): return emptymon['stats']['attack'] return self.pokelist[poke_number-1]['stats']['attack'] def get_base_defense(self, poke_number): if (poke_number > 151): return emptymon['stats']['defense'] return self.pokelist[poke_number-1]['stats']['defense'] # ... rest of the code ...
bd7c5c5544a6d09062da05a4780524e8981f1737
captainhook/checkers/block_branches.py
captainhook/checkers/block_branches.py
import argparse from .utils import bash CHECK_NAME = 'block_branch' def run(files, temp_folder, arg=None): "Check we're not committing to a blocked branch" parser = get_parser() argos = parser.parse_args(arg.split()) current_branch = bash('git symbolic-ref HEAD').value().decode('utf-8') current_branch = current_branch.replace('refs/heads/', '').strip() if current_branch in argos.branches: return ("Branch '{0}' is blocked from being " "committed to.".format(current_branch)) def get_parser(): parser = argparse.ArgumentParser() parser.add_argument('branches', metavar='B', nargs='+', help='a branch to block commits to') return parser
import argparse from .utils import bash CHECK_NAME = 'block_branch' def run(files, temp_folder, arg=None): "Check we're not committing to a blocked branch" parser = get_parser() argos = parser.parse_args(arg.split()) current_branch = bash('git symbolic-ref HEAD').value() current_branch = current_branch.replace('refs/heads/', '').strip() if current_branch in argos.branches: return ("Branch '{0}' is blocked from being " "committed to.".format(current_branch)) def get_parser(): parser = argparse.ArgumentParser() parser.add_argument('branches', metavar='B', nargs='+', help='a branch to block commits to') return parser
Remove decode from block branches check
Remove decode from block branches check It’s now done by `bash()`.
Python
bsd-3-clause
alexcouper/captainhook
import argparse from .utils import bash CHECK_NAME = 'block_branch' def run(files, temp_folder, arg=None): "Check we're not committing to a blocked branch" parser = get_parser() argos = parser.parse_args(arg.split()) - current_branch = bash('git symbolic-ref HEAD').value().decode('utf-8') + current_branch = bash('git symbolic-ref HEAD').value() current_branch = current_branch.replace('refs/heads/', '').strip() if current_branch in argos.branches: return ("Branch '{0}' is blocked from being " "committed to.".format(current_branch)) def get_parser(): parser = argparse.ArgumentParser() parser.add_argument('branches', metavar='B', nargs='+', help='a branch to block commits to') return parser
Remove decode from block branches check
## Code Before: import argparse from .utils import bash CHECK_NAME = 'block_branch' def run(files, temp_folder, arg=None): "Check we're not committing to a blocked branch" parser = get_parser() argos = parser.parse_args(arg.split()) current_branch = bash('git symbolic-ref HEAD').value().decode('utf-8') current_branch = current_branch.replace('refs/heads/', '').strip() if current_branch in argos.branches: return ("Branch '{0}' is blocked from being " "committed to.".format(current_branch)) def get_parser(): parser = argparse.ArgumentParser() parser.add_argument('branches', metavar='B', nargs='+', help='a branch to block commits to') return parser ## Instruction: Remove decode from block branches check ## Code After: import argparse from .utils import bash CHECK_NAME = 'block_branch' def run(files, temp_folder, arg=None): "Check we're not committing to a blocked branch" parser = get_parser() argos = parser.parse_args(arg.split()) current_branch = bash('git symbolic-ref HEAD').value() current_branch = current_branch.replace('refs/heads/', '').strip() if current_branch in argos.branches: return ("Branch '{0}' is blocked from being " "committed to.".format(current_branch)) def get_parser(): parser = argparse.ArgumentParser() parser.add_argument('branches', metavar='B', nargs='+', help='a branch to block commits to') return parser
... argos = parser.parse_args(arg.split()) current_branch = bash('git symbolic-ref HEAD').value() current_branch = current_branch.replace('refs/heads/', '').strip() if current_branch in argos.branches: ...
e82225201772794bf347c6e768d25f24a61b9b54
migrations/schematic_settings.py
migrations/schematic_settings.py
import sys import os # This only works if you're running schematic from the zamboni root. sys.path.insert(0, os.path.realpath('.')) # Set up zamboni. import manage from django.conf import settings config = settings.DATABASES['default'] config['HOST'] = config.get('HOST', 'localhost') config['PORT'] = config.get('PORT', '3306') if config['HOST'].endswith('.sock'): """ Oh you meant 'localhost'! """ config['HOST'] = 'localhost' s = 'mysql --silent {NAME} -h{HOST} -P{PORT} -u{USER}' if config['PASSWORD']: s += ' -p{PASSWORD}' else: del config['PASSWORD'] db = s.format(**config) table = 'schema_version'
import sys import os sys.path.insert(0, os.path.dirname(os.path.dirname(os.path.abspath(__file__)))) # Set up zamboni. import manage from django.conf import settings config = settings.DATABASES['default'] config['HOST'] = config.get('HOST', 'localhost') config['PORT'] = config.get('PORT', '3306') if not config['HOST'] or config['HOST'].endswith('.sock'): """ Oh you meant 'localhost'! """ config['HOST'] = 'localhost' s = 'mysql --silent {NAME} -h{HOST} -u{USER}' if config['PASSWORD']: s += ' -p{PASSWORD}' else: del config['PASSWORD'] if config['PORT']: s += ' -P{PORT}' else: del config['PORT'] db = s.format(**config) table = 'schema_version'
Make the settings work when there's no port, and fix up the path manipulation
Make the settings work when there's no port, and fix up the path manipulation
Python
bsd-3-clause
kumar303/zamboni,kmaglione/olympia,Prashant-Surya/addons-server,jamesthechamp/zamboni,yfdyh000/olympia,aviarypl/mozilla-l10n-addons-server,Joergen/zamboni,muffinresearch/addons-server,Jobava/zamboni,koehlermichael/olympia,clouserw/zamboni,kmaglione/olympia,mstriemer/addons-server,Hitechverma/zamboni,mstriemer/olympia,psiinon/addons-server,mozilla/addons-server,lavish205/olympia,Nolski/olympia,spasovski/zamboni,jasonthomas/zamboni,jpetto/olympia,kumar303/olympia,beni55/olympia,crdoconnor/olympia,diox/zamboni,harry-7/addons-server,andymckay/addons-server,clouserw/zamboni,andymckay/olympia,andymckay/zamboni,Nolski/olympia,kmaglione/olympia,SuriyaaKudoIsc/olympia,magopian/olympia,psiinon/addons-server,kumar303/zamboni,mstriemer/zamboni,Joergen/olympia,shahbaz17/zamboni,ayushagrawal288/zamboni,Jobava/zamboni,jbalogh/zamboni,Hitechverma/zamboni,magopian/olympia,atiqueahmedziad/addons-server,Joergen/olympia,tsl143/zamboni,elysium001/zamboni,eviljeff/zamboni,andymckay/olympia,wagnerand/olympia,mozilla/zamboni,johancz/olympia,shahbaz17/zamboni,kmaglione/olympia,SuriyaaKudoIsc/olympia,atiqueahmedziad/addons-server,luckylavish/zamboni,bqbn/addons-server,johancz/olympia,psiinon/addons-server,jamesthechamp/zamboni,robhudson/zamboni,mozilla/olympia,diox/zamboni,anaran/olympia,crdoconnor/olympia,andymckay/zamboni,mudithkr/zamboni,harry-7/addons-server,mdaif/olympia,shahbaz17/zamboni,mozilla/addons-server,atiqueahmedziad/addons-server,mozilla/olympia,Witia1/olympia,ingenioustechie/zamboni,wagnerand/olympia,spasovski/zamboni,Nolski/olympia,robhudson/zamboni,kumar303/addons-server,koehlermichael/olympia,lavish205/olympia,luckylavish/zamboni,muffinresearch/olympia,jasonthomas/zamboni,clouserw/zamboni,ddurst/zamboni,mstriemer/addons-server,diox/zamboni,elysium001/zamboni,shahbaz17/zamboni,ayushagrawal288/zamboni,koehlermichael/olympia,eviljeff/zamboni,ngokevin/zamboni,beni55/olympia,SuriyaaKudoIsc/olympia,atiqueahmedziad/addons-server,kumar303/olympia,Prashant-Surya/addons-server,ngokevin/zamboni,wagnerand/zamboni,robhudson/zamboni,andymckay/addons-server,eviljeff/zamboni,bqbn/addons-server,mrrrgn/olympia,jamesthechamp/zamboni,johancz/olympia,crdoconnor/olympia,crdoconnor/olympia,Nolski/olympia,luckylavish/zamboni,muffinresearch/olympia,robhudson/zamboni,ingenioustechie/zamboni,jpetto/olympia,Jobava/zamboni,muffinresearch/addons-server,jasonthomas/zamboni,spasovski/zamboni,Prashant-Surya/addons-server,mstriemer/zamboni,washort/zamboni,Joergen/zamboni,Prashant-Surya/addons-server,Joergen/zamboni,wagnerand/addons-server,mozilla/olympia,eviljeff/olympia,mozilla/zamboni,Witia1/olympia,tsl143/addons-server,eviljeff/olympia,jbalogh/zamboni,kumar303/addons-server,andymckay/addons-server,Witia1/olympia,wagnerand/zamboni,magopian/olympia,Joergen/zamboni,jbalogh/zamboni,mstriemer/zamboni,washort/zamboni,jpetto/olympia,andymckay/olympia,ddurst/zamboni,elysium001/zamboni,Revanth47/addons-server,aviarypl/mozilla-l10n-addons-server,spasovski/zamboni,wagnerand/zamboni,kumar303/zamboni,diox/olympia,ayushagrawal288/zamboni,lavish205/olympia,mozilla/olympia,mstriemer/olympia,yfdyh000/olympia,harikishen/addons-server,andymckay/olympia,washort/zamboni,Nolski/olympia,ddurst/zamboni,Joergen/olympia,tsl143/zamboni,harikishen/addons-server,wagnerand/olympia,muffinresearch/addons-server,tsl143/addons-server,diox/olympia,washort/zamboni,clouserw/zamboni,luckylavish/zamboni,mstriemer/addons-server,anaran/olympia,beni55/olympia,jasonthomas/zamboni,mudithkr/zamboni,mrrrgn/olympia,crdoconnor/olympia,mrrrgn/olympia,kumar303/addons-server,tsl143/addons-server,magopian/olympia,andymckay/zamboni,johancz/olympia,lavish205/olympia,tsl143/addons-server,ngokevin/zamboni,mozilla/addons-server,aviarypl/mozilla-l10n-addons-server,mozilla/zamboni,Witia1/olympia,Jobava/zamboni,beni55/olympia,muffinresearch/olympia,mrrrgn/olympia,mstriemer/olympia,Joergen/olympia,muffinresearch/addons-server,wagnerand/addons-server,Revanth47/addons-server,SuriyaaKudoIsc/olympia,beni55/olympia,diox/zamboni,Witia1/olympia,andymckay/addons-server,mstriemer/zamboni,mozilla/zamboni,mdaif/olympia,Hitechverma/zamboni,Revanth47/addons-server,diox/olympia,mudithkr/zamboni,ddurst/zamboni,anaran/olympia,ingenioustechie/zamboni,tsl143/zamboni,Joergen/zamboni,tsl143/zamboni,diox/olympia,yfdyh000/olympia,magopian/olympia,mdaif/olympia,kumar303/zamboni,jpetto/olympia,Joergen/zamboni,yfdyh000/olympia,ayushagrawal288/zamboni,kumar303/olympia,wagnerand/addons-server,mdaif/olympia,muffinresearch/olympia,harry-7/addons-server,kmaglione/olympia,elysium001/zamboni,jamesthechamp/zamboni,koehlermichael/olympia,psiinon/addons-server,johancz/olympia,yfdyh000/olympia,mudithkr/zamboni,eviljeff/olympia,mrrrgn/olympia,ingenioustechie/zamboni,anaran/olympia,wagnerand/olympia,bqbn/addons-server,wagnerand/addons-server,kumar303/olympia,eviljeff/zamboni,aviarypl/mozilla-l10n-addons-server,Revanth47/addons-server,harry-7/addons-server,mstriemer/addons-server,wagnerand/zamboni,harikishen/addons-server,mozilla/addons-server,Joergen/olympia,kumar303/addons-server,eviljeff/olympia,jbalogh/zamboni,Hitechverma/zamboni,mdaif/olympia,mstriemer/olympia,harikishen/addons-server,koehlermichael/olympia,muffinresearch/olympia,bqbn/addons-server,muffinresearch/addons-server
import sys import os + sys.path.insert(0, os.path.dirname(os.path.dirname(os.path.abspath(__file__)))) - # This only works if you're running schematic from the zamboni root. - sys.path.insert(0, os.path.realpath('.')) # Set up zamboni. import manage from django.conf import settings config = settings.DATABASES['default'] config['HOST'] = config.get('HOST', 'localhost') config['PORT'] = config.get('PORT', '3306') - if config['HOST'].endswith('.sock'): + if not config['HOST'] or config['HOST'].endswith('.sock'): """ Oh you meant 'localhost'! """ config['HOST'] = 'localhost' - s = 'mysql --silent {NAME} -h{HOST} -P{PORT} -u{USER}' + s = 'mysql --silent {NAME} -h{HOST} -u{USER}' if config['PASSWORD']: s += ' -p{PASSWORD}' else: del config['PASSWORD'] + if config['PORT']: + s += ' -P{PORT}' + else: + del config['PORT'] db = s.format(**config) table = 'schema_version'
Make the settings work when there's no port, and fix up the path manipulation
## Code Before: import sys import os # This only works if you're running schematic from the zamboni root. sys.path.insert(0, os.path.realpath('.')) # Set up zamboni. import manage from django.conf import settings config = settings.DATABASES['default'] config['HOST'] = config.get('HOST', 'localhost') config['PORT'] = config.get('PORT', '3306') if config['HOST'].endswith('.sock'): """ Oh you meant 'localhost'! """ config['HOST'] = 'localhost' s = 'mysql --silent {NAME} -h{HOST} -P{PORT} -u{USER}' if config['PASSWORD']: s += ' -p{PASSWORD}' else: del config['PASSWORD'] db = s.format(**config) table = 'schema_version' ## Instruction: Make the settings work when there's no port, and fix up the path manipulation ## Code After: import sys import os sys.path.insert(0, os.path.dirname(os.path.dirname(os.path.abspath(__file__)))) # Set up zamboni. import manage from django.conf import settings config = settings.DATABASES['default'] config['HOST'] = config.get('HOST', 'localhost') config['PORT'] = config.get('PORT', '3306') if not config['HOST'] or config['HOST'].endswith('.sock'): """ Oh you meant 'localhost'! """ config['HOST'] = 'localhost' s = 'mysql --silent {NAME} -h{HOST} -u{USER}' if config['PASSWORD']: s += ' -p{PASSWORD}' else: del config['PASSWORD'] if config['PORT']: s += ' -P{PORT}' else: del config['PORT'] db = s.format(**config) table = 'schema_version'
... import os sys.path.insert(0, os.path.dirname(os.path.dirname(os.path.abspath(__file__)))) # Set up zamboni. ... config['PORT'] = config.get('PORT', '3306') if not config['HOST'] or config['HOST'].endswith('.sock'): """ Oh you meant 'localhost'! """ config['HOST'] = 'localhost' s = 'mysql --silent {NAME} -h{HOST} -u{USER}' if config['PASSWORD']: ... else: del config['PASSWORD'] if config['PORT']: s += ' -P{PORT}' else: del config['PORT'] db = s.format(**config) ...
79c9ee6107b841986054915c23f8456c80097c5b
osgtest/tests/test_13_gridftp.py
osgtest/tests/test_13_gridftp.py
import os import osgtest.library.core as core import unittest class TestStartGridFTP(unittest.TestCase): def test_01_start_gridftp(self): core.config['gridftp.pid-file'] = '/var/run/globus-gridftp-server.pid' core.state['gridftp.started-server'] = False if not core.rpm_is_installed('globus-gridftp-server-progs'): core.skip('not installed') return if os.path.exists(core.config['gridftp.pid-file']): core.skip('apparently running') return command = ('service', 'globus-gridftp-server', 'start') stdout, _, fail = core.check_system(command, 'Start GridFTP server') self.assert_(stdout.find('FAILED') == -1, fail) self.assert_(os.path.exists(core.config['gridftp.pid-file']), 'GridFTP server PID file missing') core.state['gridftp.started-server'] = True
import os from osgtest.library import core, osgunittest import unittest class TestStartGridFTP(osgunittest.OSGTestCase): def test_01_start_gridftp(self): core.config['gridftp.pid-file'] = '/var/run/globus-gridftp-server.pid' core.state['gridftp.started-server'] = False core.skip_ok_unless_installed('globus-gridftp-server-progs') self.skip_ok_if(os.path.exists(core.config['gridftp.pid-file']), 'already running') command = ('service', 'globus-gridftp-server', 'start') stdout, _, fail = core.check_system(command, 'Start GridFTP server') self.assert_(stdout.find('FAILED') == -1, fail) self.assert_(os.path.exists(core.config['gridftp.pid-file']), 'GridFTP server PID file missing') core.state['gridftp.started-server'] = True
Update 13_gridftp to use OkSkip functionality
Update 13_gridftp to use OkSkip functionality git-svn-id: 884a03e47e2adb735d896e55bb5ad6bc3421ba19@16527 4e558342-562e-0410-864c-e07659590f8c
Python
apache-2.0
efajardo/osg-test,efajardo/osg-test
import os - import osgtest.library.core as core + from osgtest.library import core, osgunittest import unittest - class TestStartGridFTP(unittest.TestCase): + class TestStartGridFTP(osgunittest.OSGTestCase): def test_01_start_gridftp(self): core.config['gridftp.pid-file'] = '/var/run/globus-gridftp-server.pid' core.state['gridftp.started-server'] = False - if not core.rpm_is_installed('globus-gridftp-server-progs'): + core.skip_ok_unless_installed('globus-gridftp-server-progs') - core.skip('not installed') - return - if os.path.exists(core.config['gridftp.pid-file']): + self.skip_ok_if(os.path.exists(core.config['gridftp.pid-file']), 'already running') - core.skip('apparently running') - return command = ('service', 'globus-gridftp-server', 'start') stdout, _, fail = core.check_system(command, 'Start GridFTP server') self.assert_(stdout.find('FAILED') == -1, fail) self.assert_(os.path.exists(core.config['gridftp.pid-file']), 'GridFTP server PID file missing') core.state['gridftp.started-server'] = True
Update 13_gridftp to use OkSkip functionality
## Code Before: import os import osgtest.library.core as core import unittest class TestStartGridFTP(unittest.TestCase): def test_01_start_gridftp(self): core.config['gridftp.pid-file'] = '/var/run/globus-gridftp-server.pid' core.state['gridftp.started-server'] = False if not core.rpm_is_installed('globus-gridftp-server-progs'): core.skip('not installed') return if os.path.exists(core.config['gridftp.pid-file']): core.skip('apparently running') return command = ('service', 'globus-gridftp-server', 'start') stdout, _, fail = core.check_system(command, 'Start GridFTP server') self.assert_(stdout.find('FAILED') == -1, fail) self.assert_(os.path.exists(core.config['gridftp.pid-file']), 'GridFTP server PID file missing') core.state['gridftp.started-server'] = True ## Instruction: Update 13_gridftp to use OkSkip functionality ## Code After: import os from osgtest.library import core, osgunittest import unittest class TestStartGridFTP(osgunittest.OSGTestCase): def test_01_start_gridftp(self): core.config['gridftp.pid-file'] = '/var/run/globus-gridftp-server.pid' core.state['gridftp.started-server'] = False core.skip_ok_unless_installed('globus-gridftp-server-progs') self.skip_ok_if(os.path.exists(core.config['gridftp.pid-file']), 'already running') command = ('service', 'globus-gridftp-server', 'start') stdout, _, fail = core.check_system(command, 'Start GridFTP server') self.assert_(stdout.find('FAILED') == -1, fail) self.assert_(os.path.exists(core.config['gridftp.pid-file']), 'GridFTP server PID file missing') core.state['gridftp.started-server'] = True
... import os from osgtest.library import core, osgunittest import unittest class TestStartGridFTP(osgunittest.OSGTestCase): def test_01_start_gridftp(self): ... core.state['gridftp.started-server'] = False core.skip_ok_unless_installed('globus-gridftp-server-progs') self.skip_ok_if(os.path.exists(core.config['gridftp.pid-file']), 'already running') command = ('service', 'globus-gridftp-server', 'start') ...
0b28fe44514969470db926c6f38615a8a5478bf6
smoke_signal/__init__.py
smoke_signal/__init__.py
from flask import Flask, g from .main.views import main from .nojs.views import nojs from sqlalchemy import create_engine from smoke_signal.database.models import Base from sqlalchemy.orm import sessionmaker app = Flask(__name__, instance_relative_config=True) app.config.from_object("config") app.config.from_pyfile("config.py") app.register_blueprint(main) app.register_blueprint(nojs) @app.before_request def init_db(): engine = create_engine(app.config["DATABASE_PATH"]) if not engine.dialect.has_table(engine.connect(), "feed"): Base.metadata.create_all(engine) Session = sessionmaker(bind=engine) g.db = Session() @app.teardown_appcontext def shutdown_session(exception=None): db = getattr(g, 'db', None) if db is not None: g.db.close()
from flask import Flask, g from .main.views import main from sqlalchemy import create_engine from smoke_signal.database.models import Base from sqlalchemy.orm import sessionmaker app = Flask(__name__, instance_relative_config=True) app.config.from_object("config") app.config.from_pyfile("config.py") app.register_blueprint(main) @app.before_request def init_db(): engine = create_engine(app.config["DATABASE_PATH"]) if not engine.dialect.has_table(engine.connect(), "feed"): Base.metadata.create_all(engine) Session = sessionmaker(bind=engine) g.db = Session() @app.teardown_appcontext def shutdown_session(exception=None): db = getattr(g, 'db', None) if db is not None: g.db.close()
Remove the no-JS version from the app
Remove the no-JS version from the app I haven't looked into it for a long while.
Python
mit
flacerdk/smoke-signal,flacerdk/smoke-signal,flacerdk/smoke-signal
from flask import Flask, g from .main.views import main - from .nojs.views import nojs from sqlalchemy import create_engine from smoke_signal.database.models import Base from sqlalchemy.orm import sessionmaker app = Flask(__name__, instance_relative_config=True) app.config.from_object("config") app.config.from_pyfile("config.py") app.register_blueprint(main) - app.register_blueprint(nojs) @app.before_request def init_db(): engine = create_engine(app.config["DATABASE_PATH"]) if not engine.dialect.has_table(engine.connect(), "feed"): Base.metadata.create_all(engine) Session = sessionmaker(bind=engine) g.db = Session() @app.teardown_appcontext def shutdown_session(exception=None): db = getattr(g, 'db', None) if db is not None: g.db.close()
Remove the no-JS version from the app
## Code Before: from flask import Flask, g from .main.views import main from .nojs.views import nojs from sqlalchemy import create_engine from smoke_signal.database.models import Base from sqlalchemy.orm import sessionmaker app = Flask(__name__, instance_relative_config=True) app.config.from_object("config") app.config.from_pyfile("config.py") app.register_blueprint(main) app.register_blueprint(nojs) @app.before_request def init_db(): engine = create_engine(app.config["DATABASE_PATH"]) if not engine.dialect.has_table(engine.connect(), "feed"): Base.metadata.create_all(engine) Session = sessionmaker(bind=engine) g.db = Session() @app.teardown_appcontext def shutdown_session(exception=None): db = getattr(g, 'db', None) if db is not None: g.db.close() ## Instruction: Remove the no-JS version from the app ## Code After: from flask import Flask, g from .main.views import main from sqlalchemy import create_engine from smoke_signal.database.models import Base from sqlalchemy.orm import sessionmaker app = Flask(__name__, instance_relative_config=True) app.config.from_object("config") app.config.from_pyfile("config.py") app.register_blueprint(main) @app.before_request def init_db(): engine = create_engine(app.config["DATABASE_PATH"]) if not engine.dialect.has_table(engine.connect(), "feed"): Base.metadata.create_all(engine) Session = sessionmaker(bind=engine) g.db = Session() @app.teardown_appcontext def shutdown_session(exception=None): db = getattr(g, 'db', None) if db is not None: g.db.close()
// ... existing code ... from flask import Flask, g from .main.views import main from sqlalchemy import create_engine from smoke_signal.database.models import Base // ... modified code ... app.config.from_pyfile("config.py") app.register_blueprint(main) // ... rest of the code ...
3039b00e761f02eb0586dad51049377a31329491
reggae/reflect.py
reggae/reflect.py
from __future__ import (unicode_literals, division, absolute_import, print_function) from reggae.build import Build, DefaultOptions from inspect import getmembers def get_build(module): builds = [v for n, v in getmembers(module) if isinstance(v, Build)] assert len(builds) == 1 return builds[0] def get_default_options(module): opts = [v for n, v in getmembers(module) if isinstance(v, DefaultOptions)] assert len(opts) == 1 or len(opts) == 0 return opts[0] if len(opts) else None def get_dependencies(module): from modulefinder import ModuleFinder import os finder = ModuleFinder() finder.run_script(module) all_module_paths = [m.__file__ for m in finder.modules.values()] def is_in_same_path(p): return p and os.path.dirname(p).startswith(os.path.dirname(module)) return [x for x in all_module_paths if is_in_same_path(x) and x != module]
from __future__ import (unicode_literals, division, absolute_import, print_function) from reggae.build import Build, DefaultOptions from inspect import getmembers def get_build(module): builds = [v for n, v in getmembers(module) if isinstance(v, Build)] assert len(builds) == 1 return builds[0] def get_default_options(module): opts = [v for n, v in getmembers(module) if isinstance(v, DefaultOptions)] assert len(opts) == 1 or len(opts) == 0 return opts[0] if len(opts) else None def get_dependencies(module): from modulefinder import ModuleFinder import os finder = ModuleFinder() finder.run_script(module) all_module_paths = [os.path.abspath(m.__file__) for m in finder.modules.values() if m.__file__ is not None] def is_in_same_path(p): return p and os.path.dirname(p).startswith(os.path.dirname(module)) return [x for x in all_module_paths if is_in_same_path(x) and x != module]
Use absolute paths for dependencies
Use absolute paths for dependencies
Python
bsd-3-clause
atilaneves/reggae-python
from __future__ import (unicode_literals, division, absolute_import, print_function) from reggae.build import Build, DefaultOptions from inspect import getmembers def get_build(module): builds = [v for n, v in getmembers(module) if isinstance(v, Build)] assert len(builds) == 1 return builds[0] def get_default_options(module): opts = [v for n, v in getmembers(module) if isinstance(v, DefaultOptions)] assert len(opts) == 1 or len(opts) == 0 return opts[0] if len(opts) else None def get_dependencies(module): from modulefinder import ModuleFinder import os finder = ModuleFinder() finder.run_script(module) - all_module_paths = [m.__file__ for m in finder.modules.values()] + all_module_paths = [os.path.abspath(m.__file__) for + m in finder.modules.values() if m.__file__ is not None] def is_in_same_path(p): return p and os.path.dirname(p).startswith(os.path.dirname(module)) return [x for x in all_module_paths if is_in_same_path(x) and x != module]
Use absolute paths for dependencies
## Code Before: from __future__ import (unicode_literals, division, absolute_import, print_function) from reggae.build import Build, DefaultOptions from inspect import getmembers def get_build(module): builds = [v for n, v in getmembers(module) if isinstance(v, Build)] assert len(builds) == 1 return builds[0] def get_default_options(module): opts = [v for n, v in getmembers(module) if isinstance(v, DefaultOptions)] assert len(opts) == 1 or len(opts) == 0 return opts[0] if len(opts) else None def get_dependencies(module): from modulefinder import ModuleFinder import os finder = ModuleFinder() finder.run_script(module) all_module_paths = [m.__file__ for m in finder.modules.values()] def is_in_same_path(p): return p and os.path.dirname(p).startswith(os.path.dirname(module)) return [x for x in all_module_paths if is_in_same_path(x) and x != module] ## Instruction: Use absolute paths for dependencies ## Code After: from __future__ import (unicode_literals, division, absolute_import, print_function) from reggae.build import Build, DefaultOptions from inspect import getmembers def get_build(module): builds = [v for n, v in getmembers(module) if isinstance(v, Build)] assert len(builds) == 1 return builds[0] def get_default_options(module): opts = [v for n, v in getmembers(module) if isinstance(v, DefaultOptions)] assert len(opts) == 1 or len(opts) == 0 return opts[0] if len(opts) else None def get_dependencies(module): from modulefinder import ModuleFinder import os finder = ModuleFinder() finder.run_script(module) all_module_paths = [os.path.abspath(m.__file__) for m in finder.modules.values() if m.__file__ is not None] def is_in_same_path(p): return p and os.path.dirname(p).startswith(os.path.dirname(module)) return [x for x in all_module_paths if is_in_same_path(x) and x != module]
... finder = ModuleFinder() finder.run_script(module) all_module_paths = [os.path.abspath(m.__file__) for m in finder.modules.values() if m.__file__ is not None] def is_in_same_path(p): ...
194a84b4559449f0b0e3e9cc9e7026392822c0af
questions/urls.py
questions/urls.py
from django.conf.urls.defaults import * from spenglr.education.models import * # Uncomment the next two lines to enable the admin: from django.contrib import admin admin.autodiscover() urlpatterns = patterns('', (r'^m/(?P<object_id>\d+)$', 'django.views.generic.list_detail.object_detail', { 'queryset': Module.objects.all(), 'template_name': 'education/module_questions.html' } ), )
from django.conf.urls.defaults import * from spenglr.education.models import * # Uncomment the next two lines to enable the admin: from django.contrib import admin admin.autodiscover() urlpatterns = patterns('', (r'^m/(?P<object_id>\d+)$', 'django.views.generic.list_detail.object_detail', { 'queryset': Module.objects.all(), 'template_name': 'questions/question_list.html' } ), )
Change to questions template location.
Change to questions template location.
Python
bsd-3-clause
mfitzp/smrtr,mfitzp/smrtr
from django.conf.urls.defaults import * from spenglr.education.models import * # Uncomment the next two lines to enable the admin: from django.contrib import admin admin.autodiscover() urlpatterns = patterns('', - (r'^m/(?P<object_id>\d+)$', 'django.views.generic.list_detail.object_detail', { 'queryset': Module.objects.all(), 'template_name': 'education/module_questions.html' } ), + (r'^m/(?P<object_id>\d+)$', 'django.views.generic.list_detail.object_detail', { 'queryset': Module.objects.all(), 'template_name': 'questions/question_list.html' } ), )
Change to questions template location.
## Code Before: from django.conf.urls.defaults import * from spenglr.education.models import * # Uncomment the next two lines to enable the admin: from django.contrib import admin admin.autodiscover() urlpatterns = patterns('', (r'^m/(?P<object_id>\d+)$', 'django.views.generic.list_detail.object_detail', { 'queryset': Module.objects.all(), 'template_name': 'education/module_questions.html' } ), ) ## Instruction: Change to questions template location. ## Code After: from django.conf.urls.defaults import * from spenglr.education.models import * # Uncomment the next two lines to enable the admin: from django.contrib import admin admin.autodiscover() urlpatterns = patterns('', (r'^m/(?P<object_id>\d+)$', 'django.views.generic.list_detail.object_detail', { 'queryset': Module.objects.all(), 'template_name': 'questions/question_list.html' } ), )
... urlpatterns = patterns('', (r'^m/(?P<object_id>\d+)$', 'django.views.generic.list_detail.object_detail', { 'queryset': Module.objects.all(), 'template_name': 'questions/question_list.html' } ), ) ...
56e764835e75035452a6a1ea06c386ec61dbe872
src/rinoh/stylesheets/__init__.py
src/rinoh/stylesheets/__init__.py
import inspect import os import sys from .. import DATA_PATH from ..style import StyleSheetFile from .matcher import matcher __all__ = ['matcher', 'sphinx', 'sphinx_base14'] STYLESHEETS_PATH = os.path.join(DATA_PATH, 'stylesheets') def path(filename): return os.path.join(STYLESHEETS_PATH, filename) sphinx = StyleSheetFile(path('sphinx.rts')) sphinx_article = StyleSheetFile(path('sphinx_article.rts')) sphinx_base14 = StyleSheetFile(path('base14.rts')) # generate docstrings for the StyleSheet instances for name, stylesheet in inspect.getmembers(sys.modules[__name__]): if not isinstance(stylesheet, StyleSheetFile): continue stylesheet.__doc__ = (':entry point name: ``{}``\n\n{}' .format(stylesheet, stylesheet.description))
import inspect import os import sys from .. import DATA_PATH from ..style import StyleSheetFile from .matcher import matcher __all__ = ['matcher', 'sphinx', 'sphinx_base14'] STYLESHEETS_PATH = os.path.join(DATA_PATH, 'stylesheets') def path(filename): return os.path.join(STYLESHEETS_PATH, filename) sphinx = StyleSheetFile(path('sphinx.rts')) sphinx_article = StyleSheetFile(path('sphinx_article.rts')) sphinx_base14 = StyleSheetFile(path('base14.rts')) # generate docstrings for the StyleSheet instances for name, stylesheet in inspect.getmembers(sys.modules[__name__]): if not isinstance(stylesheet, StyleSheetFile): continue stylesheet.__doc__ = ('{}\n\nEntry point name: ``{}``' .format(stylesheet.description, stylesheet))
Fix the auto-generated docstrings of style sheets
Fix the auto-generated docstrings of style sheets
Python
agpl-3.0
brechtm/rinohtype,brechtm/rinohtype,brechtm/rinohtype
import inspect import os import sys from .. import DATA_PATH from ..style import StyleSheetFile from .matcher import matcher __all__ = ['matcher', 'sphinx', 'sphinx_base14'] STYLESHEETS_PATH = os.path.join(DATA_PATH, 'stylesheets') def path(filename): return os.path.join(STYLESHEETS_PATH, filename) sphinx = StyleSheetFile(path('sphinx.rts')) sphinx_article = StyleSheetFile(path('sphinx_article.rts')) sphinx_base14 = StyleSheetFile(path('base14.rts')) # generate docstrings for the StyleSheet instances for name, stylesheet in inspect.getmembers(sys.modules[__name__]): if not isinstance(stylesheet, StyleSheetFile): continue - stylesheet.__doc__ = (':entry point name: ``{}``\n\n{}' + stylesheet.__doc__ = ('{}\n\nEntry point name: ``{}``' - .format(stylesheet, stylesheet.description)) + .format(stylesheet.description, stylesheet))
Fix the auto-generated docstrings of style sheets
## Code Before: import inspect import os import sys from .. import DATA_PATH from ..style import StyleSheetFile from .matcher import matcher __all__ = ['matcher', 'sphinx', 'sphinx_base14'] STYLESHEETS_PATH = os.path.join(DATA_PATH, 'stylesheets') def path(filename): return os.path.join(STYLESHEETS_PATH, filename) sphinx = StyleSheetFile(path('sphinx.rts')) sphinx_article = StyleSheetFile(path('sphinx_article.rts')) sphinx_base14 = StyleSheetFile(path('base14.rts')) # generate docstrings for the StyleSheet instances for name, stylesheet in inspect.getmembers(sys.modules[__name__]): if not isinstance(stylesheet, StyleSheetFile): continue stylesheet.__doc__ = (':entry point name: ``{}``\n\n{}' .format(stylesheet, stylesheet.description)) ## Instruction: Fix the auto-generated docstrings of style sheets ## Code After: import inspect import os import sys from .. import DATA_PATH from ..style import StyleSheetFile from .matcher import matcher __all__ = ['matcher', 'sphinx', 'sphinx_base14'] STYLESHEETS_PATH = os.path.join(DATA_PATH, 'stylesheets') def path(filename): return os.path.join(STYLESHEETS_PATH, filename) sphinx = StyleSheetFile(path('sphinx.rts')) sphinx_article = StyleSheetFile(path('sphinx_article.rts')) sphinx_base14 = StyleSheetFile(path('base14.rts')) # generate docstrings for the StyleSheet instances for name, stylesheet in inspect.getmembers(sys.modules[__name__]): if not isinstance(stylesheet, StyleSheetFile): continue stylesheet.__doc__ = ('{}\n\nEntry point name: ``{}``' .format(stylesheet.description, stylesheet))
... if not isinstance(stylesheet, StyleSheetFile): continue stylesheet.__doc__ = ('{}\n\nEntry point name: ``{}``' .format(stylesheet.description, stylesheet)) ...
d7a227ae5f0f53b5c620864df08c7b883402e968
netmiko/brocade/brocade_nos_ssh.py
netmiko/brocade/brocade_nos_ssh.py
"""Support for Brocade NOS/VDX.""" from __future__ import unicode_literals import time from netmiko.cisco_base_connection import CiscoSSHConnection class BrocadeNosSSH(CiscoSSHConnection): """Support for Brocade NOS/VDX.""" def enable(self, *args, **kwargs): """No enable mode on Brocade VDX.""" pass def exit_enable_mode(self, *args, **kwargs): """No enable mode on Brocade VDX.""" pass def special_login_handler(self, delay_factor=1): """Adding a delay after login.""" delay_factor = self.select_delay_factor(delay_factor) self.write_channel(self.RETURN) time.sleep(1 * delay_factor)
"""Support for Brocade NOS/VDX.""" from __future__ import unicode_literals import time from netmiko.cisco_base_connection import CiscoSSHConnection class BrocadeNosSSH(CiscoSSHConnection): """Support for Brocade NOS/VDX.""" def enable(self, *args, **kwargs): """No enable mode on Brocade VDX.""" pass def exit_enable_mode(self, *args, **kwargs): """No enable mode on Brocade VDX.""" pass def special_login_handler(self, delay_factor=1): """Adding a delay after login.""" delay_factor = self.select_delay_factor(delay_factor) self.write_channel(self.RETURN) time.sleep(1 * delay_factor) def save_config(self): """Save Config for Brocade VDX.""" self.send_command('copy running-config startup-config', '[Y/N]') self.send_command('y')
Add save_config for brocade VDX
Add save_config for brocade VDX
Python
mit
ktbyers/netmiko,ktbyers/netmiko
"""Support for Brocade NOS/VDX.""" from __future__ import unicode_literals import time from netmiko.cisco_base_connection import CiscoSSHConnection class BrocadeNosSSH(CiscoSSHConnection): """Support for Brocade NOS/VDX.""" def enable(self, *args, **kwargs): """No enable mode on Brocade VDX.""" pass def exit_enable_mode(self, *args, **kwargs): """No enable mode on Brocade VDX.""" pass def special_login_handler(self, delay_factor=1): """Adding a delay after login.""" delay_factor = self.select_delay_factor(delay_factor) self.write_channel(self.RETURN) time.sleep(1 * delay_factor) + def save_config(self): + """Save Config for Brocade VDX.""" + self.send_command('copy running-config startup-config', '[Y/N]') + self.send_command('y') +
Add save_config for brocade VDX
## Code Before: """Support for Brocade NOS/VDX.""" from __future__ import unicode_literals import time from netmiko.cisco_base_connection import CiscoSSHConnection class BrocadeNosSSH(CiscoSSHConnection): """Support for Brocade NOS/VDX.""" def enable(self, *args, **kwargs): """No enable mode on Brocade VDX.""" pass def exit_enable_mode(self, *args, **kwargs): """No enable mode on Brocade VDX.""" pass def special_login_handler(self, delay_factor=1): """Adding a delay after login.""" delay_factor = self.select_delay_factor(delay_factor) self.write_channel(self.RETURN) time.sleep(1 * delay_factor) ## Instruction: Add save_config for brocade VDX ## Code After: """Support for Brocade NOS/VDX.""" from __future__ import unicode_literals import time from netmiko.cisco_base_connection import CiscoSSHConnection class BrocadeNosSSH(CiscoSSHConnection): """Support for Brocade NOS/VDX.""" def enable(self, *args, **kwargs): """No enable mode on Brocade VDX.""" pass def exit_enable_mode(self, *args, **kwargs): """No enable mode on Brocade VDX.""" pass def special_login_handler(self, delay_factor=1): """Adding a delay after login.""" delay_factor = self.select_delay_factor(delay_factor) self.write_channel(self.RETURN) time.sleep(1 * delay_factor) def save_config(self): """Save Config for Brocade VDX.""" self.send_command('copy running-config startup-config', '[Y/N]') self.send_command('y')
... self.write_channel(self.RETURN) time.sleep(1 * delay_factor) def save_config(self): """Save Config for Brocade VDX.""" self.send_command('copy running-config startup-config', '[Y/N]') self.send_command('y') ...
81b1cf6973dde3ca23bbe5ac071d3decad81079a
pydsa/sleep_sort.py
pydsa/sleep_sort.py
from time import sleep from threading import Timer # Sleep Sort ;) # Complexity: O(max(input)+n) def sleep_sort(a): """ Sorts the list 'a' using Sleep sort algorithm >>> from pydsa import sleep_sort >>> a = [3, 4, 2] >>> sleep_sort(a) [2, 3, 4] """ sleep_sort.result = [] def add1(x): sleep_sort.result.append(x) mx = a[0] for v in a: if mx < v: mx = v Timer(v, add1, [v]).start() sleep(mx + 1) return sleep_sort.result
from time import sleep from threading import Timer # Sleep Sort ;) # Complexity: O(max(input)+n) def sleep_sort(a): """ Sorts the list 'a' using Sleep sort algorithm >>> from pydsa import sleep_sort >>> a = [3, 4, 2] >>> sleep_sort(a) [2, 3, 4] """ sleep_sort.result = [] def add1(x): sleep_sort.result.append(x) mx = a[0] for v in a: if mx < v: mx = v Timer(v, add1, [v]).start() sleep(mx + 1) return sleep_sort.result
Format code according to PEP8
Format code according to PEP8
Python
bsd-3-clause
rehassachdeva/pydsa,aktech/pydsa
from time import sleep from threading import Timer # Sleep Sort ;) # Complexity: O(max(input)+n) + def sleep_sort(a): """ Sorts the list 'a' using Sleep sort algorithm >>> from pydsa import sleep_sort >>> a = [3, 4, 2] >>> sleep_sort(a) [2, 3, 4] """ sleep_sort.result = [] + def add1(x): sleep_sort.result.append(x) + mx = a[0] for v in a: - if mx < v: mx = v + if mx < v: + mx = v Timer(v, add1, [v]).start() sleep(mx + 1) return sleep_sort.result
Format code according to PEP8
## Code Before: from time import sleep from threading import Timer # Sleep Sort ;) # Complexity: O(max(input)+n) def sleep_sort(a): """ Sorts the list 'a' using Sleep sort algorithm >>> from pydsa import sleep_sort >>> a = [3, 4, 2] >>> sleep_sort(a) [2, 3, 4] """ sleep_sort.result = [] def add1(x): sleep_sort.result.append(x) mx = a[0] for v in a: if mx < v: mx = v Timer(v, add1, [v]).start() sleep(mx + 1) return sleep_sort.result ## Instruction: Format code according to PEP8 ## Code After: from time import sleep from threading import Timer # Sleep Sort ;) # Complexity: O(max(input)+n) def sleep_sort(a): """ Sorts the list 'a' using Sleep sort algorithm >>> from pydsa import sleep_sort >>> a = [3, 4, 2] >>> sleep_sort(a) [2, 3, 4] """ sleep_sort.result = [] def add1(x): sleep_sort.result.append(x) mx = a[0] for v in a: if mx < v: mx = v Timer(v, add1, [v]).start() sleep(mx + 1) return sleep_sort.result
... # Sleep Sort ;) # Complexity: O(max(input)+n) def sleep_sort(a): ... """ sleep_sort.result = [] def add1(x): sleep_sort.result.append(x) mx = a[0] for v in a: if mx < v: mx = v Timer(v, add1, [v]).start() sleep(mx + 1) ...
39c777d6fc5555534628113190bb543c6225c07e
uncurl/bin.py
uncurl/bin.py
from __future__ import print_function import sys from .api import parse def main(): result = parse(sys.argv[1]) print(result)
from __future__ import print_function import sys from .api import parse def main(): if sys.stdin.isatty(): result = parse(sys.argv[1]) else: result = parse(sys.stdin.read()) print(result)
Read from stdin if available.
Read from stdin if available.
Python
apache-2.0
weinerjm/uncurl,spulec/uncurl
from __future__ import print_function import sys from .api import parse def main(): + if sys.stdin.isatty(): - result = parse(sys.argv[1]) + result = parse(sys.argv[1]) + else: + result = parse(sys.stdin.read()) print(result)
Read from stdin if available.
## Code Before: from __future__ import print_function import sys from .api import parse def main(): result = parse(sys.argv[1]) print(result) ## Instruction: Read from stdin if available. ## Code After: from __future__ import print_function import sys from .api import parse def main(): if sys.stdin.isatty(): result = parse(sys.argv[1]) else: result = parse(sys.stdin.read()) print(result)
// ... existing code ... def main(): if sys.stdin.isatty(): result = parse(sys.argv[1]) else: result = parse(sys.stdin.read()) print(result) // ... rest of the code ...
6807e5a5966f1f37f69a54e255a9981918cc8fb6
tests/test_cmd.py
tests/test_cmd.py
import base64 import os from distutils.core import Command class TestCommand(Command): description = "Launch all tests under fusion_tables app" user_options = [] def initialize_options(self): pass def finalize_options(self): pass def create_client_secret_file(self): client_secret = open("/tmp/client_secret.json", "w") data = os.environ.get("CLIENT_SECRET") client_secret.write(base64.b64decode(data)) client_secret.close() def configure_settings(self): from django.conf import settings settings.configure( DATABASES={ "default": { "NAME": ":memory:", "ENGINE": "django.db.backends.sqlite3", "TEST": { "NAME": ":memory:" } } }, INSTALLED_APPS=( "django.contrib.contenttypes", "fusion_tables", ), ROOT_URLCONF="tests.urls", MODELS_TO_SYNC=("fusion_tables.SampleModel", ), CLIENT_SECRET_JSON_FILEPATH="/tmp/client_secret.json", LOCATION_FIELDS=("TextField", ) ) def run(self): import django from django.core.management import call_command self.create_client_secret_file() self.configure_settings() django.setup() call_command("test", "fusion_tables")
import base64 import os from distutils.core import Command class TestCommand(Command): description = "Launch all tests under fusion_tables app" user_options = [] def initialize_options(self): pass def finalize_options(self): pass def create_client_secret_file(self): client_secret = open("/tmp/client_secret.json", "w") data = os.environ.get("CLIENT_SECRET").decode("utf-8") client_secret.write(base64.b64decode(data)) client_secret.close() def configure_settings(self): from django.conf import settings settings.configure( DATABASES={ "default": { "NAME": ":memory:", "ENGINE": "django.db.backends.sqlite3", "TEST": { "NAME": ":memory:" } } }, INSTALLED_APPS=( "django.contrib.contenttypes", "fusion_tables", ), ROOT_URLCONF="tests.urls", MODELS_TO_SYNC=("fusion_tables.SampleModel", ), CLIENT_SECRET_JSON_FILEPATH="/tmp/client_secret.json", LOCATION_FIELDS=("TextField", ) ) def run(self): import django from django.core.management import call_command self.create_client_secret_file() self.configure_settings() django.setup() call_command("test", "fusion_tables")
Fix unit test python3 compatibility.
Fix unit test python3 compatibility.
Python
mit
bsvetchine/django-fusion-tables
import base64 import os from distutils.core import Command class TestCommand(Command): description = "Launch all tests under fusion_tables app" user_options = [] def initialize_options(self): pass def finalize_options(self): pass def create_client_secret_file(self): client_secret = open("/tmp/client_secret.json", "w") - data = os.environ.get("CLIENT_SECRET") + data = os.environ.get("CLIENT_SECRET").decode("utf-8") client_secret.write(base64.b64decode(data)) client_secret.close() def configure_settings(self): from django.conf import settings settings.configure( DATABASES={ "default": { "NAME": ":memory:", "ENGINE": "django.db.backends.sqlite3", "TEST": { "NAME": ":memory:" } } }, INSTALLED_APPS=( "django.contrib.contenttypes", "fusion_tables", ), ROOT_URLCONF="tests.urls", MODELS_TO_SYNC=("fusion_tables.SampleModel", ), CLIENT_SECRET_JSON_FILEPATH="/tmp/client_secret.json", LOCATION_FIELDS=("TextField", ) ) def run(self): import django from django.core.management import call_command self.create_client_secret_file() self.configure_settings() django.setup() call_command("test", "fusion_tables")
Fix unit test python3 compatibility.
## Code Before: import base64 import os from distutils.core import Command class TestCommand(Command): description = "Launch all tests under fusion_tables app" user_options = [] def initialize_options(self): pass def finalize_options(self): pass def create_client_secret_file(self): client_secret = open("/tmp/client_secret.json", "w") data = os.environ.get("CLIENT_SECRET") client_secret.write(base64.b64decode(data)) client_secret.close() def configure_settings(self): from django.conf import settings settings.configure( DATABASES={ "default": { "NAME": ":memory:", "ENGINE": "django.db.backends.sqlite3", "TEST": { "NAME": ":memory:" } } }, INSTALLED_APPS=( "django.contrib.contenttypes", "fusion_tables", ), ROOT_URLCONF="tests.urls", MODELS_TO_SYNC=("fusion_tables.SampleModel", ), CLIENT_SECRET_JSON_FILEPATH="/tmp/client_secret.json", LOCATION_FIELDS=("TextField", ) ) def run(self): import django from django.core.management import call_command self.create_client_secret_file() self.configure_settings() django.setup() call_command("test", "fusion_tables") ## Instruction: Fix unit test python3 compatibility. ## Code After: import base64 import os from distutils.core import Command class TestCommand(Command): description = "Launch all tests under fusion_tables app" user_options = [] def initialize_options(self): pass def finalize_options(self): pass def create_client_secret_file(self): client_secret = open("/tmp/client_secret.json", "w") data = os.environ.get("CLIENT_SECRET").decode("utf-8") client_secret.write(base64.b64decode(data)) client_secret.close() def configure_settings(self): from django.conf import settings settings.configure( DATABASES={ "default": { "NAME": ":memory:", "ENGINE": "django.db.backends.sqlite3", "TEST": { "NAME": ":memory:" } } }, INSTALLED_APPS=( "django.contrib.contenttypes", "fusion_tables", ), ROOT_URLCONF="tests.urls", MODELS_TO_SYNC=("fusion_tables.SampleModel", ), CLIENT_SECRET_JSON_FILEPATH="/tmp/client_secret.json", LOCATION_FIELDS=("TextField", ) ) def run(self): import django from django.core.management import call_command self.create_client_secret_file() self.configure_settings() django.setup() call_command("test", "fusion_tables")
... def create_client_secret_file(self): client_secret = open("/tmp/client_secret.json", "w") data = os.environ.get("CLIENT_SECRET").decode("utf-8") client_secret.write(base64.b64decode(data)) client_secret.close() ...
d89747e26371b1986b4cec5a7514ba2c99480487
tests/test_codec.py
tests/test_codec.py
from .common import * from av.codec import Codec, Encoder, Decoder class TestCodecs(TestCase): def test_codec_mpeg4(self): for cls in (Encoder, Decoder): c = cls('mpeg4') self.assertEqual(c.name, 'mpeg4') self.assertEqual(c.long_name, 'MPEG-4 part 2') self.assertEqual(c.type, 'video') self.assertEqual(c.id, 13) formats = c.video_formats self.assertEqual(len(formats), 1) self.assertEqual(formats[0].name, 'yuv420p')
from .common import * from av.codec import Codec, Encoder, Decoder class TestCodecs(TestCase): def test_codec_mpeg4(self): for cls in (Encoder, Decoder): c = cls('mpeg4') self.assertEqual(c.name, 'mpeg4') self.assertEqual(c.long_name, 'MPEG-4 part 2') self.assertEqual(c.type, 'video') self.assertEqual(c.id, 13) formats = c.video_formats self.assertTrue(formats) self.assertTrue(any(f.name == 'yuv420p' for f in formats))
Allow codec test to have more than just the one format
Allow codec test to have more than just the one format
Python
bsd-3-clause
mcpv/PyAV,danielballan/PyAV,pupil-labs/PyAV,PyAV-Org/PyAV,markreidvfx/PyAV,PyAV-Org/PyAV,xxr3376/PyAV,pupil-labs/PyAV,pupil-labs/PyAV,xxr3376/PyAV,mikeboers/PyAV,markreidvfx/PyAV,danielballan/PyAV,danielballan/PyAV,xxr3376/PyAV,markreidvfx/PyAV,mcpv/PyAV,mikeboers/PyAV,mcpv/PyAV,pupil-labs/PyAV
from .common import * from av.codec import Codec, Encoder, Decoder class TestCodecs(TestCase): def test_codec_mpeg4(self): for cls in (Encoder, Decoder): c = cls('mpeg4') self.assertEqual(c.name, 'mpeg4') self.assertEqual(c.long_name, 'MPEG-4 part 2') self.assertEqual(c.type, 'video') self.assertEqual(c.id, 13) formats = c.video_formats - self.assertEqual(len(formats), 1) + self.assertTrue(formats) - self.assertEqual(formats[0].name, 'yuv420p') + self.assertTrue(any(f.name == 'yuv420p' for f in formats))
Allow codec test to have more than just the one format
## Code Before: from .common import * from av.codec import Codec, Encoder, Decoder class TestCodecs(TestCase): def test_codec_mpeg4(self): for cls in (Encoder, Decoder): c = cls('mpeg4') self.assertEqual(c.name, 'mpeg4') self.assertEqual(c.long_name, 'MPEG-4 part 2') self.assertEqual(c.type, 'video') self.assertEqual(c.id, 13) formats = c.video_formats self.assertEqual(len(formats), 1) self.assertEqual(formats[0].name, 'yuv420p') ## Instruction: Allow codec test to have more than just the one format ## Code After: from .common import * from av.codec import Codec, Encoder, Decoder class TestCodecs(TestCase): def test_codec_mpeg4(self): for cls in (Encoder, Decoder): c = cls('mpeg4') self.assertEqual(c.name, 'mpeg4') self.assertEqual(c.long_name, 'MPEG-4 part 2') self.assertEqual(c.type, 'video') self.assertEqual(c.id, 13) formats = c.video_formats self.assertTrue(formats) self.assertTrue(any(f.name == 'yuv420p' for f in formats))
// ... existing code ... formats = c.video_formats self.assertTrue(formats) self.assertTrue(any(f.name == 'yuv420p' for f in formats)) // ... rest of the code ...
0c2a7bfebbb6d427ffea66f4a8df534c5b8be974
timed/subscription/admin.py
timed/subscription/admin.py
from django.contrib import admin from . import models @admin.register(models.Package) class PackageAdmin(admin.ModelAdmin): list_display = ['billing_type', 'duration', 'price']
from django import forms from django.contrib import admin from django.utils.translation import ugettext_lazy as _ from timed.forms import DurationInHoursField from . import models class PackageForm(forms.ModelForm): model = models.Package duration = DurationInHoursField( label=_('Duration in hours'), required=True, ) @admin.register(models.Package) class PackageAdmin(admin.ModelAdmin): list_display = ['billing_type', 'duration', 'price'] form = PackageForm
Configure duration field on subscription package in hours
Configure duration field on subscription package in hours
Python
agpl-3.0
adfinis-sygroup/timed-backend,adfinis-sygroup/timed-backend,adfinis-sygroup/timed-backend
+ from django import forms from django.contrib import admin + from django.utils.translation import ugettext_lazy as _ + + from timed.forms import DurationInHoursField from . import models + + + class PackageForm(forms.ModelForm): + model = models.Package + duration = DurationInHoursField( + label=_('Duration in hours'), + required=True, + ) @admin.register(models.Package) class PackageAdmin(admin.ModelAdmin): list_display = ['billing_type', 'duration', 'price'] + form = PackageForm
Configure duration field on subscription package in hours
## Code Before: from django.contrib import admin from . import models @admin.register(models.Package) class PackageAdmin(admin.ModelAdmin): list_display = ['billing_type', 'duration', 'price'] ## Instruction: Configure duration field on subscription package in hours ## Code After: from django import forms from django.contrib import admin from django.utils.translation import ugettext_lazy as _ from timed.forms import DurationInHoursField from . import models class PackageForm(forms.ModelForm): model = models.Package duration = DurationInHoursField( label=_('Duration in hours'), required=True, ) @admin.register(models.Package) class PackageAdmin(admin.ModelAdmin): list_display = ['billing_type', 'duration', 'price'] form = PackageForm
// ... existing code ... from django import forms from django.contrib import admin from django.utils.translation import ugettext_lazy as _ from timed.forms import DurationInHoursField from . import models class PackageForm(forms.ModelForm): model = models.Package duration = DurationInHoursField( label=_('Duration in hours'), required=True, ) // ... modified code ... class PackageAdmin(admin.ModelAdmin): list_display = ['billing_type', 'duration', 'price'] form = PackageForm // ... rest of the code ...
d187a8434c9d64171f76efa3055bdc06afbc8981
scripts/pystart.py
scripts/pystart.py
import os,sys,re from time import sleep from pprint import pprint home = os.path.expanduser('~') from math import log,ceil def clog2(num): return int(ceil(log(num,2))) if (sys.version_info > (3, 0)): # Python 3 code in this block exec(open(home+'/homedir/scripts/hexecho.py').read()) else: # Python 2 code in this block execfile(home+'/homedir/scripts/hexecho.py') hexoff print ("Imported os,sys,re,sleep,pprint. Defined clog2,hexon/hexoff")
import os,sys,re from time import sleep from pprint import pprint home = os.path.expanduser('~') from math import log,ceil sys.ps1 = '\001\033[96m\002>>> \001\033[0m\002' sys.ps2 = '\001\033[96m\002... \001\033[0m\002' def clog2(num): return int(ceil(log(num,2))) if (sys.version_info > (3, 0)): # Python 3 code in this block exec(open(home+'/homedir/scripts/hexecho.py').read()) else: # Python 2 code in this block execfile(home+'/homedir/scripts/hexecho.py') hexoff print ("Imported os,sys,re,sleep,pprint. Defined clog2,hexon/hexoff")
Add color to python prompt
Add color to python prompt
Python
mit
jdanders/homedir,jdanders/homedir,jdanders/homedir,jdanders/homedir
import os,sys,re from time import sleep from pprint import pprint home = os.path.expanduser('~') from math import log,ceil + sys.ps1 = '\001\033[96m\002>>> \001\033[0m\002' + sys.ps2 = '\001\033[96m\002... \001\033[0m\002' def clog2(num): return int(ceil(log(num,2))) if (sys.version_info > (3, 0)): # Python 3 code in this block exec(open(home+'/homedir/scripts/hexecho.py').read()) else: # Python 2 code in this block execfile(home+'/homedir/scripts/hexecho.py') hexoff print ("Imported os,sys,re,sleep,pprint. Defined clog2,hexon/hexoff")
Add color to python prompt
## Code Before: import os,sys,re from time import sleep from pprint import pprint home = os.path.expanduser('~') from math import log,ceil def clog2(num): return int(ceil(log(num,2))) if (sys.version_info > (3, 0)): # Python 3 code in this block exec(open(home+'/homedir/scripts/hexecho.py').read()) else: # Python 2 code in this block execfile(home+'/homedir/scripts/hexecho.py') hexoff print ("Imported os,sys,re,sleep,pprint. Defined clog2,hexon/hexoff") ## Instruction: Add color to python prompt ## Code After: import os,sys,re from time import sleep from pprint import pprint home = os.path.expanduser('~') from math import log,ceil sys.ps1 = '\001\033[96m\002>>> \001\033[0m\002' sys.ps2 = '\001\033[96m\002... \001\033[0m\002' def clog2(num): return int(ceil(log(num,2))) if (sys.version_info > (3, 0)): # Python 3 code in this block exec(open(home+'/homedir/scripts/hexecho.py').read()) else: # Python 2 code in this block execfile(home+'/homedir/scripts/hexecho.py') hexoff print ("Imported os,sys,re,sleep,pprint. Defined clog2,hexon/hexoff")
... home = os.path.expanduser('~') from math import log,ceil sys.ps1 = '\001\033[96m\002>>> \001\033[0m\002' sys.ps2 = '\001\033[96m\002... \001\033[0m\002' def clog2(num): return int(ceil(log(num,2))) ...
2f140327c24a8efab5482a975793dddedd0ebfc4
nucleus/wsgi.py
nucleus/wsgi.py
# newrelic.agent must be imported and initialized first # https://docs.newrelic.com/docs/agents/python-agent/installation/python-agent-advanced-integration#manual-integration import newrelic.agent newrelic.agent.initialize('newrelic.ini') import os os.environ.setdefault('DJANGO_SETTINGS_MODULE', 'nucleus.settings') # NOQA from django.core.handlers.wsgi import WSGIRequest from django.core.wsgi import get_wsgi_application from decouple import config IS_HTTPS = config('HTTPS', default='off', cast=bool) class WSGIHTTPSRequest(WSGIRequest): def _get_scheme(self): if IS_HTTPS: return 'https' return super(WSGIHTTPSRequest, self)._get_scheme() application = get_wsgi_application() application.request_class = WSGIHTTPSRequest if config('SENTRY_DSN', None): from raven.contrib.django.raven_compat.middleware.wsgi import Sentry application = Sentry(application) newrelic_license_key = config('NEW_RELIC_LICENSE_KEY', default=None) if newrelic_license_key: application = newrelic.agent.WSGIApplicationWrapper(application)
import newrelic.agent newrelic.agent.initialize('newrelic.ini') import os os.environ.setdefault('DJANGO_SETTINGS_MODULE', 'nucleus.settings') # NOQA from django.core.handlers.wsgi import WSGIRequest from django.core.wsgi import get_wsgi_application from decouple import config IS_HTTPS = config('HTTPS', default='off', cast=bool) class WSGIHTTPSRequest(WSGIRequest): def _get_scheme(self): if IS_HTTPS: return 'https' return super(WSGIHTTPSRequest, self)._get_scheme() application = get_wsgi_application() application.request_class = WSGIHTTPSRequest if config('SENTRY_DSN', None): from raven.contrib.django.raven_compat.middleware.wsgi import Sentry application = Sentry(application) newrelic_license_key = config('NEW_RELIC_LICENSE_KEY', default=None) if newrelic_license_key: application = newrelic.agent.WSGIApplicationWrapper(application)
Remove old docstring with link to old django docs
Remove old docstring with link to old django docs
Python
mpl-2.0
mozilla/nucleus,mozilla/nucleus,mozilla/nucleus,mozilla/nucleus
- # newrelic.agent must be imported and initialized first - # https://docs.newrelic.com/docs/agents/python-agent/installation/python-agent-advanced-integration#manual-integration import newrelic.agent newrelic.agent.initialize('newrelic.ini') import os os.environ.setdefault('DJANGO_SETTINGS_MODULE', 'nucleus.settings') # NOQA from django.core.handlers.wsgi import WSGIRequest from django.core.wsgi import get_wsgi_application from decouple import config IS_HTTPS = config('HTTPS', default='off', cast=bool) class WSGIHTTPSRequest(WSGIRequest): def _get_scheme(self): if IS_HTTPS: return 'https' return super(WSGIHTTPSRequest, self)._get_scheme() application = get_wsgi_application() application.request_class = WSGIHTTPSRequest if config('SENTRY_DSN', None): from raven.contrib.django.raven_compat.middleware.wsgi import Sentry application = Sentry(application) newrelic_license_key = config('NEW_RELIC_LICENSE_KEY', default=None) if newrelic_license_key: application = newrelic.agent.WSGIApplicationWrapper(application)
Remove old docstring with link to old django docs
## Code Before: # newrelic.agent must be imported and initialized first # https://docs.newrelic.com/docs/agents/python-agent/installation/python-agent-advanced-integration#manual-integration import newrelic.agent newrelic.agent.initialize('newrelic.ini') import os os.environ.setdefault('DJANGO_SETTINGS_MODULE', 'nucleus.settings') # NOQA from django.core.handlers.wsgi import WSGIRequest from django.core.wsgi import get_wsgi_application from decouple import config IS_HTTPS = config('HTTPS', default='off', cast=bool) class WSGIHTTPSRequest(WSGIRequest): def _get_scheme(self): if IS_HTTPS: return 'https' return super(WSGIHTTPSRequest, self)._get_scheme() application = get_wsgi_application() application.request_class = WSGIHTTPSRequest if config('SENTRY_DSN', None): from raven.contrib.django.raven_compat.middleware.wsgi import Sentry application = Sentry(application) newrelic_license_key = config('NEW_RELIC_LICENSE_KEY', default=None) if newrelic_license_key: application = newrelic.agent.WSGIApplicationWrapper(application) ## Instruction: Remove old docstring with link to old django docs ## Code After: import newrelic.agent newrelic.agent.initialize('newrelic.ini') import os os.environ.setdefault('DJANGO_SETTINGS_MODULE', 'nucleus.settings') # NOQA from django.core.handlers.wsgi import WSGIRequest from django.core.wsgi import get_wsgi_application from decouple import config IS_HTTPS = config('HTTPS', default='off', cast=bool) class WSGIHTTPSRequest(WSGIRequest): def _get_scheme(self): if IS_HTTPS: return 'https' return super(WSGIHTTPSRequest, self)._get_scheme() application = get_wsgi_application() application.request_class = WSGIHTTPSRequest if config('SENTRY_DSN', None): from raven.contrib.django.raven_compat.middleware.wsgi import Sentry application = Sentry(application) newrelic_license_key = config('NEW_RELIC_LICENSE_KEY', default=None) if newrelic_license_key: application = newrelic.agent.WSGIApplicationWrapper(application)
... import newrelic.agent newrelic.agent.initialize('newrelic.ini') ...
df227a375c1cf5fdd0ad23505799e7c6f7177b9c
InvenTree/InvenTree/validators.py
InvenTree/InvenTree/validators.py
from django.core.exceptions import ValidationError from django.utils.translation import gettext_lazy as _ def validate_part_name(value): # Prevent some illegal characters in part names for c in ['/', '\\', '|', '#', '$']: if c in str(value): raise ValidationError( _('Invalid character in part name') )
from django.core.exceptions import ValidationError from django.utils.translation import gettext_lazy as _ def validate_part_name(value): # Prevent some illegal characters in part names for c in ['|', '#', '$']: if c in str(value): raise ValidationError( _('Invalid character in part name') )
Allow some more chars in part names
Allow some more chars in part names
Python
mit
inventree/InvenTree,inventree/InvenTree,SchrodingersGat/InvenTree,inventree/InvenTree,SchrodingersGat/InvenTree,inventree/InvenTree,SchrodingersGat/InvenTree,SchrodingersGat/InvenTree
from django.core.exceptions import ValidationError from django.utils.translation import gettext_lazy as _ def validate_part_name(value): # Prevent some illegal characters in part names - for c in ['/', '\\', '|', '#', '$']: + for c in ['|', '#', '$']: if c in str(value): raise ValidationError( _('Invalid character in part name') )
Allow some more chars in part names
## Code Before: from django.core.exceptions import ValidationError from django.utils.translation import gettext_lazy as _ def validate_part_name(value): # Prevent some illegal characters in part names for c in ['/', '\\', '|', '#', '$']: if c in str(value): raise ValidationError( _('Invalid character in part name') ) ## Instruction: Allow some more chars in part names ## Code After: from django.core.exceptions import ValidationError from django.utils.translation import gettext_lazy as _ def validate_part_name(value): # Prevent some illegal characters in part names for c in ['|', '#', '$']: if c in str(value): raise ValidationError( _('Invalid character in part name') )
// ... existing code ... def validate_part_name(value): # Prevent some illegal characters in part names for c in ['|', '#', '$']: if c in str(value): raise ValidationError( // ... rest of the code ...
7d8283b2d233a8fbee97de122f0b4ba293cf788d
app/emails.py
app/emails.py
from flask import render_template,g from flask.ext.mail import Message from app import mail, db from .models import User from config import MAIL_SENDER # Wrapper function for sending mails using flask-mail plugin def send_email(subject, sender, recipients, text_body): msg = Message(subject, sender=sender, recipients=recipients) msg.body = text_body mail.send(msg) def add_movie_notification(movie): users = User.query.all() you_user = False for cur_user in users: # Check if the cur_user is the logged user who added the movie # in order to change the mail text if cur_user.id==g.user.id: you_user=True send_email('[Cineapp] - Ajout d\'un film' , MAIL_SENDER,[ cur_user.email ] , render_template('add_movie_notification.txt', dest_user=cur_user, add_user=g.user,movie=movie,you_user=you_user))
from flask import render_template,g from flask.ext.mail import Message from app import mail, db from .models import User from config import MAIL_SENDER from threading import Thread from app import app # Send mail into a dedicated thread in order to avoir the web app to wait def send_async_email(app, msg): with app.app_context(): mail.send(msg) # Wrapper function for sending mails using flask-mail plugin def send_email(subject, sender, recipients, text_body): msg = Message(subject, sender=sender, recipients=recipients) msg.body = text_body thr = Thread(target=send_async_email, args=[app, msg]) thr.start() # Function which sends notifications to users when a movie is added def add_movie_notification(movie): users = User.query.all() you_user = False for cur_user in users: # Check if the cur_user is the logged user who added the movie # in order to change the mail text if cur_user.id==g.user.id: you_user=True send_email('[Cineapp] - Ajout d\'un film' , MAIL_SENDER,[ cur_user.email ] , render_template('add_movie_notification.txt', dest_user=cur_user, add_user=g.user,movie=movie,you_user=you_user))
Send email notifications in asynchronous mode
Send email notifications in asynchronous mode Each mail notification is done on a thread in order to not block the main thread of the web app.
Python
mit
ptitoliv/cineapp,ptitoliv/cineapp,ptitoliv/cineapp
from flask import render_template,g from flask.ext.mail import Message from app import mail, db from .models import User from config import MAIL_SENDER + from threading import Thread + from app import app + + # Send mail into a dedicated thread in order to avoir the web app to wait + def send_async_email(app, msg): + with app.app_context(): + mail.send(msg) # Wrapper function for sending mails using flask-mail plugin def send_email(subject, sender, recipients, text_body): msg = Message(subject, sender=sender, recipients=recipients) msg.body = text_body - mail.send(msg) + thr = Thread(target=send_async_email, args=[app, msg]) + thr.start() + # Function which sends notifications to users when a movie is added def add_movie_notification(movie): users = User.query.all() you_user = False for cur_user in users: # Check if the cur_user is the logged user who added the movie # in order to change the mail text if cur_user.id==g.user.id: you_user=True send_email('[Cineapp] - Ajout d\'un film' , MAIL_SENDER,[ cur_user.email ] , render_template('add_movie_notification.txt', dest_user=cur_user, add_user=g.user,movie=movie,you_user=you_user))
Send email notifications in asynchronous mode
## Code Before: from flask import render_template,g from flask.ext.mail import Message from app import mail, db from .models import User from config import MAIL_SENDER # Wrapper function for sending mails using flask-mail plugin def send_email(subject, sender, recipients, text_body): msg = Message(subject, sender=sender, recipients=recipients) msg.body = text_body mail.send(msg) def add_movie_notification(movie): users = User.query.all() you_user = False for cur_user in users: # Check if the cur_user is the logged user who added the movie # in order to change the mail text if cur_user.id==g.user.id: you_user=True send_email('[Cineapp] - Ajout d\'un film' , MAIL_SENDER,[ cur_user.email ] , render_template('add_movie_notification.txt', dest_user=cur_user, add_user=g.user,movie=movie,you_user=you_user)) ## Instruction: Send email notifications in asynchronous mode ## Code After: from flask import render_template,g from flask.ext.mail import Message from app import mail, db from .models import User from config import MAIL_SENDER from threading import Thread from app import app # Send mail into a dedicated thread in order to avoir the web app to wait def send_async_email(app, msg): with app.app_context(): mail.send(msg) # Wrapper function for sending mails using flask-mail plugin def send_email(subject, sender, recipients, text_body): msg = Message(subject, sender=sender, recipients=recipients) msg.body = text_body thr = Thread(target=send_async_email, args=[app, msg]) thr.start() # Function which sends notifications to users when a movie is added def add_movie_notification(movie): users = User.query.all() you_user = False for cur_user in users: # Check if the cur_user is the logged user who added the movie # in order to change the mail text if cur_user.id==g.user.id: you_user=True send_email('[Cineapp] - Ajout d\'un film' , MAIL_SENDER,[ cur_user.email ] , render_template('add_movie_notification.txt', dest_user=cur_user, add_user=g.user,movie=movie,you_user=you_user))
// ... existing code ... from .models import User from config import MAIL_SENDER from threading import Thread from app import app # Send mail into a dedicated thread in order to avoir the web app to wait def send_async_email(app, msg): with app.app_context(): mail.send(msg) # Wrapper function for sending mails using flask-mail plugin // ... modified code ... msg = Message(subject, sender=sender, recipients=recipients) msg.body = text_body thr = Thread(target=send_async_email, args=[app, msg]) thr.start() # Function which sends notifications to users when a movie is added def add_movie_notification(movie): users = User.query.all() // ... rest of the code ...
be964b02036159567efcaecce5b5d905f23985af
deduper/scanfiles.py
deduper/scanfiles.py
import os from sqlalchemy.ext.declarative import declarative_base Base = declarative_base() from .models import ImageFile from .util import Util def ScanFiles(session, FOLDER): for root, dirs, files in os.walk(FOLDER): if root.split('/')[-1].startswith('.'): continue for count, filename in enumerate(files, start=1): if filename.startswith('.'): continue fullpath = os.path.join(root, filename) if Util.file_record_exists(session, fullpath): print('{count} of {length}: Skipping {filename}'.format( count=count, length=len(files), filename=filename)) else: print('{count} of {length}: Processing {filename}'.format( count=count, length=len(files), filename=filename)) new_file = ImageFile(name=filename, fullpath=fullpath, filehash=Util.hash_file(fullpath)) session.add(new_file) session.commit() session.close()
import os from sqlalchemy.ext.declarative import declarative_base Base = declarative_base() from .models import ImageFile from .util import Util def ScanFiles(session, FOLDER): for root, dirs, files in os.walk(FOLDER): if root.split('/')[-1].startswith('.'): continue for count, filename in enumerate(files, start=1): if filename.startswith('.'): continue fullpath = os.path.join(root, filename) if not os.path.isfile(fullpath): continue if Util.file_record_exists(session, fullpath): print('{count} of {length}: Skipping {filename}'.format( count=count, length=len(files), filename=filename)) else: print('{count} of {length}: Processing {filename}'.format( count=count, length=len(files), filename=filename)) new_file = ImageFile(name=filename, fullpath=fullpath, filehash=Util.hash_file(fullpath)) session.add(new_file) session.commit() session.close()
Check that fullpath is a regular file before continuing
Check that fullpath is a regular file before continuing
Python
bsd-3-clause
cgspeck/filededuper
import os from sqlalchemy.ext.declarative import declarative_base Base = declarative_base() from .models import ImageFile from .util import Util def ScanFiles(session, FOLDER): for root, dirs, files in os.walk(FOLDER): if root.split('/')[-1].startswith('.'): continue for count, filename in enumerate(files, start=1): if filename.startswith('.'): continue fullpath = os.path.join(root, filename) + if not os.path.isfile(fullpath): + continue + if Util.file_record_exists(session, fullpath): print('{count} of {length}: Skipping {filename}'.format( count=count, length=len(files), filename=filename)) else: print('{count} of {length}: Processing {filename}'.format( count=count, length=len(files), filename=filename)) new_file = ImageFile(name=filename, fullpath=fullpath, filehash=Util.hash_file(fullpath)) session.add(new_file) session.commit() session.close()
Check that fullpath is a regular file before continuing
## Code Before: import os from sqlalchemy.ext.declarative import declarative_base Base = declarative_base() from .models import ImageFile from .util import Util def ScanFiles(session, FOLDER): for root, dirs, files in os.walk(FOLDER): if root.split('/')[-1].startswith('.'): continue for count, filename in enumerate(files, start=1): if filename.startswith('.'): continue fullpath = os.path.join(root, filename) if Util.file_record_exists(session, fullpath): print('{count} of {length}: Skipping {filename}'.format( count=count, length=len(files), filename=filename)) else: print('{count} of {length}: Processing {filename}'.format( count=count, length=len(files), filename=filename)) new_file = ImageFile(name=filename, fullpath=fullpath, filehash=Util.hash_file(fullpath)) session.add(new_file) session.commit() session.close() ## Instruction: Check that fullpath is a regular file before continuing ## Code After: import os from sqlalchemy.ext.declarative import declarative_base Base = declarative_base() from .models import ImageFile from .util import Util def ScanFiles(session, FOLDER): for root, dirs, files in os.walk(FOLDER): if root.split('/')[-1].startswith('.'): continue for count, filename in enumerate(files, start=1): if filename.startswith('.'): continue fullpath = os.path.join(root, filename) if not os.path.isfile(fullpath): continue if Util.file_record_exists(session, fullpath): print('{count} of {length}: Skipping {filename}'.format( count=count, length=len(files), filename=filename)) else: print('{count} of {length}: Processing {filename}'.format( count=count, length=len(files), filename=filename)) new_file = ImageFile(name=filename, fullpath=fullpath, filehash=Util.hash_file(fullpath)) session.add(new_file) session.commit() session.close()
// ... existing code ... fullpath = os.path.join(root, filename) if not os.path.isfile(fullpath): continue if Util.file_record_exists(session, fullpath): print('{count} of {length}: Skipping {filename}'.format( // ... rest of the code ...
c5239c6bbb40ede4279b33b965c5ded26a78b2ae
app/tests/manual/test_twitter_api.py
app/tests/manual/test_twitter_api.py
from __future__ import absolute_import from unittest import TestCase from lib.twitter_api import authentication class TestAuth(TestCase): def test_generateAppAccessToken(self): auth = authentication._generateAppAccessToken() def test_getTweepyConnection(self): auth = authentication._generateAppAccessToken() api = authentication._getTweepyConnection(auth) def test_getAPIConnection(self): """ Test that App Access token can be used to connect to Twitter API. """ api = authentication.getAPIConnection(userFlow=False) def test_getAppOnlyConnection(self): """ Test App-only token. """ api = authentication.getAppOnlyConnection()
from __future__ import absolute_import import os import sys import unittest from unittest import TestCase # Allow imports to be done when executing this file directly. sys.path.insert(0, os.path.abspath(os.path.join( os.path.dirname(__file__), os.path.pardir, os.path.pardir) )) from lib.twitter_api import authentication class TestAuth(TestCase): def test_generateAppAccessToken(self): auth = authentication._generateAppAccessToken() def test_getTweepyConnection(self): auth = authentication._generateAppAccessToken() api = authentication._getTweepyConnection(auth) def test_getAPIConnection(self): """ Test that App Access token can be used to connect to Twitter API. """ api = authentication.getAPIConnection(userFlow=False) def test_getAppOnlyConnection(self): """ Test App-only token. """ api = authentication.getAppOnlyConnection() if __name__ == '__main__': unittest.main()
Update Twitter auth test to run directly
test: Update Twitter auth test to run directly
Python
mit
MichaelCurrin/twitterverse,MichaelCurrin/twitterverse
from __future__ import absolute_import + import os + import sys + import unittest from unittest import TestCase + + # Allow imports to be done when executing this file directly. + sys.path.insert(0, os.path.abspath(os.path.join( + os.path.dirname(__file__), os.path.pardir, os.path.pardir) + )) + from lib.twitter_api import authentication class TestAuth(TestCase): def test_generateAppAccessToken(self): auth = authentication._generateAppAccessToken() def test_getTweepyConnection(self): auth = authentication._generateAppAccessToken() api = authentication._getTweepyConnection(auth) def test_getAPIConnection(self): """ Test that App Access token can be used to connect to Twitter API. """ api = authentication.getAPIConnection(userFlow=False) def test_getAppOnlyConnection(self): """ Test App-only token. """ api = authentication.getAppOnlyConnection() + + if __name__ == '__main__': + unittest.main() +
Update Twitter auth test to run directly
## Code Before: from __future__ import absolute_import from unittest import TestCase from lib.twitter_api import authentication class TestAuth(TestCase): def test_generateAppAccessToken(self): auth = authentication._generateAppAccessToken() def test_getTweepyConnection(self): auth = authentication._generateAppAccessToken() api = authentication._getTweepyConnection(auth) def test_getAPIConnection(self): """ Test that App Access token can be used to connect to Twitter API. """ api = authentication.getAPIConnection(userFlow=False) def test_getAppOnlyConnection(self): """ Test App-only token. """ api = authentication.getAppOnlyConnection() ## Instruction: Update Twitter auth test to run directly ## Code After: from __future__ import absolute_import import os import sys import unittest from unittest import TestCase # Allow imports to be done when executing this file directly. sys.path.insert(0, os.path.abspath(os.path.join( os.path.dirname(__file__), os.path.pardir, os.path.pardir) )) from lib.twitter_api import authentication class TestAuth(TestCase): def test_generateAppAccessToken(self): auth = authentication._generateAppAccessToken() def test_getTweepyConnection(self): auth = authentication._generateAppAccessToken() api = authentication._getTweepyConnection(auth) def test_getAPIConnection(self): """ Test that App Access token can be used to connect to Twitter API. """ api = authentication.getAPIConnection(userFlow=False) def test_getAppOnlyConnection(self): """ Test App-only token. """ api = authentication.getAppOnlyConnection() if __name__ == '__main__': unittest.main()
// ... existing code ... from __future__ import absolute_import import os import sys import unittest from unittest import TestCase # Allow imports to be done when executing this file directly. sys.path.insert(0, os.path.abspath(os.path.join( os.path.dirname(__file__), os.path.pardir, os.path.pardir) )) from lib.twitter_api import authentication // ... modified code ... """ api = authentication.getAppOnlyConnection() if __name__ == '__main__': unittest.main() // ... rest of the code ...
8b538c452242050e468b71ca937e3d4feb57887b
mopidy/backends/stream/__init__.py
mopidy/backends/stream/__init__.py
from __future__ import unicode_literals import mopidy from mopidy import ext __doc__ = """A backend for playing music for streaming music. This backend will handle streaming of URIs in :attr:`mopidy.settings.STREAM_PROTOCOLS` assuming the right plugins are installed. **Issues:** https://github.com/mopidy/mopidy/issues?labels=Stream+backend **Dependencies:** - None **Settings:** - :attr:`mopidy.settings.STREAM_PROTOCOLS` """ class Extension(ext.Extension): name = 'Mopidy-Stream' version = mopidy.__version__ def get_default_config(self): return '[ext.stream]' def validate_config(self, config): pass def validate_environment(self): pass def get_backend_classes(self): from .actor import StreamBackend return [StreamBackend]
from __future__ import unicode_literals import mopidy from mopidy import ext from mopidy.utils import config, formatting default_config = """ [ext.stream] # If the stream extension should be enabled or not enabled = true # Whitelist of URI schemas to support streaming from protocols = http https mms rtmp rtmps rtsp """ __doc__ = """A backend for playing music for streaming music. This backend will handle streaming of URIs in :attr:`mopidy.settings.STREAM_PROTOCOLS` assuming the right plugins are installed. **Issues:** https://github.com/mopidy/mopidy/issues?labels=Stream+backend **Dependencies:** - None **Default config:** .. code-block:: ini %(config)s """ % {'config': formatting.indent(default_config)} class Extension(ext.Extension): name = 'Mopidy-Stream' version = mopidy.__version__ def get_default_config(self): return default_config def get_config_schema(self): schema = config.ExtensionConfigSchema() schema['protocols'] = config.List() return schema def validate_environment(self): pass def get_backend_classes(self): from .actor import StreamBackend return [StreamBackend]
Add default config and config schema
stream: Add default config and config schema
Python
apache-2.0
tkem/mopidy,jcass77/mopidy,jmarsik/mopidy,ZenithDK/mopidy,swak/mopidy,vrs01/mopidy,diandiankan/mopidy,quartz55/mopidy,adamcik/mopidy,abarisain/mopidy,liamw9534/mopidy,vrs01/mopidy,tkem/mopidy,dbrgn/mopidy,liamw9534/mopidy,SuperStarPL/mopidy,diandiankan/mopidy,abarisain/mopidy,glogiotatidis/mopidy,hkariti/mopidy,mopidy/mopidy,mokieyue/mopidy,mopidy/mopidy,vrs01/mopidy,ali/mopidy,ZenithDK/mopidy,tkem/mopidy,SuperStarPL/mopidy,hkariti/mopidy,ali/mopidy,glogiotatidis/mopidy,adamcik/mopidy,jmarsik/mopidy,kingosticks/mopidy,ZenithDK/mopidy,kingosticks/mopidy,pacificIT/mopidy,glogiotatidis/mopidy,priestd09/mopidy,rawdlite/mopidy,bacontext/mopidy,diandiankan/mopidy,SuperStarPL/mopidy,hkariti/mopidy,mopidy/mopidy,kingosticks/mopidy,diandiankan/mopidy,mokieyue/mopidy,rawdlite/mopidy,swak/mopidy,priestd09/mopidy,ali/mopidy,bacontext/mopidy,bencevans/mopidy,jodal/mopidy,quartz55/mopidy,mokieyue/mopidy,jodal/mopidy,bencevans/mopidy,quartz55/mopidy,quartz55/mopidy,dbrgn/mopidy,mokieyue/mopidy,jmarsik/mopidy,rawdlite/mopidy,bencevans/mopidy,vrs01/mopidy,ZenithDK/mopidy,tkem/mopidy,dbrgn/mopidy,hkariti/mopidy,rawdlite/mopidy,priestd09/mopidy,woutervanwijk/mopidy,jmarsik/mopidy,SuperStarPL/mopidy,bacontext/mopidy,dbrgn/mopidy,jodal/mopidy,swak/mopidy,jcass77/mopidy,pacificIT/mopidy,adamcik/mopidy,bencevans/mopidy,ali/mopidy,bacontext/mopidy,swak/mopidy,pacificIT/mopidy,glogiotatidis/mopidy,pacificIT/mopidy,jcass77/mopidy,woutervanwijk/mopidy
from __future__ import unicode_literals import mopidy from mopidy import ext + from mopidy.utils import config, formatting + + default_config = """ + [ext.stream] + + # If the stream extension should be enabled or not + enabled = true + + # Whitelist of URI schemas to support streaming from + protocols = + http + https + mms + rtmp + rtmps + rtsp + """ __doc__ = """A backend for playing music for streaming music. This backend will handle streaming of URIs in :attr:`mopidy.settings.STREAM_PROTOCOLS` assuming the right plugins are installed. **Issues:** https://github.com/mopidy/mopidy/issues?labels=Stream+backend **Dependencies:** - None - **Settings:** + **Default config:** - - :attr:`mopidy.settings.STREAM_PROTOCOLS` - """ + .. code-block:: ini + + %(config)s + """ % {'config': formatting.indent(default_config)} class Extension(ext.Extension): name = 'Mopidy-Stream' version = mopidy.__version__ def get_default_config(self): - return '[ext.stream]' + return default_config - def validate_config(self, config): - pass + def get_config_schema(self): + schema = config.ExtensionConfigSchema() + schema['protocols'] = config.List() + return schema def validate_environment(self): pass def get_backend_classes(self): from .actor import StreamBackend return [StreamBackend]
Add default config and config schema
## Code Before: from __future__ import unicode_literals import mopidy from mopidy import ext __doc__ = """A backend for playing music for streaming music. This backend will handle streaming of URIs in :attr:`mopidy.settings.STREAM_PROTOCOLS` assuming the right plugins are installed. **Issues:** https://github.com/mopidy/mopidy/issues?labels=Stream+backend **Dependencies:** - None **Settings:** - :attr:`mopidy.settings.STREAM_PROTOCOLS` """ class Extension(ext.Extension): name = 'Mopidy-Stream' version = mopidy.__version__ def get_default_config(self): return '[ext.stream]' def validate_config(self, config): pass def validate_environment(self): pass def get_backend_classes(self): from .actor import StreamBackend return [StreamBackend] ## Instruction: Add default config and config schema ## Code After: from __future__ import unicode_literals import mopidy from mopidy import ext from mopidy.utils import config, formatting default_config = """ [ext.stream] # If the stream extension should be enabled or not enabled = true # Whitelist of URI schemas to support streaming from protocols = http https mms rtmp rtmps rtsp """ __doc__ = """A backend for playing music for streaming music. This backend will handle streaming of URIs in :attr:`mopidy.settings.STREAM_PROTOCOLS` assuming the right plugins are installed. **Issues:** https://github.com/mopidy/mopidy/issues?labels=Stream+backend **Dependencies:** - None **Default config:** .. code-block:: ini %(config)s """ % {'config': formatting.indent(default_config)} class Extension(ext.Extension): name = 'Mopidy-Stream' version = mopidy.__version__ def get_default_config(self): return default_config def get_config_schema(self): schema = config.ExtensionConfigSchema() schema['protocols'] = config.List() return schema def validate_environment(self): pass def get_backend_classes(self): from .actor import StreamBackend return [StreamBackend]
// ... existing code ... import mopidy from mopidy import ext from mopidy.utils import config, formatting default_config = """ [ext.stream] # If the stream extension should be enabled or not enabled = true # Whitelist of URI schemas to support streaming from protocols = http https mms rtmp rtmps rtsp """ __doc__ = """A backend for playing music for streaming music. // ... modified code ... - None **Default config:** .. code-block:: ini %(config)s """ % {'config': formatting.indent(default_config)} ... def get_default_config(self): return default_config def get_config_schema(self): schema = config.ExtensionConfigSchema() schema['protocols'] = config.List() return schema def validate_environment(self): // ... rest of the code ...
7b05ce75c0dd16944b26f2c53f1508aa3f771d27
migrations/versions/0177_add_virus_scan_statuses.py
migrations/versions/0177_add_virus_scan_statuses.py
from alembic import op revision = '0176_alter_billing_columns' down_revision = '0175_drop_job_statistics_table' def upgrade(): op.execute("INSERT INTO notification_status_types (name) VALUES ('pending-virus-check')") op.execute("INSERT INTO notification_status_types (name) VALUES ('virus-scan-failed')") def downgrade(): op.execute("UPDATE notifications SET notification_status = 'created' WHERE notification_status = 'pending-virus-check'") op.execute("UPDATE notification_history SET notification_status = 'created' WHERE notification_status = 'pending-virus-check'") op.execute("UPDATE notifications SET notification_status = 'permanent-failure' WHERE notification_status = 'virus-scan-failed'") op.execute("UPDATE notification_history SET notification_status = 'permanent-failure' WHERE notification_status = 'virus-scan-failed'") op.execute("DELETE FROM notification_status_types WHERE name in ('pending-virus-check', 'virus-scan-failed')")
from alembic import op revision = '0177_add_virus_scan_statuses' down_revision = '0176_alter_billing_columns' def upgrade(): op.execute("INSERT INTO notification_status_types (name) VALUES ('pending-virus-check')") op.execute("INSERT INTO notification_status_types (name) VALUES ('virus-scan-failed')") def downgrade(): op.execute("UPDATE notifications SET notification_status = 'created' WHERE notification_status = 'pending-virus-check'") op.execute("UPDATE notification_history SET notification_status = 'created' WHERE notification_status = 'pending-virus-check'") op.execute("UPDATE notifications SET notification_status = 'permanent-failure' WHERE notification_status = 'virus-scan-failed'") op.execute("UPDATE notification_history SET notification_status = 'permanent-failure' WHERE notification_status = 'virus-scan-failed'") op.execute("DELETE FROM notification_status_types WHERE name in ('pending-virus-check', 'virus-scan-failed')")
Fix revision numbers in migration 0177
Fix revision numbers in migration 0177
Python
mit
alphagov/notifications-api,alphagov/notifications-api
from alembic import op + revision = '0177_add_virus_scan_statuses' - revision = '0176_alter_billing_columns' + down_revision = '0176_alter_billing_columns' - down_revision = '0175_drop_job_statistics_table' def upgrade(): op.execute("INSERT INTO notification_status_types (name) VALUES ('pending-virus-check')") op.execute("INSERT INTO notification_status_types (name) VALUES ('virus-scan-failed')") def downgrade(): op.execute("UPDATE notifications SET notification_status = 'created' WHERE notification_status = 'pending-virus-check'") op.execute("UPDATE notification_history SET notification_status = 'created' WHERE notification_status = 'pending-virus-check'") op.execute("UPDATE notifications SET notification_status = 'permanent-failure' WHERE notification_status = 'virus-scan-failed'") op.execute("UPDATE notification_history SET notification_status = 'permanent-failure' WHERE notification_status = 'virus-scan-failed'") op.execute("DELETE FROM notification_status_types WHERE name in ('pending-virus-check', 'virus-scan-failed')")
Fix revision numbers in migration 0177
## Code Before: from alembic import op revision = '0176_alter_billing_columns' down_revision = '0175_drop_job_statistics_table' def upgrade(): op.execute("INSERT INTO notification_status_types (name) VALUES ('pending-virus-check')") op.execute("INSERT INTO notification_status_types (name) VALUES ('virus-scan-failed')") def downgrade(): op.execute("UPDATE notifications SET notification_status = 'created' WHERE notification_status = 'pending-virus-check'") op.execute("UPDATE notification_history SET notification_status = 'created' WHERE notification_status = 'pending-virus-check'") op.execute("UPDATE notifications SET notification_status = 'permanent-failure' WHERE notification_status = 'virus-scan-failed'") op.execute("UPDATE notification_history SET notification_status = 'permanent-failure' WHERE notification_status = 'virus-scan-failed'") op.execute("DELETE FROM notification_status_types WHERE name in ('pending-virus-check', 'virus-scan-failed')") ## Instruction: Fix revision numbers in migration 0177 ## Code After: from alembic import op revision = '0177_add_virus_scan_statuses' down_revision = '0176_alter_billing_columns' def upgrade(): op.execute("INSERT INTO notification_status_types (name) VALUES ('pending-virus-check')") op.execute("INSERT INTO notification_status_types (name) VALUES ('virus-scan-failed')") def downgrade(): op.execute("UPDATE notifications SET notification_status = 'created' WHERE notification_status = 'pending-virus-check'") op.execute("UPDATE notification_history SET notification_status = 'created' WHERE notification_status = 'pending-virus-check'") op.execute("UPDATE notifications SET notification_status = 'permanent-failure' WHERE notification_status = 'virus-scan-failed'") op.execute("UPDATE notification_history SET notification_status = 'permanent-failure' WHERE notification_status = 'virus-scan-failed'") op.execute("DELETE FROM notification_status_types WHERE name in ('pending-virus-check', 'virus-scan-failed')")
# ... existing code ... revision = '0177_add_virus_scan_statuses' down_revision = '0176_alter_billing_columns' # ... rest of the code ...
38b4af0b3c1c6105d68ff453d86107758ef9d751
preconditions.py
preconditions.py
class PreconditionError (TypeError): pass def preconditions(*precs): def decorate(f): def g(*a, **kw): return f(*a, **kw) return g return decorate
import inspect class PreconditionError (TypeError): pass def preconditions(*precs): precinfo = [] for p in precs: spec = inspect.getargspec(p) if spec.varargs or spec.keywords: raise PreconditionError( 'Precondition {!r} must not accept * nor ** args.'.format(p)) i = -len(spec.defaults) appargs, closureargs = spec.args[:i], spec.args[i:] precinfo.append( (appargs, closureargs, p) ) def decorate(f): def g(*a, **kw): return f(*a, **kw) return g return decorate
Implement two of the "early" InvalidPreconditionTests which can be checked prior to seeing the wrapping function.
Implement two of the "early" InvalidPreconditionTests which can be checked prior to seeing the wrapping function.
Python
mit
nejucomo/preconditions
+ import inspect + + class PreconditionError (TypeError): pass def preconditions(*precs): + + precinfo = [] + for p in precs: + spec = inspect.getargspec(p) + if spec.varargs or spec.keywords: + raise PreconditionError( + 'Precondition {!r} must not accept * nor ** args.'.format(p)) + + i = -len(spec.defaults) + appargs, closureargs = spec.args[:i], spec.args[i:] + precinfo.append( (appargs, closureargs, p) ) + def decorate(f): def g(*a, **kw): return f(*a, **kw) return g return decorate
Implement two of the "early" InvalidPreconditionTests which can be checked prior to seeing the wrapping function.
## Code Before: class PreconditionError (TypeError): pass def preconditions(*precs): def decorate(f): def g(*a, **kw): return f(*a, **kw) return g return decorate ## Instruction: Implement two of the "early" InvalidPreconditionTests which can be checked prior to seeing the wrapping function. ## Code After: import inspect class PreconditionError (TypeError): pass def preconditions(*precs): precinfo = [] for p in precs: spec = inspect.getargspec(p) if spec.varargs or spec.keywords: raise PreconditionError( 'Precondition {!r} must not accept * nor ** args.'.format(p)) i = -len(spec.defaults) appargs, closureargs = spec.args[:i], spec.args[i:] precinfo.append( (appargs, closureargs, p) ) def decorate(f): def g(*a, **kw): return f(*a, **kw) return g return decorate
... import inspect class PreconditionError (TypeError): pass ... def preconditions(*precs): precinfo = [] for p in precs: spec = inspect.getargspec(p) if spec.varargs or spec.keywords: raise PreconditionError( 'Precondition {!r} must not accept * nor ** args.'.format(p)) i = -len(spec.defaults) appargs, closureargs = spec.args[:i], spec.args[i:] precinfo.append( (appargs, closureargs, p) ) def decorate(f): def g(*a, **kw): ...
45116fc996b097176bcfa2dcd7fb8c9710f6d66e
tests/test_basics.py
tests/test_basics.py
import os from xml.etree import ElementTree from utils import with_app, pretty_print_xml #============================================================================= # Tests @with_app(buildername="xml", srcdir="basics") def test_basics(app, status, warning): app.build() tree = ElementTree.parse(app.outdir / "index.xml") pretty_print_xml(tree.getroot()) # Verify that 2 traceables are found. assert len(tree.findall(".//target")) == 2 assert len(tree.findall(".//index")) == 2 assert len(tree.findall(".//admonition")) == 2 assert len(tree.findall(".//admonition")) == 2 # Verify that child-parent relationship are made. assert len(tree.findall(".//field_list")) == 2 parent_fields, child_fields = tree.findall(".//field_list") for field in parent_fields: field_name = field.findall("./field_name")[0] if field_name.text == "child": break else: assert False, "Parent's child field not found!" for field in child_fields: field_name = field.findall("./field_name")[0] if field_name.text == "parent": break else: assert False, "Child's parent field not found!" # Verify that a warning is emitted for unknown traceable tag. assert (warning.getvalue().find( "WARNING: Traceables: no traceable with tag" " 'NONEXISTENT' found!") > 0)
import os from xml.etree import ElementTree from utils import with_app, pretty_print_xml #============================================================================= # Tests @with_app(buildername="xml", srcdir="basics") def test_basics(app, status, warning): app.build() tree = ElementTree.parse(app.outdir / "index.xml") # Verify that 2 traceables are found. assert len(tree.findall(".//target")) == 2 assert len(tree.findall(".//index")) == 2 assert len(tree.findall(".//admonition")) == 2 assert len(tree.findall(".//admonition")) == 2 # Verify that child-parent relationship are made. assert len(tree.findall(".//field_list")) == 2 parent_fields, child_fields = tree.findall(".//field_list") for field in parent_fields: field_name = field.findall("./field_name")[0] if field_name.text == "child": break else: assert False, "Parent's child field not found!" for field in child_fields: field_name = field.findall("./field_name")[0] if field_name.text == "parent": break else: assert False, "Child's parent field not found!" # Verify that a warning is emitted for unknown traceable tag. assert (warning.getvalue().find( "WARNING: Traceables: no traceable with tag" " 'NONEXISTENT' found!") > 0)
Remove debug printing from test case
Remove debug printing from test case
Python
apache-2.0
t4ngo/sphinxcontrib-traceables
import os from xml.etree import ElementTree from utils import with_app, pretty_print_xml #============================================================================= # Tests @with_app(buildername="xml", srcdir="basics") def test_basics(app, status, warning): app.build() tree = ElementTree.parse(app.outdir / "index.xml") - pretty_print_xml(tree.getroot()) # Verify that 2 traceables are found. assert len(tree.findall(".//target")) == 2 assert len(tree.findall(".//index")) == 2 assert len(tree.findall(".//admonition")) == 2 assert len(tree.findall(".//admonition")) == 2 # Verify that child-parent relationship are made. assert len(tree.findall(".//field_list")) == 2 parent_fields, child_fields = tree.findall(".//field_list") for field in parent_fields: field_name = field.findall("./field_name")[0] if field_name.text == "child": break else: assert False, "Parent's child field not found!" for field in child_fields: field_name = field.findall("./field_name")[0] if field_name.text == "parent": break else: assert False, "Child's parent field not found!" # Verify that a warning is emitted for unknown traceable tag. assert (warning.getvalue().find( "WARNING: Traceables: no traceable with tag" " 'NONEXISTENT' found!") > 0)
Remove debug printing from test case
## Code Before: import os from xml.etree import ElementTree from utils import with_app, pretty_print_xml #============================================================================= # Tests @with_app(buildername="xml", srcdir="basics") def test_basics(app, status, warning): app.build() tree = ElementTree.parse(app.outdir / "index.xml") pretty_print_xml(tree.getroot()) # Verify that 2 traceables are found. assert len(tree.findall(".//target")) == 2 assert len(tree.findall(".//index")) == 2 assert len(tree.findall(".//admonition")) == 2 assert len(tree.findall(".//admonition")) == 2 # Verify that child-parent relationship are made. assert len(tree.findall(".//field_list")) == 2 parent_fields, child_fields = tree.findall(".//field_list") for field in parent_fields: field_name = field.findall("./field_name")[0] if field_name.text == "child": break else: assert False, "Parent's child field not found!" for field in child_fields: field_name = field.findall("./field_name")[0] if field_name.text == "parent": break else: assert False, "Child's parent field not found!" # Verify that a warning is emitted for unknown traceable tag. assert (warning.getvalue().find( "WARNING: Traceables: no traceable with tag" " 'NONEXISTENT' found!") > 0) ## Instruction: Remove debug printing from test case ## Code After: import os from xml.etree import ElementTree from utils import with_app, pretty_print_xml #============================================================================= # Tests @with_app(buildername="xml", srcdir="basics") def test_basics(app, status, warning): app.build() tree = ElementTree.parse(app.outdir / "index.xml") # Verify that 2 traceables are found. assert len(tree.findall(".//target")) == 2 assert len(tree.findall(".//index")) == 2 assert len(tree.findall(".//admonition")) == 2 assert len(tree.findall(".//admonition")) == 2 # Verify that child-parent relationship are made. assert len(tree.findall(".//field_list")) == 2 parent_fields, child_fields = tree.findall(".//field_list") for field in parent_fields: field_name = field.findall("./field_name")[0] if field_name.text == "child": break else: assert False, "Parent's child field not found!" for field in child_fields: field_name = field.findall("./field_name")[0] if field_name.text == "parent": break else: assert False, "Child's parent field not found!" # Verify that a warning is emitted for unknown traceable tag. assert (warning.getvalue().find( "WARNING: Traceables: no traceable with tag" " 'NONEXISTENT' found!") > 0)
... app.build() tree = ElementTree.parse(app.outdir / "index.xml") # Verify that 2 traceables are found. ...
8932d0717bf57c86b81b6744353d6387821b8b15
wsgi/setup.py
wsgi/setup.py
import subprocess import sys import setup_util import os def start(args): subprocess.Popen("gunicorn hello:app -b 0.0.0.0:8080 -w " + str((args.max_threads * 2)) + " --log-level=critical", shell=True, cwd="wsgi") return 0 def stop(): p = subprocess.Popen(['ps', 'aux'], stdout=subprocess.PIPE) out, err = p.communicate() for line in out.splitlines(): if 'gunicorn' in line: try: pid = int(line.split(None, 2)[1]) os.kill(pid, 9) except OSError: pass return 0
import subprocess import sys import setup_util import os def start(args): subprocess.Popen('gunicorn hello:app --worker-class="egg:meinheld#gunicorn_worker" -b 0.0.0.0:8080 -w ' + str((args.max_threads * 2)) + " --log-level=critical", shell=True, cwd="wsgi") return 0 def stop(): p = subprocess.Popen(['ps', 'aux'], stdout=subprocess.PIPE) out, err = p.communicate() for line in out.splitlines(): if 'gunicorn' in line: try: pid = int(line.split(None, 2)[1]) os.kill(pid, 9) except OSError: pass return 0
Use meinheld worker (same as other Python Frameworks)
wsgi: Use meinheld worker (same as other Python Frameworks)
Python
bsd-3-clause
jamming/FrameworkBenchmarks,s-ludwig/FrameworkBenchmarks,Ocramius/FrameworkBenchmarks,diablonhn/FrameworkBenchmarks,ashawnbandy-te-tfb/FrameworkBenchmarks,zloster/FrameworkBenchmarks,donovanmuller/FrameworkBenchmarks,zdanek/FrameworkBenchmarks,markkolich/FrameworkBenchmarks,circlespainter/FrameworkBenchmarks,saturday06/FrameworkBenchmarks,sgml/FrameworkBenchmarks,mfirry/FrameworkBenchmarks,nathana1/FrameworkBenchmarks,s-ludwig/FrameworkBenchmarks,victorbriz/FrameworkBenchmarks,ratpack/FrameworkBenchmarks,diablonhn/FrameworkBenchmarks,s-ludwig/FrameworkBenchmarks,jetty-project/FrameworkBenchmarks,ashawnbandy-te-tfb/FrameworkBenchmarks,nathana1/FrameworkBenchmarks,leafo/FrameworkBenchmarks,zane-techempower/FrameworkBenchmarks,greg-hellings/FrameworkBenchmarks,Synchro/FrameworkBenchmarks,F3Community/FrameworkBenchmarks,knewmanTE/FrameworkBenchmarks,sgml/FrameworkBenchmarks,circlespainter/FrameworkBenchmarks,xitrum-framework/FrameworkBenchmarks,ratpack/FrameworkBenchmarks,marko-asplund/FrameworkBenchmarks,nathana1/FrameworkBenchmarks,F3Community/FrameworkBenchmarks,kbrock/FrameworkBenchmarks,nbrady-techempower/FrameworkBenchmarks,jeevatkm/FrameworkBenchmarks,julienschmidt/FrameworkBenchmarks,alubbe/FrameworkBenchmarks,joshk/FrameworkBenchmarks,zdanek/FrameworkBenchmarks,alubbe/FrameworkBenchmarks,saturday06/FrameworkBenchmarks,youprofit/FrameworkBenchmarks,actframework/FrameworkBenchmarks,doom369/FrameworkBenchmarks,saturday06/FrameworkBenchmarks,joshk/FrameworkBenchmarks,greenlaw110/FrameworkBenchmarks,denkab/FrameworkBenchmarks,martin-g/FrameworkBenchmarks,waiteb3/FrameworkBenchmarks,marko-asplund/FrameworkBenchmarks,Eyepea/FrameworkBenchmarks,zloster/FrameworkBenchmarks,PermeAgility/FrameworkBenchmarks,Ocramius/FrameworkBenchmarks,RockinRoel/FrameworkBenchmarks,knewmanTE/FrameworkBenchmarks,jetty-project/FrameworkBenchmarks,nbrady-techempower/FrameworkBenchmarks,stefanocasazza/FrameworkBenchmarks,fabianmurariu/FrameworkBenchmarks,stefanocasazza/FrameworkBenchmarks,lcp0578/FrameworkBenchmarks,zloster/FrameworkBenchmarks,marko-asplund/FrameworkBenchmarks,sxend/FrameworkBenchmarks,jamming/FrameworkBenchmarks,zdanek/FrameworkBenchmarks,Eyepea/FrameworkBenchmarks,herloct/FrameworkBenchmarks,F3Community/FrameworkBenchmarks,grob/FrameworkBenchmarks,grob/FrameworkBenchmarks,hamiltont/FrameworkBenchmarks,mfirry/FrameworkBenchmarks,raziel057/FrameworkBenchmarks,khellang/FrameworkBenchmarks,stefanocasazza/FrameworkBenchmarks,sagenschneider/FrameworkBenchmarks,herloct/FrameworkBenchmarks,donovanmuller/FrameworkBenchmarks,hamiltont/FrameworkBenchmarks,F3Community/FrameworkBenchmarks,zapov/FrameworkBenchmarks,Eyepea/FrameworkBenchmarks,actframework/FrameworkBenchmarks,Ocramius/FrameworkBenchmarks,zapov/FrameworkBenchmarks,knewmanTE/FrameworkBenchmarks,greenlaw110/FrameworkBenchmarks,k-r-g/FrameworkBenchmarks,doom369/FrameworkBenchmarks,steveklabnik/FrameworkBenchmarks,saturday06/FrameworkBenchmarks,sgml/FrameworkBenchmarks,denkab/FrameworkBenchmarks,testn/FrameworkBenchmarks,denkab/FrameworkBenchmarks,psfblair/FrameworkBenchmarks,xitrum-framework/FrameworkBenchmarks,PermeAgility/FrameworkBenchmarks,nbrady-techempower/FrameworkBenchmarks,circlespainter/FrameworkBenchmarks,testn/FrameworkBenchmarks,s-ludwig/FrameworkBenchmarks,ashawnbandy-te-tfb/FrameworkBenchmarks,markkolich/FrameworkBenchmarks,jaguililla/FrameworkBenchmarks,zdanek/FrameworkBenchmarks,Verber/FrameworkBenchmarks,zloster/FrameworkBenchmarks,psfblair/FrameworkBenchmarks,zdanek/FrameworkBenchmarks,kostya-sh/FrameworkBenchmarks,greenlaw110/FrameworkBenchmarks,torhve/FrameworkBenchmarks,nathana1/FrameworkBenchmarks,Rayne/FrameworkBenchmarks,youprofit/FrameworkBenchmarks,diablonhn/FrameworkBenchmarks,raziel057/FrameworkBenchmarks,Jesterovskiy/FrameworkBenchmarks,grob/FrameworkBenchmarks,testn/FrameworkBenchmarks,Verber/FrameworkBenchmarks,nkasvosve/FrameworkBenchmarks,nkasvosve/FrameworkBenchmarks,methane/FrameworkBenchmarks,k-r-g/FrameworkBenchmarks,seem-sky/FrameworkBenchmarks,testn/FrameworkBenchmarks,Jesterovskiy/FrameworkBenchmarks,Rayne/FrameworkBenchmarks,Verber/FrameworkBenchmarks,psfblair/FrameworkBenchmarks,Eyepea/FrameworkBenchmarks,zloster/FrameworkBenchmarks,sanjoydesk/FrameworkBenchmarks,jebbstewart/FrameworkBenchmarks,alubbe/FrameworkBenchmarks,dmacd/FB-try1,kellabyte/FrameworkBenchmarks,nkasvosve/FrameworkBenchmarks,nathana1/FrameworkBenchmarks,actframework/FrameworkBenchmarks,knewmanTE/FrameworkBenchmarks,zane-techempower/FrameworkBenchmarks,marko-asplund/FrameworkBenchmarks,seem-sky/FrameworkBenchmarks,zapov/FrameworkBenchmarks,RockinRoel/FrameworkBenchmarks,diablonhn/FrameworkBenchmarks,zloster/FrameworkBenchmarks,steveklabnik/FrameworkBenchmarks,dmacd/FB-try1,alubbe/FrameworkBenchmarks,Rayne/FrameworkBenchmarks,grob/FrameworkBenchmarks,zloster/FrameworkBenchmarks,Dith3r/FrameworkBenchmarks,lcp0578/FrameworkBenchmarks,RockinRoel/FrameworkBenchmarks,sagenschneider/FrameworkBenchmarks,donovanmuller/FrameworkBenchmarks,saturday06/FrameworkBenchmarks,steveklabnik/FrameworkBenchmarks,PermeAgility/FrameworkBenchmarks,jetty-project/FrameworkBenchmarks,Verber/FrameworkBenchmarks,denkab/FrameworkBenchmarks,mfirry/FrameworkBenchmarks,seem-sky/FrameworkBenchmarks,ratpack/FrameworkBenchmarks,nbrady-techempower/FrameworkBenchmarks,zapov/FrameworkBenchmarks,Eyepea/FrameworkBenchmarks,stefanocasazza/FrameworkBenchmarks,testn/FrameworkBenchmarks,lcp0578/FrameworkBenchmarks,valyala/FrameworkBenchmarks,julienschmidt/FrameworkBenchmarks,torhve/FrameworkBenchmarks,valyala/FrameworkBenchmarks,ashawnbandy-te-tfb/FrameworkBenchmarks,joshk/FrameworkBenchmarks,MTDdk/FrameworkBenchmarks,denkab/FrameworkBenchmarks,Jesterovskiy/FrameworkBenchmarks,thousandsofthem/FrameworkBenchmarks,steveklabnik/FrameworkBenchmarks,diablonhn/FrameworkBenchmarks,steveklabnik/FrameworkBenchmarks,kellabyte/FrameworkBenchmarks,dmacd/FB-try1,valyala/FrameworkBenchmarks,doom369/FrameworkBenchmarks,steveklabnik/FrameworkBenchmarks,jeevatkm/FrameworkBenchmarks,herloct/FrameworkBenchmarks,waiteb3/FrameworkBenchmarks,markkolich/FrameworkBenchmarks,nathana1/FrameworkBenchmarks,waiteb3/FrameworkBenchmarks,jamming/FrameworkBenchmarks,mfirry/FrameworkBenchmarks,MTDdk/FrameworkBenchmarks,greg-hellings/FrameworkBenchmarks,zhuochenKIDD/FrameworkBenchmarks,jamming/FrameworkBenchmarks,diablonhn/FrameworkBenchmarks,Ocramius/FrameworkBenchmarks,denkab/FrameworkBenchmarks,nkasvosve/FrameworkBenchmarks,zapov/FrameworkBenchmarks,donovanmuller/FrameworkBenchmarks,psfblair/FrameworkBenchmarks,jaguililla/FrameworkBenchmarks,greenlaw110/FrameworkBenchmarks,Rydgel/FrameworkBenchmarks,alubbe/FrameworkBenchmarks,yunspace/FrameworkBenchmarks,jaguililla/FrameworkBenchmarks,hperadin/FrameworkBenchmarks,hperadin/FrameworkBenchmarks,nbrady-techempower/FrameworkBenchmarks,nathana1/FrameworkBenchmarks,jebbstewart/FrameworkBenchmarks,Ocramius/FrameworkBenchmarks,nkasvosve/FrameworkBenchmarks,denkab/FrameworkBenchmarks,greg-hellings/FrameworkBenchmarks,jebbstewart/FrameworkBenchmarks,xitrum-framework/FrameworkBenchmarks,PermeAgility/FrameworkBenchmarks,markkolich/FrameworkBenchmarks,hamiltont/FrameworkBenchmarks,Dith3r/FrameworkBenchmarks,xitrum-framework/FrameworkBenchmarks,testn/FrameworkBenchmarks,sxend/FrameworkBenchmarks,kbrock/FrameworkBenchmarks,kostya-sh/FrameworkBenchmarks,doom369/FrameworkBenchmarks,nathana1/FrameworkBenchmarks,raziel057/FrameworkBenchmarks,knewmanTE/FrameworkBenchmarks,hamiltont/FrameworkBenchmarks,psfblair/FrameworkBenchmarks,raziel057/FrameworkBenchmarks,alubbe/FrameworkBenchmarks,Ocramius/FrameworkBenchmarks,alubbe/FrameworkBenchmarks,PermeAgility/FrameworkBenchmarks,k-r-g/FrameworkBenchmarks,methane/FrameworkBenchmarks,waiteb3/FrameworkBenchmarks,RockinRoel/FrameworkBenchmarks,jetty-project/FrameworkBenchmarks,circlespainter/FrameworkBenchmarks,F3Community/FrameworkBenchmarks,nathana1/FrameworkBenchmarks,k-r-g/FrameworkBenchmarks,jaguililla/FrameworkBenchmarks,lcp0578/FrameworkBenchmarks,Ocramius/FrameworkBenchmarks,ashawnbandy-te-tfb/FrameworkBenchmarks,kellabyte/FrameworkBenchmarks,s-ludwig/FrameworkBenchmarks,jaguililla/FrameworkBenchmarks,martin-g/FrameworkBenchmarks,khellang/FrameworkBenchmarks,khellang/FrameworkBenchmarks,grob/FrameworkBenchmarks,greg-hellings/FrameworkBenchmarks,methane/FrameworkBenchmarks,jeevatkm/FrameworkBenchmarks,khellang/FrameworkBenchmarks,seem-sky/FrameworkBenchmarks,hamiltont/FrameworkBenchmarks,sanjoydesk/FrameworkBenchmarks,dmacd/FB-try1,sanjoydesk/FrameworkBenchmarks,victorbriz/FrameworkBenchmarks,Eyepea/FrameworkBenchmarks,sxend/FrameworkBenchmarks,testn/FrameworkBenchmarks,victorbriz/FrameworkBenchmarks,fabianmurariu/FrameworkBenchmarks,alubbe/FrameworkBenchmarks,greg-hellings/FrameworkBenchmarks,diablonhn/FrameworkBenchmarks,Dith3r/FrameworkBenchmarks,yunspace/FrameworkBenchmarks,Synchro/FrameworkBenchmarks,grob/FrameworkBenchmarks,donovanmuller/FrameworkBenchmarks,herloct/FrameworkBenchmarks,donovanmuller/FrameworkBenchmarks,yunspace/FrameworkBenchmarks,jeevatkm/FrameworkBenchmarks,steveklabnik/FrameworkBenchmarks,Rayne/FrameworkBenchmarks,sanjoydesk/FrameworkBenchmarks,nkasvosve/FrameworkBenchmarks,sagenschneider/FrameworkBenchmarks,herloct/FrameworkBenchmarks,zhuochenKIDD/FrameworkBenchmarks,psfblair/FrameworkBenchmarks,marko-asplund/FrameworkBenchmarks,thousandsofthem/FrameworkBenchmarks,zloster/FrameworkBenchmarks,jetty-project/FrameworkBenchmarks,hamiltont/FrameworkBenchmarks,zane-techempower/FrameworkBenchmarks,torhve/FrameworkBenchmarks,herloct/FrameworkBenchmarks,knewmanTE/FrameworkBenchmarks,zapov/FrameworkBenchmarks,victorbriz/FrameworkBenchmarks,kellabyte/FrameworkBenchmarks,sagenschneider/FrameworkBenchmarks,zhuochenKIDD/FrameworkBenchmarks,diablonhn/FrameworkBenchmarks,zhuochenKIDD/FrameworkBenchmarks,F3Community/FrameworkBenchmarks,jeevatkm/FrameworkBenchmarks,ashawnbandy-te-tfb/FrameworkBenchmarks,jetty-project/FrameworkBenchmarks,zane-techempower/FrameworkBenchmarks,leafo/FrameworkBenchmarks,nathana1/FrameworkBenchmarks,sagenschneider/FrameworkBenchmarks,greg-hellings/FrameworkBenchmarks,zloster/FrameworkBenchmarks,Eyepea/FrameworkBenchmarks,Rydgel/FrameworkBenchmarks,PermeAgility/FrameworkBenchmarks,ashawnbandy-te-tfb/FrameworkBenchmarks,sagenschneider/FrameworkBenchmarks,Verber/FrameworkBenchmarks,martin-g/FrameworkBenchmarks,sxend/FrameworkBenchmarks,yunspace/FrameworkBenchmarks,stefanocasazza/FrameworkBenchmarks,methane/FrameworkBenchmarks,julienschmidt/FrameworkBenchmarks,hperadin/FrameworkBenchmarks,seem-sky/FrameworkBenchmarks,ashawnbandy-te-tfb/FrameworkBenchmarks,zane-techempower/FrameworkBenchmarks,s-ludwig/FrameworkBenchmarks,zloster/FrameworkBenchmarks,F3Community/FrameworkBenchmarks,Dith3r/FrameworkBenchmarks,thousandsofthem/FrameworkBenchmarks,mfirry/FrameworkBenchmarks,MTDdk/FrameworkBenchmarks,raziel057/FrameworkBenchmarks,Synchro/FrameworkBenchmarks,Verber/FrameworkBenchmarks,kostya-sh/FrameworkBenchmarks,ratpack/FrameworkBenchmarks,stefanocasazza/FrameworkBenchmarks,greenlaw110/FrameworkBenchmarks,kostya-sh/FrameworkBenchmarks,fabianmurariu/FrameworkBenchmarks,knewmanTE/FrameworkBenchmarks,xitrum-framework/FrameworkBenchmarks,stefanocasazza/FrameworkBenchmarks,Synchro/FrameworkBenchmarks,nbrady-techempower/FrameworkBenchmarks,joshk/FrameworkBenchmarks,ratpack/FrameworkBenchmarks,mfirry/FrameworkBenchmarks,torhve/FrameworkBenchmarks,raziel057/FrameworkBenchmarks,victorbriz/FrameworkBenchmarks,Dith3r/FrameworkBenchmarks,xitrum-framework/FrameworkBenchmarks,kostya-sh/FrameworkBenchmarks,Eyepea/FrameworkBenchmarks,waiteb3/FrameworkBenchmarks,zane-techempower/FrameworkBenchmarks,actframework/FrameworkBenchmarks,steveklabnik/FrameworkBenchmarks,seem-sky/FrameworkBenchmarks,martin-g/FrameworkBenchmarks,zdanek/FrameworkBenchmarks,stefanocasazza/FrameworkBenchmarks,jetty-project/FrameworkBenchmarks,jeevatkm/FrameworkBenchmarks,diablonhn/FrameworkBenchmarks,RockinRoel/FrameworkBenchmarks,sxend/FrameworkBenchmarks,jeevatkm/FrameworkBenchmarks,Synchro/FrameworkBenchmarks,julienschmidt/FrameworkBenchmarks,jeevatkm/FrameworkBenchmarks,jeevatkm/FrameworkBenchmarks,hperadin/FrameworkBenchmarks,jamming/FrameworkBenchmarks,MTDdk/FrameworkBenchmarks,waiteb3/FrameworkBenchmarks,fabianmurariu/FrameworkBenchmarks,k-r-g/FrameworkBenchmarks,xitrum-framework/FrameworkBenchmarks,sgml/FrameworkBenchmarks,martin-g/FrameworkBenchmarks,PermeAgility/FrameworkBenchmarks,RockinRoel/FrameworkBenchmarks,RockinRoel/FrameworkBenchmarks,leafo/FrameworkBenchmarks,zdanek/FrameworkBenchmarks,methane/FrameworkBenchmarks,fabianmurariu/FrameworkBenchmarks,Ocramius/FrameworkBenchmarks,youprofit/FrameworkBenchmarks,seem-sky/FrameworkBenchmarks,herloct/FrameworkBenchmarks,sxend/FrameworkBenchmarks,nkasvosve/FrameworkBenchmarks,ratpack/FrameworkBenchmarks,testn/FrameworkBenchmarks,stefanocasazza/FrameworkBenchmarks,knewmanTE/FrameworkBenchmarks,nbrady-techempower/FrameworkBenchmarks,nbrady-techempower/FrameworkBenchmarks,zane-techempower/FrameworkBenchmarks,MTDdk/FrameworkBenchmarks,RockinRoel/FrameworkBenchmarks,herloct/FrameworkBenchmarks,zdanek/FrameworkBenchmarks,jeevatkm/FrameworkBenchmarks,psfblair/FrameworkBenchmarks,saturday06/FrameworkBenchmarks,doom369/FrameworkBenchmarks,sxend/FrameworkBenchmarks,saturday06/FrameworkBenchmarks,grob/FrameworkBenchmarks,Ocramius/FrameworkBenchmarks,Jesterovskiy/FrameworkBenchmarks,lcp0578/FrameworkBenchmarks,kostya-sh/FrameworkBenchmarks,actframework/FrameworkBenchmarks,ratpack/FrameworkBenchmarks,s-ludwig/FrameworkBenchmarks,circlespainter/FrameworkBenchmarks,PermeAgility/FrameworkBenchmarks,saturday06/FrameworkBenchmarks,thousandsofthem/FrameworkBenchmarks,sanjoydesk/FrameworkBenchmarks,yunspace/FrameworkBenchmarks,markkolich/FrameworkBenchmarks,nkasvosve/FrameworkBenchmarks,Jesterovskiy/FrameworkBenchmarks,waiteb3/FrameworkBenchmarks,sgml/FrameworkBenchmarks,jeevatkm/FrameworkBenchmarks,victorbriz/FrameworkBenchmarks,F3Community/FrameworkBenchmarks,herloct/FrameworkBenchmarks,greenlaw110/FrameworkBenchmarks,waiteb3/FrameworkBenchmarks,Dith3r/FrameworkBenchmarks,Eyepea/FrameworkBenchmarks,Rydgel/FrameworkBenchmarks,actframework/FrameworkBenchmarks,Rydgel/FrameworkBenchmarks,sxend/FrameworkBenchmarks,yunspace/FrameworkBenchmarks,youprofit/FrameworkBenchmarks,youprofit/FrameworkBenchmarks,knewmanTE/FrameworkBenchmarks,hperadin/FrameworkBenchmarks,zloster/FrameworkBenchmarks,fabianmurariu/FrameworkBenchmarks,lcp0578/FrameworkBenchmarks,Verber/FrameworkBenchmarks,sgml/FrameworkBenchmarks,donovanmuller/FrameworkBenchmarks,kostya-sh/FrameworkBenchmarks,Eyepea/FrameworkBenchmarks,doom369/FrameworkBenchmarks,julienschmidt/FrameworkBenchmarks,testn/FrameworkBenchmarks,Eyepea/FrameworkBenchmarks,Jesterovskiy/FrameworkBenchmarks,Verber/FrameworkBenchmarks,grob/FrameworkBenchmarks,PermeAgility/FrameworkBenchmarks,herloct/FrameworkBenchmarks,donovanmuller/FrameworkBenchmarks,markkolich/FrameworkBenchmarks,Rayne/FrameworkBenchmarks,ashawnbandy-te-tfb/FrameworkBenchmarks,methane/FrameworkBenchmarks,Verber/FrameworkBenchmarks,denkab/FrameworkBenchmarks,markkolich/FrameworkBenchmarks,nkasvosve/FrameworkBenchmarks,saturday06/FrameworkBenchmarks,mfirry/FrameworkBenchmarks,khellang/FrameworkBenchmarks,Synchro/FrameworkBenchmarks,k-r-g/FrameworkBenchmarks,valyala/FrameworkBenchmarks,youprofit/FrameworkBenchmarks,Rayne/FrameworkBenchmarks,seem-sky/FrameworkBenchmarks,julienschmidt/FrameworkBenchmarks,Rayne/FrameworkBenchmarks,jebbstewart/FrameworkBenchmarks,zapov/FrameworkBenchmarks,kbrock/FrameworkBenchmarks,donovanmuller/FrameworkBenchmarks,kbrock/FrameworkBenchmarks,raziel057/FrameworkBenchmarks,diablonhn/FrameworkBenchmarks,greg-hellings/FrameworkBenchmarks,hperadin/FrameworkBenchmarks,kostya-sh/FrameworkBenchmarks,youprofit/FrameworkBenchmarks,testn/FrameworkBenchmarks,testn/FrameworkBenchmarks,psfblair/FrameworkBenchmarks,Dith3r/FrameworkBenchmarks,stefanocasazza/FrameworkBenchmarks,jaguililla/FrameworkBenchmarks,F3Community/FrameworkBenchmarks,torhve/FrameworkBenchmarks,jetty-project/FrameworkBenchmarks,kellabyte/FrameworkBenchmarks,Rydgel/FrameworkBenchmarks,martin-g/FrameworkBenchmarks,zapov/FrameworkBenchmarks,circlespainter/FrameworkBenchmarks,marko-asplund/FrameworkBenchmarks,xitrum-framework/FrameworkBenchmarks,stefanocasazza/FrameworkBenchmarks,seem-sky/FrameworkBenchmarks,mfirry/FrameworkBenchmarks,alubbe/FrameworkBenchmarks,khellang/FrameworkBenchmarks,methane/FrameworkBenchmarks,s-ludwig/FrameworkBenchmarks,Rydgel/FrameworkBenchmarks,raziel057/FrameworkBenchmarks,leafo/FrameworkBenchmarks,mfirry/FrameworkBenchmarks,kellabyte/FrameworkBenchmarks,thousandsofthem/FrameworkBenchmarks,sagenschneider/FrameworkBenchmarks,khellang/FrameworkBenchmarks,jebbstewart/FrameworkBenchmarks,thousandsofthem/FrameworkBenchmarks,nkasvosve/FrameworkBenchmarks,khellang/FrameworkBenchmarks,dmacd/FB-try1,martin-g/FrameworkBenchmarks,yunspace/FrameworkBenchmarks,Rydgel/FrameworkBenchmarks,nkasvosve/FrameworkBenchmarks,sagenschneider/FrameworkBenchmarks,jamming/FrameworkBenchmarks,jetty-project/FrameworkBenchmarks,ratpack/FrameworkBenchmarks,markkolich/FrameworkBenchmarks,greenlaw110/FrameworkBenchmarks,nbrady-techempower/FrameworkBenchmarks,greenlaw110/FrameworkBenchmarks,actframework/FrameworkBenchmarks,Rydgel/FrameworkBenchmarks,xitrum-framework/FrameworkBenchmarks,raziel057/FrameworkBenchmarks,Eyepea/FrameworkBenchmarks,RockinRoel/FrameworkBenchmarks,stefanocasazza/FrameworkBenchmarks,mfirry/FrameworkBenchmarks,martin-g/FrameworkBenchmarks,joshk/FrameworkBenchmarks,denkab/FrameworkBenchmarks,psfblair/FrameworkBenchmarks,circlespainter/FrameworkBenchmarks,youprofit/FrameworkBenchmarks,thousandsofthem/FrameworkBenchmarks,seem-sky/FrameworkBenchmarks,marko-asplund/FrameworkBenchmarks,jaguililla/FrameworkBenchmarks,sanjoydesk/FrameworkBenchmarks,khellang/FrameworkBenchmarks,yunspace/FrameworkBenchmarks,sxend/FrameworkBenchmarks,Jesterovskiy/FrameworkBenchmarks,leafo/FrameworkBenchmarks,MTDdk/FrameworkBenchmarks,dmacd/FB-try1,kostya-sh/FrameworkBenchmarks,PermeAgility/FrameworkBenchmarks,khellang/FrameworkBenchmarks,psfblair/FrameworkBenchmarks,sanjoydesk/FrameworkBenchmarks,methane/FrameworkBenchmarks,jaguililla/FrameworkBenchmarks,kbrock/FrameworkBenchmarks,markkolich/FrameworkBenchmarks,Dith3r/FrameworkBenchmarks,s-ludwig/FrameworkBenchmarks,julienschmidt/FrameworkBenchmarks,grob/FrameworkBenchmarks,PermeAgility/FrameworkBenchmarks,zhuochenKIDD/FrameworkBenchmarks,jebbstewart/FrameworkBenchmarks,torhve/FrameworkBenchmarks,MTDdk/FrameworkBenchmarks,jaguililla/FrameworkBenchmarks,Jesterovskiy/FrameworkBenchmarks,marko-asplund/FrameworkBenchmarks,sanjoydesk/FrameworkBenchmarks,sgml/FrameworkBenchmarks,sgml/FrameworkBenchmarks,circlespainter/FrameworkBenchmarks,markkolich/FrameworkBenchmarks,grob/FrameworkBenchmarks,nkasvosve/FrameworkBenchmarks,zhuochenKIDD/FrameworkBenchmarks,joshk/FrameworkBenchmarks,martin-g/FrameworkBenchmarks,F3Community/FrameworkBenchmarks,seem-sky/FrameworkBenchmarks,kostya-sh/FrameworkBenchmarks,k-r-g/FrameworkBenchmarks,jetty-project/FrameworkBenchmarks,yunspace/FrameworkBenchmarks,s-ludwig/FrameworkBenchmarks,s-ludwig/FrameworkBenchmarks,Ocramius/FrameworkBenchmarks,herloct/FrameworkBenchmarks,knewmanTE/FrameworkBenchmarks,alubbe/FrameworkBenchmarks,steveklabnik/FrameworkBenchmarks,zane-techempower/FrameworkBenchmarks,Verber/FrameworkBenchmarks,knewmanTE/FrameworkBenchmarks,martin-g/FrameworkBenchmarks,k-r-g/FrameworkBenchmarks,k-r-g/FrameworkBenchmarks,markkolich/FrameworkBenchmarks,jebbstewart/FrameworkBenchmarks,donovanmuller/FrameworkBenchmarks,alubbe/FrameworkBenchmarks,torhve/FrameworkBenchmarks,Ocramius/FrameworkBenchmarks,Ocramius/FrameworkBenchmarks,leafo/FrameworkBenchmarks,diablonhn/FrameworkBenchmarks,k-r-g/FrameworkBenchmarks,sgml/FrameworkBenchmarks,kellabyte/FrameworkBenchmarks,xitrum-framework/FrameworkBenchmarks,nathana1/FrameworkBenchmarks,stefanocasazza/FrameworkBenchmarks,zane-techempower/FrameworkBenchmarks,hperadin/FrameworkBenchmarks,saturday06/FrameworkBenchmarks,RockinRoel/FrameworkBenchmarks,zhuochenKIDD/FrameworkBenchmarks,Rayne/FrameworkBenchmarks,kbrock/FrameworkBenchmarks,hperadin/FrameworkBenchmarks,actframework/FrameworkBenchmarks,hperadin/FrameworkBenchmarks,MTDdk/FrameworkBenchmarks,valyala/FrameworkBenchmarks,doom369/FrameworkBenchmarks,Rydgel/FrameworkBenchmarks,xitrum-framework/FrameworkBenchmarks,markkolich/FrameworkBenchmarks,jamming/FrameworkBenchmarks,zhuochenKIDD/FrameworkBenchmarks,hamiltont/FrameworkBenchmarks,MTDdk/FrameworkBenchmarks,marko-asplund/FrameworkBenchmarks,jebbstewart/FrameworkBenchmarks,thousandsofthem/FrameworkBenchmarks,Dith3r/FrameworkBenchmarks,victorbriz/FrameworkBenchmarks,greg-hellings/FrameworkBenchmarks,RockinRoel/FrameworkBenchmarks,kbrock/FrameworkBenchmarks,ratpack/FrameworkBenchmarks,valyala/FrameworkBenchmarks,victorbriz/FrameworkBenchmarks,grob/FrameworkBenchmarks,khellang/FrameworkBenchmarks,torhve/FrameworkBenchmarks,MTDdk/FrameworkBenchmarks,zloster/FrameworkBenchmarks,zdanek/FrameworkBenchmarks,zapov/FrameworkBenchmarks,doom369/FrameworkBenchmarks,doom369/FrameworkBenchmarks,psfblair/FrameworkBenchmarks,stefanocasazza/FrameworkBenchmarks,Rayne/FrameworkBenchmarks,sxend/FrameworkBenchmarks,zapov/FrameworkBenchmarks,ratpack/FrameworkBenchmarks,mfirry/FrameworkBenchmarks,raziel057/FrameworkBenchmarks,knewmanTE/FrameworkBenchmarks,ashawnbandy-te-tfb/FrameworkBenchmarks,zloster/FrameworkBenchmarks,martin-g/FrameworkBenchmarks,zhuochenKIDD/FrameworkBenchmarks,xitrum-framework/FrameworkBenchmarks,sagenschneider/FrameworkBenchmarks,nathana1/FrameworkBenchmarks,xitrum-framework/FrameworkBenchmarks,greg-hellings/FrameworkBenchmarks,joshk/FrameworkBenchmarks,thousandsofthem/FrameworkBenchmarks,raziel057/FrameworkBenchmarks,steveklabnik/FrameworkBenchmarks,Eyepea/FrameworkBenchmarks,Synchro/FrameworkBenchmarks,julienschmidt/FrameworkBenchmarks,doom369/FrameworkBenchmarks,circlespainter/FrameworkBenchmarks,youprofit/FrameworkBenchmarks,zhuochenKIDD/FrameworkBenchmarks,Eyepea/FrameworkBenchmarks,kbrock/FrameworkBenchmarks,victorbriz/FrameworkBenchmarks,k-r-g/FrameworkBenchmarks,doom369/FrameworkBenchmarks,nathana1/FrameworkBenchmarks,nbrady-techempower/FrameworkBenchmarks,Rydgel/FrameworkBenchmarks,marko-asplund/FrameworkBenchmarks,Jesterovskiy/FrameworkBenchmarks,nathana1/FrameworkBenchmarks,sxend/FrameworkBenchmarks,Eyepea/FrameworkBenchmarks,Rydgel/FrameworkBenchmarks,stefanocasazza/FrameworkBenchmarks,valyala/FrameworkBenchmarks,s-ludwig/FrameworkBenchmarks,waiteb3/FrameworkBenchmarks,jebbstewart/FrameworkBenchmarks,k-r-g/FrameworkBenchmarks,fabianmurariu/FrameworkBenchmarks,sagenschneider/FrameworkBenchmarks,zapov/FrameworkBenchmarks,zdanek/FrameworkBenchmarks,methane/FrameworkBenchmarks,sxend/FrameworkBenchmarks,jaguililla/FrameworkBenchmarks,kbrock/FrameworkBenchmarks,jebbstewart/FrameworkBenchmarks,jetty-project/FrameworkBenchmarks,Dith3r/FrameworkBenchmarks,kostya-sh/FrameworkBenchmarks,Synchro/FrameworkBenchmarks,zane-techempower/FrameworkBenchmarks,torhve/FrameworkBenchmarks,sxend/FrameworkBenchmarks,fabianmurariu/FrameworkBenchmarks,grob/FrameworkBenchmarks,F3Community/FrameworkBenchmarks,jeevatkm/FrameworkBenchmarks,Rydgel/FrameworkBenchmarks,kellabyte/FrameworkBenchmarks,actframework/FrameworkBenchmarks,donovanmuller/FrameworkBenchmarks,jeevatkm/FrameworkBenchmarks,marko-asplund/FrameworkBenchmarks,nbrady-techempower/FrameworkBenchmarks,nathana1/FrameworkBenchmarks,thousandsofthem/FrameworkBenchmarks,alubbe/FrameworkBenchmarks,actframework/FrameworkBenchmarks,PermeAgility/FrameworkBenchmarks,MTDdk/FrameworkBenchmarks,dmacd/FB-try1,Synchro/FrameworkBenchmarks,julienschmidt/FrameworkBenchmarks,zdanek/FrameworkBenchmarks,ashawnbandy-te-tfb/FrameworkBenchmarks,psfblair/FrameworkBenchmarks,zane-techempower/FrameworkBenchmarks,kellabyte/FrameworkBenchmarks,khellang/FrameworkBenchmarks,MTDdk/FrameworkBenchmarks,jamming/FrameworkBenchmarks,sgml/FrameworkBenchmarks,doom369/FrameworkBenchmarks,circlespainter/FrameworkBenchmarks,greenlaw110/FrameworkBenchmarks,grob/FrameworkBenchmarks,sanjoydesk/FrameworkBenchmarks,methane/FrameworkBenchmarks,doom369/FrameworkBenchmarks,denkab/FrameworkBenchmarks,hperadin/FrameworkBenchmarks,sanjoydesk/FrameworkBenchmarks,hperadin/FrameworkBenchmarks,sagenschneider/FrameworkBenchmarks,fabianmurariu/FrameworkBenchmarks,jamming/FrameworkBenchmarks,Rydgel/FrameworkBenchmarks,jetty-project/FrameworkBenchmarks,leafo/FrameworkBenchmarks,jaguililla/FrameworkBenchmarks,kellabyte/FrameworkBenchmarks,greenlaw110/FrameworkBenchmarks,kellabyte/FrameworkBenchmarks,doom369/FrameworkBenchmarks,jeevatkm/FrameworkBenchmarks,victorbriz/FrameworkBenchmarks,kbrock/FrameworkBenchmarks,youprofit/FrameworkBenchmarks,knewmanTE/FrameworkBenchmarks,mfirry/FrameworkBenchmarks,jeevatkm/FrameworkBenchmarks,jeevatkm/FrameworkBenchmarks,kostya-sh/FrameworkBenchmarks,fabianmurariu/FrameworkBenchmarks,mfirry/FrameworkBenchmarks,waiteb3/FrameworkBenchmarks,zdanek/FrameworkBenchmarks,PermeAgility/FrameworkBenchmarks,methane/FrameworkBenchmarks,Verber/FrameworkBenchmarks,jetty-project/FrameworkBenchmarks,julienschmidt/FrameworkBenchmarks,sgml/FrameworkBenchmarks,RockinRoel/FrameworkBenchmarks,ratpack/FrameworkBenchmarks,dmacd/FB-try1,jaguililla/FrameworkBenchmarks,joshk/FrameworkBenchmarks,jetty-project/FrameworkBenchmarks,jetty-project/FrameworkBenchmarks,kellabyte/FrameworkBenchmarks,zapov/FrameworkBenchmarks,lcp0578/FrameworkBenchmarks,nbrady-techempower/FrameworkBenchmarks,herloct/FrameworkBenchmarks,steveklabnik/FrameworkBenchmarks,jetty-project/FrameworkBenchmarks,kellabyte/FrameworkBenchmarks,nkasvosve/FrameworkBenchmarks,sanjoydesk/FrameworkBenchmarks,RockinRoel/FrameworkBenchmarks,sxend/FrameworkBenchmarks,actframework/FrameworkBenchmarks,mfirry/FrameworkBenchmarks,nbrady-techempower/FrameworkBenchmarks,marko-asplund/FrameworkBenchmarks,diablonhn/FrameworkBenchmarks,greenlaw110/FrameworkBenchmarks,sxend/FrameworkBenchmarks,markkolich/FrameworkBenchmarks,yunspace/FrameworkBenchmarks,ashawnbandy-te-tfb/FrameworkBenchmarks,fabianmurariu/FrameworkBenchmarks,denkab/FrameworkBenchmarks,sanjoydesk/FrameworkBenchmarks,doom369/FrameworkBenchmarks,Verber/FrameworkBenchmarks,sxend/FrameworkBenchmarks,marko-asplund/FrameworkBenchmarks,actframework/FrameworkBenchmarks,markkolich/FrameworkBenchmarks,circlespainter/FrameworkBenchmarks,saturday06/FrameworkBenchmarks,raziel057/FrameworkBenchmarks,raziel057/FrameworkBenchmarks,joshk/FrameworkBenchmarks,MTDdk/FrameworkBenchmarks,waiteb3/FrameworkBenchmarks,donovanmuller/FrameworkBenchmarks,valyala/FrameworkBenchmarks,martin-g/FrameworkBenchmarks,nbrady-techempower/FrameworkBenchmarks,saturday06/FrameworkBenchmarks,s-ludwig/FrameworkBenchmarks,steveklabnik/FrameworkBenchmarks,hamiltont/FrameworkBenchmarks,Verber/FrameworkBenchmarks,xitrum-framework/FrameworkBenchmarks,sgml/FrameworkBenchmarks,kbrock/FrameworkBenchmarks,nkasvosve/FrameworkBenchmarks,RockinRoel/FrameworkBenchmarks,sanjoydesk/FrameworkBenchmarks,actframework/FrameworkBenchmarks,zane-techempower/FrameworkBenchmarks,greg-hellings/FrameworkBenchmarks,diablonhn/FrameworkBenchmarks,Eyepea/FrameworkBenchmarks,youprofit/FrameworkBenchmarks,saturday06/FrameworkBenchmarks,psfblair/FrameworkBenchmarks,greenlaw110/FrameworkBenchmarks,khellang/FrameworkBenchmarks,jamming/FrameworkBenchmarks,zloster/FrameworkBenchmarks,zloster/FrameworkBenchmarks,thousandsofthem/FrameworkBenchmarks,steveklabnik/FrameworkBenchmarks,jeevatkm/FrameworkBenchmarks,s-ludwig/FrameworkBenchmarks,MTDdk/FrameworkBenchmarks,zhuochenKIDD/FrameworkBenchmarks,actframework/FrameworkBenchmarks,Jesterovskiy/FrameworkBenchmarks,zapov/FrameworkBenchmarks,doom369/FrameworkBenchmarks,steveklabnik/FrameworkBenchmarks,joshk/FrameworkBenchmarks,martin-g/FrameworkBenchmarks,steveklabnik/FrameworkBenchmarks,stefanocasazza/FrameworkBenchmarks,Synchro/FrameworkBenchmarks,jaguililla/FrameworkBenchmarks,knewmanTE/FrameworkBenchmarks,kbrock/FrameworkBenchmarks,zhuochenKIDD/FrameworkBenchmarks,testn/FrameworkBenchmarks,donovanmuller/FrameworkBenchmarks,greg-hellings/FrameworkBenchmarks,RockinRoel/FrameworkBenchmarks,denkab/FrameworkBenchmarks,methane/FrameworkBenchmarks,torhve/FrameworkBenchmarks,valyala/FrameworkBenchmarks,greg-hellings/FrameworkBenchmarks,zhuochenKIDD/FrameworkBenchmarks,youprofit/FrameworkBenchmarks,valyala/FrameworkBenchmarks,ashawnbandy-te-tfb/FrameworkBenchmarks,dmacd/FB-try1,leafo/FrameworkBenchmarks,hamiltont/FrameworkBenchmarks,victorbriz/FrameworkBenchmarks,nbrady-techempower/FrameworkBenchmarks,mfirry/FrameworkBenchmarks,MTDdk/FrameworkBenchmarks,nbrady-techempower/FrameworkBenchmarks,dmacd/FB-try1,fabianmurariu/FrameworkBenchmarks,Jesterovskiy/FrameworkBenchmarks,dmacd/FB-try1,jetty-project/FrameworkBenchmarks,leafo/FrameworkBenchmarks,lcp0578/FrameworkBenchmarks,valyala/FrameworkBenchmarks,zhuochenKIDD/FrameworkBenchmarks,fabianmurariu/FrameworkBenchmarks,thousandsofthem/FrameworkBenchmarks,ratpack/FrameworkBenchmarks,waiteb3/FrameworkBenchmarks,victorbriz/FrameworkBenchmarks,zloster/FrameworkBenchmarks,RockinRoel/FrameworkBenchmarks,raziel057/FrameworkBenchmarks,waiteb3/FrameworkBenchmarks,jamming/FrameworkBenchmarks,knewmanTE/FrameworkBenchmarks,alubbe/FrameworkBenchmarks,doom369/FrameworkBenchmarks,sanjoydesk/FrameworkBenchmarks,greg-hellings/FrameworkBenchmarks,greenlaw110/FrameworkBenchmarks,herloct/FrameworkBenchmarks,Jesterovskiy/FrameworkBenchmarks,sagenschneider/FrameworkBenchmarks,hamiltont/FrameworkBenchmarks,lcp0578/FrameworkBenchmarks,Synchro/FrameworkBenchmarks,MTDdk/FrameworkBenchmarks,waiteb3/FrameworkBenchmarks,hperadin/FrameworkBenchmarks,jebbstewart/FrameworkBenchmarks,Synchro/FrameworkBenchmarks,ashawnbandy-te-tfb/FrameworkBenchmarks,actframework/FrameworkBenchmarks,saturday06/FrameworkBenchmarks,thousandsofthem/FrameworkBenchmarks,lcp0578/FrameworkBenchmarks,greg-hellings/FrameworkBenchmarks,seem-sky/FrameworkBenchmarks,khellang/FrameworkBenchmarks,Dith3r/FrameworkBenchmarks,julienschmidt/FrameworkBenchmarks,lcp0578/FrameworkBenchmarks,martin-g/FrameworkBenchmarks,valyala/FrameworkBenchmarks,Dith3r/FrameworkBenchmarks,leafo/FrameworkBenchmarks,RockinRoel/FrameworkBenchmarks,Dith3r/FrameworkBenchmarks,sagenschneider/FrameworkBenchmarks,testn/FrameworkBenchmarks,k-r-g/FrameworkBenchmarks,denkab/FrameworkBenchmarks,actframework/FrameworkBenchmarks,mfirry/FrameworkBenchmarks,k-r-g/FrameworkBenchmarks,ashawnbandy-te-tfb/FrameworkBenchmarks,yunspace/FrameworkBenchmarks,fabianmurariu/FrameworkBenchmarks,marko-asplund/FrameworkBenchmarks,jebbstewart/FrameworkBenchmarks,seem-sky/FrameworkBenchmarks,Rayne/FrameworkBenchmarks,circlespainter/FrameworkBenchmarks,methane/FrameworkBenchmarks,joshk/FrameworkBenchmarks,raziel057/FrameworkBenchmarks,torhve/FrameworkBenchmarks,circlespainter/FrameworkBenchmarks,julienschmidt/FrameworkBenchmarks,Dith3r/FrameworkBenchmarks,k-r-g/FrameworkBenchmarks,jaguililla/FrameworkBenchmarks,PermeAgility/FrameworkBenchmarks,zane-techempower/FrameworkBenchmarks,zapov/FrameworkBenchmarks,Dith3r/FrameworkBenchmarks,dmacd/FB-try1,sxend/FrameworkBenchmarks,zdanek/FrameworkBenchmarks,greenlaw110/FrameworkBenchmarks,sgml/FrameworkBenchmarks,Jesterovskiy/FrameworkBenchmarks,Rydgel/FrameworkBenchmarks,Verber/FrameworkBenchmarks,zloster/FrameworkBenchmarks,kostya-sh/FrameworkBenchmarks,methane/FrameworkBenchmarks,yunspace/FrameworkBenchmarks,steveklabnik/FrameworkBenchmarks,Synchro/FrameworkBenchmarks,victorbriz/FrameworkBenchmarks,hamiltont/FrameworkBenchmarks,testn/FrameworkBenchmarks,joshk/FrameworkBenchmarks,Rayne/FrameworkBenchmarks,MTDdk/FrameworkBenchmarks,F3Community/FrameworkBenchmarks,Rayne/FrameworkBenchmarks,zane-techempower/FrameworkBenchmarks,zapov/FrameworkBenchmarks,stefanocasazza/FrameworkBenchmarks,herloct/FrameworkBenchmarks,herloct/FrameworkBenchmarks,MTDdk/FrameworkBenchmarks,circlespainter/FrameworkBenchmarks,F3Community/FrameworkBenchmarks,nbrady-techempower/FrameworkBenchmarks,zloster/FrameworkBenchmarks,jebbstewart/FrameworkBenchmarks,sagenschneider/FrameworkBenchmarks,jamming/FrameworkBenchmarks,alubbe/FrameworkBenchmarks,steveklabnik/FrameworkBenchmarks,sagenschneider/FrameworkBenchmarks,jaguililla/FrameworkBenchmarks,joshk/FrameworkBenchmarks,sagenschneider/FrameworkBenchmarks,greenlaw110/FrameworkBenchmarks,valyala/FrameworkBenchmarks,sxend/FrameworkBenchmarks,raziel057/FrameworkBenchmarks,sagenschneider/FrameworkBenchmarks,k-r-g/FrameworkBenchmarks,youprofit/FrameworkBenchmarks,Jesterovskiy/FrameworkBenchmarks,Dith3r/FrameworkBenchmarks,k-r-g/FrameworkBenchmarks,PermeAgility/FrameworkBenchmarks,torhve/FrameworkBenchmarks,Rayne/FrameworkBenchmarks,joshk/FrameworkBenchmarks,yunspace/FrameworkBenchmarks,jetty-project/FrameworkBenchmarks,sgml/FrameworkBenchmarks,lcp0578/FrameworkBenchmarks,greenlaw110/FrameworkBenchmarks,victorbriz/FrameworkBenchmarks,psfblair/FrameworkBenchmarks,nbrady-techempower/FrameworkBenchmarks,greenlaw110/FrameworkBenchmarks,yunspace/FrameworkBenchmarks,lcp0578/FrameworkBenchmarks,denkab/FrameworkBenchmarks,stefanocasazza/FrameworkBenchmarks,Jesterovskiy/FrameworkBenchmarks,Rayne/FrameworkBenchmarks,raziel057/FrameworkBenchmarks,jamming/FrameworkBenchmarks,jeevatkm/FrameworkBenchmarks,thousandsofthem/FrameworkBenchmarks,zdanek/FrameworkBenchmarks,donovanmuller/FrameworkBenchmarks,valyala/FrameworkBenchmarks,Synchro/FrameworkBenchmarks,hamiltont/FrameworkBenchmarks,F3Community/FrameworkBenchmarks,lcp0578/FrameworkBenchmarks,kostya-sh/FrameworkBenchmarks,hamiltont/FrameworkBenchmarks,jebbstewart/FrameworkBenchmarks,yunspace/FrameworkBenchmarks,greenlaw110/FrameworkBenchmarks,youprofit/FrameworkBenchmarks,leafo/FrameworkBenchmarks,martin-g/FrameworkBenchmarks,zapov/FrameworkBenchmarks,Jesterovskiy/FrameworkBenchmarks,leafo/FrameworkBenchmarks,diablonhn/FrameworkBenchmarks,jamming/FrameworkBenchmarks,grob/FrameworkBenchmarks,martin-g/FrameworkBenchmarks,Dith3r/FrameworkBenchmarks,s-ludwig/FrameworkBenchmarks,hperadin/FrameworkBenchmarks,s-ludwig/FrameworkBenchmarks
import subprocess import sys import setup_util import os def start(args): - subprocess.Popen("gunicorn hello:app -b 0.0.0.0:8080 -w " + str((args.max_threads * 2)) + " --log-level=critical", shell=True, cwd="wsgi") + subprocess.Popen('gunicorn hello:app --worker-class="egg:meinheld#gunicorn_worker" -b 0.0.0.0:8080 -w ' + + str((args.max_threads * 2)) + " --log-level=critical", shell=True, cwd="wsgi") return 0 def stop(): p = subprocess.Popen(['ps', 'aux'], stdout=subprocess.PIPE) out, err = p.communicate() for line in out.splitlines(): if 'gunicorn' in line: try: pid = int(line.split(None, 2)[1]) os.kill(pid, 9) except OSError: pass return 0 +
Use meinheld worker (same as other Python Frameworks)
## Code Before: import subprocess import sys import setup_util import os def start(args): subprocess.Popen("gunicorn hello:app -b 0.0.0.0:8080 -w " + str((args.max_threads * 2)) + " --log-level=critical", shell=True, cwd="wsgi") return 0 def stop(): p = subprocess.Popen(['ps', 'aux'], stdout=subprocess.PIPE) out, err = p.communicate() for line in out.splitlines(): if 'gunicorn' in line: try: pid = int(line.split(None, 2)[1]) os.kill(pid, 9) except OSError: pass return 0 ## Instruction: Use meinheld worker (same as other Python Frameworks) ## Code After: import subprocess import sys import setup_util import os def start(args): subprocess.Popen('gunicorn hello:app --worker-class="egg:meinheld#gunicorn_worker" -b 0.0.0.0:8080 -w ' + str((args.max_threads * 2)) + " --log-level=critical", shell=True, cwd="wsgi") return 0 def stop(): p = subprocess.Popen(['ps', 'aux'], stdout=subprocess.PIPE) out, err = p.communicate() for line in out.splitlines(): if 'gunicorn' in line: try: pid = int(line.split(None, 2)[1]) os.kill(pid, 9) except OSError: pass return 0
# ... existing code ... def start(args): subprocess.Popen('gunicorn hello:app --worker-class="egg:meinheld#gunicorn_worker" -b 0.0.0.0:8080 -w ' + str((args.max_threads * 2)) + " --log-level=critical", shell=True, cwd="wsgi") return 0 def stop(): # ... rest of the code ...
2046d82addab9ec83dbb85a2d08c727a52065d8b
deckglue/models.py
deckglue/models.py
from django.db import models # Create your models here.
from django.contrib.auth.models import Permission from django.db.models.signals import post_save, pre_delete from django.dispatch import receiver from cardbox.card_model import Card from cardbox.deck_model import Deck from guardian.shortcuts import assign_perm, get_users_with_perms from guardian.models import UserObjectPermission from memorize.models import Practice from django.contrib.auth.models import User @receiver(post_save, sender=UserObjectPermission) def create_practice_objects_for_new_viewers(sender, **kwargs): if kwargs['instance'].permission_id == Permission.objects.get(codename="view_deck").id: for card in Card.objects.filter(deck=kwargs['instance'].object_pk): Practice(item=card, user=User.objects.get(id = kwargs['instance'].user_id)).save() @receiver(pre_delete, sender=UserObjectPermission) def delete_practice_objects_for_removed_viewers(sender, **kwargs): if kwargs['instance'].permission_id == Permission.objects.get(codename="view_deck").id: for card in Card.objects.filter(deck=kwargs['instance'].object_pk): Practice.objects.get(object_id=card.ID, user=User.objects.get(id = kwargs['instance'].user_id)).delete() @receiver(post_save, sender=Card) def create_practice_objects_for_new_card(sender,update_fields, **kwargs): """Creates practice objects for all users with permission to view the card. """ perm_users = get_users_with_perms(kwargs['instance'].deck) for user in perm_users: practice = Practice(item = kwargs['instance'], user = user) if Practice.objects.filter(object_id = kwargs['instance'].ID, user=user).count() == 0: practice.save() @receiver(pre_delete, sender=Card) def delete_practice_objects(sender, **kwargs): """Deletes all practice objects for a card once it is deleted. """ Practice.objects.filter(object_id = kwargs['instance'].ID).delete()
Add signal hooks to create practice objects
Add signal hooks to create practice objects
Python
mit
DummyDivision/Tsune,DummyDivision/Tsune,DummyDivision/Tsune
- from django.db import models + from django.contrib.auth.models import Permission + from django.db.models.signals import post_save, pre_delete + from django.dispatch import receiver + from cardbox.card_model import Card + from cardbox.deck_model import Deck + from guardian.shortcuts import assign_perm, get_users_with_perms + from guardian.models import UserObjectPermission + from memorize.models import Practice + from django.contrib.auth.models import User - # Create your models here. + @receiver(post_save, sender=UserObjectPermission) + def create_practice_objects_for_new_viewers(sender, **kwargs): + if kwargs['instance'].permission_id == Permission.objects.get(codename="view_deck").id: + for card in Card.objects.filter(deck=kwargs['instance'].object_pk): + Practice(item=card, user=User.objects.get(id = kwargs['instance'].user_id)).save() + @receiver(pre_delete, sender=UserObjectPermission) + def delete_practice_objects_for_removed_viewers(sender, **kwargs): + if kwargs['instance'].permission_id == Permission.objects.get(codename="view_deck").id: + for card in Card.objects.filter(deck=kwargs['instance'].object_pk): + Practice.objects.get(object_id=card.ID, user=User.objects.get(id = kwargs['instance'].user_id)).delete() + + @receiver(post_save, sender=Card) + def create_practice_objects_for_new_card(sender,update_fields, **kwargs): + """Creates practice objects for all users with permission to view the card. + + """ + perm_users = get_users_with_perms(kwargs['instance'].deck) + for user in perm_users: + practice = Practice(item = kwargs['instance'], user = user) + if Practice.objects.filter(object_id = kwargs['instance'].ID, user=user).count() == 0: + practice.save() + + + + @receiver(pre_delete, sender=Card) + def delete_practice_objects(sender, **kwargs): + """Deletes all practice objects for a card once it is deleted. + + """ + Practice.objects.filter(object_id = kwargs['instance'].ID).delete() + +
Add signal hooks to create practice objects
## Code Before: from django.db import models # Create your models here. ## Instruction: Add signal hooks to create practice objects ## Code After: from django.contrib.auth.models import Permission from django.db.models.signals import post_save, pre_delete from django.dispatch import receiver from cardbox.card_model import Card from cardbox.deck_model import Deck from guardian.shortcuts import assign_perm, get_users_with_perms from guardian.models import UserObjectPermission from memorize.models import Practice from django.contrib.auth.models import User @receiver(post_save, sender=UserObjectPermission) def create_practice_objects_for_new_viewers(sender, **kwargs): if kwargs['instance'].permission_id == Permission.objects.get(codename="view_deck").id: for card in Card.objects.filter(deck=kwargs['instance'].object_pk): Practice(item=card, user=User.objects.get(id = kwargs['instance'].user_id)).save() @receiver(pre_delete, sender=UserObjectPermission) def delete_practice_objects_for_removed_viewers(sender, **kwargs): if kwargs['instance'].permission_id == Permission.objects.get(codename="view_deck").id: for card in Card.objects.filter(deck=kwargs['instance'].object_pk): Practice.objects.get(object_id=card.ID, user=User.objects.get(id = kwargs['instance'].user_id)).delete() @receiver(post_save, sender=Card) def create_practice_objects_for_new_card(sender,update_fields, **kwargs): """Creates practice objects for all users with permission to view the card. """ perm_users = get_users_with_perms(kwargs['instance'].deck) for user in perm_users: practice = Practice(item = kwargs['instance'], user = user) if Practice.objects.filter(object_id = kwargs['instance'].ID, user=user).count() == 0: practice.save() @receiver(pre_delete, sender=Card) def delete_practice_objects(sender, **kwargs): """Deletes all practice objects for a card once it is deleted. """ Practice.objects.filter(object_id = kwargs['instance'].ID).delete()
// ... existing code ... from django.contrib.auth.models import Permission from django.db.models.signals import post_save, pre_delete from django.dispatch import receiver from cardbox.card_model import Card from cardbox.deck_model import Deck from guardian.shortcuts import assign_perm, get_users_with_perms from guardian.models import UserObjectPermission from memorize.models import Practice from django.contrib.auth.models import User @receiver(post_save, sender=UserObjectPermission) def create_practice_objects_for_new_viewers(sender, **kwargs): if kwargs['instance'].permission_id == Permission.objects.get(codename="view_deck").id: for card in Card.objects.filter(deck=kwargs['instance'].object_pk): Practice(item=card, user=User.objects.get(id = kwargs['instance'].user_id)).save() @receiver(pre_delete, sender=UserObjectPermission) def delete_practice_objects_for_removed_viewers(sender, **kwargs): if kwargs['instance'].permission_id == Permission.objects.get(codename="view_deck").id: for card in Card.objects.filter(deck=kwargs['instance'].object_pk): Practice.objects.get(object_id=card.ID, user=User.objects.get(id = kwargs['instance'].user_id)).delete() @receiver(post_save, sender=Card) def create_practice_objects_for_new_card(sender,update_fields, **kwargs): """Creates practice objects for all users with permission to view the card. """ perm_users = get_users_with_perms(kwargs['instance'].deck) for user in perm_users: practice = Practice(item = kwargs['instance'], user = user) if Practice.objects.filter(object_id = kwargs['instance'].ID, user=user).count() == 0: practice.save() @receiver(pre_delete, sender=Card) def delete_practice_objects(sender, **kwargs): """Deletes all practice objects for a card once it is deleted. """ Practice.objects.filter(object_id = kwargs['instance'].ID).delete() // ... rest of the code ...
db13de154fa44f3ef0bf1e365d2ee0d7a6951700
cellcounter/accounts/urls.py
cellcounter/accounts/urls.py
from django.conf.urls import patterns, url from cellcounter.accounts import views urlpatterns = patterns('', url('^new/$', views.RegistrationView.as_view(), name='register'), url('^(?P<pk>[0-9]+)/$', views.UserDetailView.as_view(), name='user-detail'), url('^(?P<pk>[0-9]+)/delete/$', views.UserDeleteView.as_view(), name='user-delete'), url('^(?P<pk>[0-9]+)/edit/$', views.UserUpdateView.as_view(), name='user-update'), url('^password/reset/$', views.PasswordResetView.as_view(), name='password-reset'), url('^password/reset/confirm/(?P<uidb64>[0-9A-Za-z_\-]+)/(?P<token>[\d\w\-]+)/$', views.PasswordResetConfirmView.as_view(), name='password-reset-confirm'), url('^password/change/$', views.PasswordChangeView.as_view(), name='change-password'), )
from django.conf.urls import patterns, url from cellcounter.accounts import views urlpatterns = patterns('', url('^new/$', views.RegistrationView.as_view(), name='register'), url('^(?P<pk>[0-9]+)/$', views.UserDetailView.as_view(), name='user-detail'), url('^(?P<pk>[0-9]+)/delete/$', views.UserDeleteView.as_view(), name='user-delete'), url('^(?P<pk>[0-9]+)/edit/$', views.UserUpdateView.as_view(), name='user-update'), url('^password/reset/$', views.PasswordResetView.as_view(), name='password-reset'), url('^password/reset/confirm/(?P<uidb64>[0-9A-Za-z_\-]+)/(?P<token>[0-9A-Za-z]{1,13}-[0-9A-Za-z]{1,20})/$', views.PasswordResetConfirmView.as_view(), name='password-reset-confirm'), url('^password/change/$', views.PasswordChangeView.as_view(), name='change-password'), )
Use URL regex as per main Django project
Use URL regex as per main Django project
Python
mit
cellcounter/cellcounter,haematologic/cellcounter,haematologic/cellcounter,cellcounter/cellcounter,cellcounter/cellcounter,cellcounter/cellcounter,haematologic/cellcounter
from django.conf.urls import patterns, url from cellcounter.accounts import views urlpatterns = patterns('', url('^new/$', views.RegistrationView.as_view(), name='register'), url('^(?P<pk>[0-9]+)/$', views.UserDetailView.as_view(), name='user-detail'), url('^(?P<pk>[0-9]+)/delete/$', views.UserDeleteView.as_view(), name='user-delete'), url('^(?P<pk>[0-9]+)/edit/$', views.UserUpdateView.as_view(), name='user-update'), url('^password/reset/$', views.PasswordResetView.as_view(), name='password-reset'), - url('^password/reset/confirm/(?P<uidb64>[0-9A-Za-z_\-]+)/(?P<token>[\d\w\-]+)/$', + url('^password/reset/confirm/(?P<uidb64>[0-9A-Za-z_\-]+)/(?P<token>[0-9A-Za-z]{1,13}-[0-9A-Za-z]{1,20})/$', views.PasswordResetConfirmView.as_view(), name='password-reset-confirm'), url('^password/change/$', views.PasswordChangeView.as_view(), name='change-password'), )
Use URL regex as per main Django project
## Code Before: from django.conf.urls import patterns, url from cellcounter.accounts import views urlpatterns = patterns('', url('^new/$', views.RegistrationView.as_view(), name='register'), url('^(?P<pk>[0-9]+)/$', views.UserDetailView.as_view(), name='user-detail'), url('^(?P<pk>[0-9]+)/delete/$', views.UserDeleteView.as_view(), name='user-delete'), url('^(?P<pk>[0-9]+)/edit/$', views.UserUpdateView.as_view(), name='user-update'), url('^password/reset/$', views.PasswordResetView.as_view(), name='password-reset'), url('^password/reset/confirm/(?P<uidb64>[0-9A-Za-z_\-]+)/(?P<token>[\d\w\-]+)/$', views.PasswordResetConfirmView.as_view(), name='password-reset-confirm'), url('^password/change/$', views.PasswordChangeView.as_view(), name='change-password'), ) ## Instruction: Use URL regex as per main Django project ## Code After: from django.conf.urls import patterns, url from cellcounter.accounts import views urlpatterns = patterns('', url('^new/$', views.RegistrationView.as_view(), name='register'), url('^(?P<pk>[0-9]+)/$', views.UserDetailView.as_view(), name='user-detail'), url('^(?P<pk>[0-9]+)/delete/$', views.UserDeleteView.as_view(), name='user-delete'), url('^(?P<pk>[0-9]+)/edit/$', views.UserUpdateView.as_view(), name='user-update'), url('^password/reset/$', views.PasswordResetView.as_view(), name='password-reset'), url('^password/reset/confirm/(?P<uidb64>[0-9A-Za-z_\-]+)/(?P<token>[0-9A-Za-z]{1,13}-[0-9A-Za-z]{1,20})/$', views.PasswordResetConfirmView.as_view(), name='password-reset-confirm'), url('^password/change/$', views.PasswordChangeView.as_view(), name='change-password'), )
# ... existing code ... url('^password/reset/$', views.PasswordResetView.as_view(), name='password-reset'), url('^password/reset/confirm/(?P<uidb64>[0-9A-Za-z_\-]+)/(?P<token>[0-9A-Za-z]{1,13}-[0-9A-Za-z]{1,20})/$', views.PasswordResetConfirmView.as_view(), name='password-reset-confirm'), # ... rest of the code ...
079e7cbbd59266e1dc8b161989c90202caa4c5a8
flaskbb/utils/views.py
flaskbb/utils/views.py
from flask import render_template from flask.views import View class RenderableView(View): def __init__(self, template, view): self.template = template self.view = view def dispatch_request(self, *args, **kwargs): view_model = self.view(*args, **kwargs) return render_template(self.template, **view_model)
from flaskbb.utils.helpers import render_template from flask.views import View class RenderableView(View): def __init__(self, template, view): self.template = template self.view = view def dispatch_request(self, *args, **kwargs): view_model = self.view(*args, **kwargs) return render_template(self.template, **view_model)
Use local render_template than Flask's native
Use local render_template than Flask's native TODO: Provide a renderer argument at instantation?
Python
bsd-3-clause
realityone/flaskbb,realityone/flaskbb,dromanow/flaskbb,dromanow/flaskbb,realityone/flaskbb,dromanow/flaskbb
- from flask import render_template + from flaskbb.utils.helpers import render_template from flask.views import View class RenderableView(View): def __init__(self, template, view): self.template = template self.view = view def dispatch_request(self, *args, **kwargs): view_model = self.view(*args, **kwargs) return render_template(self.template, **view_model)
Use local render_template than Flask's native
## Code Before: from flask import render_template from flask.views import View class RenderableView(View): def __init__(self, template, view): self.template = template self.view = view def dispatch_request(self, *args, **kwargs): view_model = self.view(*args, **kwargs) return render_template(self.template, **view_model) ## Instruction: Use local render_template than Flask's native ## Code After: from flaskbb.utils.helpers import render_template from flask.views import View class RenderableView(View): def __init__(self, template, view): self.template = template self.view = view def dispatch_request(self, *args, **kwargs): view_model = self.view(*args, **kwargs) return render_template(self.template, **view_model)
... from flaskbb.utils.helpers import render_template from flask.views import View ...