commit
stringlengths 40
40
| old_file
stringlengths 4
118
| new_file
stringlengths 4
118
| old_contents
stringlengths 10
3.52k
| new_contents
stringlengths 21
3.18k
| subject
stringlengths 16
444
| message
stringlengths 17
2.63k
| lang
stringclasses 1
value | license
stringclasses 13
values | repos
stringlengths 7
43k
| ndiff
stringlengths 52
3.32k
| instruction
stringlengths 16
444
| content
stringlengths 133
4.32k
| fuzzy_diff
stringlengths 17
3.24k
|
---|---|---|---|---|---|---|---|---|---|---|---|---|---|
52d9ed9c08ef0686a891e3428349b70d74a7ecf8 | scripts/munge_fah_data.py | scripts/munge_fah_data.py | import numpy as np
import os
import glob
import mdtraj as md
import fahmunge
import pandas as pd
projects = pd.read_csv("./projects.csv", index_col=0)
output_path = "/data/choderalab/fah/munged/"
for (project, location, pdb) in projects.itertuples():
print(project, location, pdb)
allatom_output_path = os.path.join(output_path, str(project), "allatoms/")
protein_output_path = os.path.join(output_path, str(project), "protein/")
fahmunge.automation.make_path(allatom_output_path)
fahmunge.automation.make_path(protein_output_path)
fahmunge.automation.merge_fah_trajectories(location, allatom_output_path, pdb)
trj0 = md.load(pdb) # Hacky temporary solution.
top, bonds = trj0.top.to_dataframe()
protein_atom_indices = top.index[top.chainID == 0].values
fahmunge.automation.strip_water(allatom_output_path, protein_output_path, protein_atom_indices)
| import numpy as np
import os
import glob
import mdtraj as md
import fahmunge
import pandas as pd
projects = pd.read_csv("./projects.csv", index_col=0)
output_path = "/data/choderalab/fah/munged/"
for (project, location, pdb) in projects.itertuples():
print(project, location, pdb)
allatom_output_path = os.path.join(output_path, "allatoms/", "%s/" % project)
protein_output_path = os.path.join(output_path, "protein/", "%s/" % project)
fahmunge.automation.make_path(allatom_output_path)
fahmunge.automation.make_path(protein_output_path)
fahmunge.automation.merge_fah_trajectories(location, allatom_output_path, pdb)
trj0 = md.load(pdb) # Hacky temporary solution.
top, bonds = trj0.top.to_dataframe()
protein_atom_indices = top.index[top.chainID == 0].values
fahmunge.automation.strip_water(allatom_output_path, protein_output_path, protein_atom_indices)
| Change output data structure to support faster rsync | Change output data structure to support faster rsync
| Python | lgpl-2.1 | steven-albanese/FAHMunge,kyleabeauchamp/FAHMunge,choderalab/FAHMunge | import numpy as np
import os
import glob
import mdtraj as md
import fahmunge
import pandas as pd
projects = pd.read_csv("./projects.csv", index_col=0)
output_path = "/data/choderalab/fah/munged/"
for (project, location, pdb) in projects.itertuples():
print(project, location, pdb)
- allatom_output_path = os.path.join(output_path, str(project), "allatoms/")
+ allatom_output_path = os.path.join(output_path, "allatoms/", "%s/" % project)
- protein_output_path = os.path.join(output_path, str(project), "protein/")
+ protein_output_path = os.path.join(output_path, "protein/", "%s/" % project)
fahmunge.automation.make_path(allatom_output_path)
fahmunge.automation.make_path(protein_output_path)
fahmunge.automation.merge_fah_trajectories(location, allatom_output_path, pdb)
trj0 = md.load(pdb) # Hacky temporary solution.
top, bonds = trj0.top.to_dataframe()
protein_atom_indices = top.index[top.chainID == 0].values
fahmunge.automation.strip_water(allatom_output_path, protein_output_path, protein_atom_indices)
| Change output data structure to support faster rsync | ## Code Before:
import numpy as np
import os
import glob
import mdtraj as md
import fahmunge
import pandas as pd
projects = pd.read_csv("./projects.csv", index_col=0)
output_path = "/data/choderalab/fah/munged/"
for (project, location, pdb) in projects.itertuples():
print(project, location, pdb)
allatom_output_path = os.path.join(output_path, str(project), "allatoms/")
protein_output_path = os.path.join(output_path, str(project), "protein/")
fahmunge.automation.make_path(allatom_output_path)
fahmunge.automation.make_path(protein_output_path)
fahmunge.automation.merge_fah_trajectories(location, allatom_output_path, pdb)
trj0 = md.load(pdb) # Hacky temporary solution.
top, bonds = trj0.top.to_dataframe()
protein_atom_indices = top.index[top.chainID == 0].values
fahmunge.automation.strip_water(allatom_output_path, protein_output_path, protein_atom_indices)
## Instruction:
Change output data structure to support faster rsync
## Code After:
import numpy as np
import os
import glob
import mdtraj as md
import fahmunge
import pandas as pd
projects = pd.read_csv("./projects.csv", index_col=0)
output_path = "/data/choderalab/fah/munged/"
for (project, location, pdb) in projects.itertuples():
print(project, location, pdb)
allatom_output_path = os.path.join(output_path, "allatoms/", "%s/" % project)
protein_output_path = os.path.join(output_path, "protein/", "%s/" % project)
fahmunge.automation.make_path(allatom_output_path)
fahmunge.automation.make_path(protein_output_path)
fahmunge.automation.merge_fah_trajectories(location, allatom_output_path, pdb)
trj0 = md.load(pdb) # Hacky temporary solution.
top, bonds = trj0.top.to_dataframe()
protein_atom_indices = top.index[top.chainID == 0].values
fahmunge.automation.strip_water(allatom_output_path, protein_output_path, protein_atom_indices)
| // ... existing code ...
for (project, location, pdb) in projects.itertuples():
print(project, location, pdb)
allatom_output_path = os.path.join(output_path, "allatoms/", "%s/" % project)
protein_output_path = os.path.join(output_path, "protein/", "%s/" % project)
fahmunge.automation.make_path(allatom_output_path)
fahmunge.automation.make_path(protein_output_path)
// ... rest of the code ... |
096f9e86755a6967d732986c51ae00855551cf4d | project_name/urls.py | project_name/urls.py | from django.conf import settings
from django.conf.urls import include, url # noqa
from django.contrib import admin
from django.views.generic import TemplateView
import django_js_reverse.views
urlpatterns = [
url(r'^admin/', admin.site.urls),
url(r'^jsreverse/$', django_js_reverse.views.urls_js, name='js_reverse'),
url(r'^$', TemplateView.as_view(template_name='exampleapp/itworks.html'), name='home'),
]
if settings.DEBUG:
import debug_toolbar
urlpatterns = [
url(r'^__debug__/', include(debug_toolbar.urls)),
] + urlpatterns
| from django.conf.urls import include, url # noqa
from django.contrib import admin
from django.views.generic import TemplateView
import django_js_reverse.views
urlpatterns = [
url(r'^admin/', admin.site.urls),
url(r'^jsreverse/$', django_js_reverse.views.urls_js, name='js_reverse'),
url(r'^$', TemplateView.as_view(template_name='exampleapp/itworks.html'), name='home'),
]
| Remove usage from debug toolbar | Remove usage from debug toolbar
| Python | mit | vintasoftware/django-react-boilerplate,vintasoftware/django-react-boilerplate,vintasoftware/django-react-boilerplate,vintasoftware/django-react-boilerplate | - from django.conf import settings
from django.conf.urls import include, url # noqa
from django.contrib import admin
from django.views.generic import TemplateView
import django_js_reverse.views
urlpatterns = [
url(r'^admin/', admin.site.urls),
url(r'^jsreverse/$', django_js_reverse.views.urls_js, name='js_reverse'),
url(r'^$', TemplateView.as_view(template_name='exampleapp/itworks.html'), name='home'),
]
- if settings.DEBUG:
- import debug_toolbar
- urlpatterns = [
- url(r'^__debug__/', include(debug_toolbar.urls)),
- ] + urlpatterns
- | Remove usage from debug toolbar | ## Code Before:
from django.conf import settings
from django.conf.urls import include, url # noqa
from django.contrib import admin
from django.views.generic import TemplateView
import django_js_reverse.views
urlpatterns = [
url(r'^admin/', admin.site.urls),
url(r'^jsreverse/$', django_js_reverse.views.urls_js, name='js_reverse'),
url(r'^$', TemplateView.as_view(template_name='exampleapp/itworks.html'), name='home'),
]
if settings.DEBUG:
import debug_toolbar
urlpatterns = [
url(r'^__debug__/', include(debug_toolbar.urls)),
] + urlpatterns
## Instruction:
Remove usage from debug toolbar
## Code After:
from django.conf.urls import include, url # noqa
from django.contrib import admin
from django.views.generic import TemplateView
import django_js_reverse.views
urlpatterns = [
url(r'^admin/', admin.site.urls),
url(r'^jsreverse/$', django_js_reverse.views.urls_js, name='js_reverse'),
url(r'^$', TemplateView.as_view(template_name='exampleapp/itworks.html'), name='home'),
]
| ...
from django.conf.urls import include, url # noqa
from django.contrib import admin
...
url(r'^$', TemplateView.as_view(template_name='exampleapp/itworks.html'), name='home'),
]
... |
f112e7754e4f4368f0a82c3aae3a58f5300176f0 | spacy/language_data/tag_map.py | spacy/language_data/tag_map.py | from __future__ import unicode_literals
from ..symbols import *
TAG_MAP = {
"ADV": {POS: ADV},
"NOUN": {POS: NOUN},
"ADP": {POS: ADP},
"PRON": {POS: PRON},
"SCONJ": {POS: SCONJ},
"PROPN": {POS: PROPN},
"DET": {POS: DET},
"SYM": {POS: SYM},
"INTJ": {POS: INTJ},
"PUNCT": {POS: PUNCT},
"NUM": {POS: NUM},
"AUX": {POS: AUX},
"X": {POS: X},
"CONJ": {POS: CONJ},
"ADJ": {POS: ADJ},
"VERB": {POS: VERB}
}
| from __future__ import unicode_literals
from ..symbols import *
TAG_MAP = {
"ADV": {POS: ADV},
"NOUN": {POS: NOUN},
"ADP": {POS: ADP},
"PRON": {POS: PRON},
"SCONJ": {POS: SCONJ},
"PROPN": {POS: PROPN},
"DET": {POS: DET},
"SYM": {POS: SYM},
"INTJ": {POS: INTJ},
"PUNCT": {POS: PUNCT},
"NUM": {POS: NUM},
"AUX": {POS: AUX},
"X": {POS: X},
"CONJ": {POS: CONJ},
"ADJ": {POS: ADJ},
"VERB": {POS: VERB},
"PART": {POS: PART}
}
| Add PART to tag map | Add PART to tag map
16 of the 17 PoS tags in the UD tag set is added; PART is missing. | Python | mit | banglakit/spaCy,raphael0202/spaCy,recognai/spaCy,spacy-io/spaCy,explosion/spaCy,spacy-io/spaCy,spacy-io/spaCy,honnibal/spaCy,recognai/spaCy,Gregory-Howard/spaCy,aikramer2/spaCy,Gregory-Howard/spaCy,oroszgy/spaCy.hu,oroszgy/spaCy.hu,raphael0202/spaCy,recognai/spaCy,Gregory-Howard/spaCy,aikramer2/spaCy,explosion/spaCy,honnibal/spaCy,explosion/spaCy,Gregory-Howard/spaCy,banglakit/spaCy,banglakit/spaCy,Gregory-Howard/spaCy,banglakit/spaCy,spacy-io/spaCy,oroszgy/spaCy.hu,raphael0202/spaCy,aikramer2/spaCy,honnibal/spaCy,banglakit/spaCy,recognai/spaCy,explosion/spaCy,aikramer2/spaCy,recognai/spaCy,raphael0202/spaCy,oroszgy/spaCy.hu,raphael0202/spaCy,recognai/spaCy,aikramer2/spaCy,spacy-io/spaCy,oroszgy/spaCy.hu,explosion/spaCy,oroszgy/spaCy.hu,raphael0202/spaCy,honnibal/spaCy,banglakit/spaCy,aikramer2/spaCy,Gregory-Howard/spaCy,spacy-io/spaCy,explosion/spaCy | from __future__ import unicode_literals
from ..symbols import *
TAG_MAP = {
"ADV": {POS: ADV},
"NOUN": {POS: NOUN},
"ADP": {POS: ADP},
"PRON": {POS: PRON},
"SCONJ": {POS: SCONJ},
"PROPN": {POS: PROPN},
"DET": {POS: DET},
"SYM": {POS: SYM},
"INTJ": {POS: INTJ},
"PUNCT": {POS: PUNCT},
"NUM": {POS: NUM},
"AUX": {POS: AUX},
"X": {POS: X},
"CONJ": {POS: CONJ},
"ADJ": {POS: ADJ},
- "VERB": {POS: VERB}
+ "VERB": {POS: VERB},
+ "PART": {POS: PART}
}
| Add PART to tag map | ## Code Before:
from __future__ import unicode_literals
from ..symbols import *
TAG_MAP = {
"ADV": {POS: ADV},
"NOUN": {POS: NOUN},
"ADP": {POS: ADP},
"PRON": {POS: PRON},
"SCONJ": {POS: SCONJ},
"PROPN": {POS: PROPN},
"DET": {POS: DET},
"SYM": {POS: SYM},
"INTJ": {POS: INTJ},
"PUNCT": {POS: PUNCT},
"NUM": {POS: NUM},
"AUX": {POS: AUX},
"X": {POS: X},
"CONJ": {POS: CONJ},
"ADJ": {POS: ADJ},
"VERB": {POS: VERB}
}
## Instruction:
Add PART to tag map
## Code After:
from __future__ import unicode_literals
from ..symbols import *
TAG_MAP = {
"ADV": {POS: ADV},
"NOUN": {POS: NOUN},
"ADP": {POS: ADP},
"PRON": {POS: PRON},
"SCONJ": {POS: SCONJ},
"PROPN": {POS: PROPN},
"DET": {POS: DET},
"SYM": {POS: SYM},
"INTJ": {POS: INTJ},
"PUNCT": {POS: PUNCT},
"NUM": {POS: NUM},
"AUX": {POS: AUX},
"X": {POS: X},
"CONJ": {POS: CONJ},
"ADJ": {POS: ADJ},
"VERB": {POS: VERB},
"PART": {POS: PART}
}
| // ... existing code ...
"CONJ": {POS: CONJ},
"ADJ": {POS: ADJ},
"VERB": {POS: VERB},
"PART": {POS: PART}
}
// ... rest of the code ... |
1b9c4935b2edf6601c2d75d8a2d318266de2d456 | circuits/tools/__init__.py | circuits/tools/__init__.py |
try:
from cStringIO import StringIO
except ImportError:
from StringIO import StringIO
def graph(x):
s = StringIO()
d = 0
i = 0
done = False
stack = []
visited = set()
children = list(x.components)
while not done:
if x not in visited:
if d:
s.write("%s%s\n" % (" " * d, "|"))
s.write("%s%s%s\n" % (" " * d, "|-", x))
else:
s.write(" .%s\n" % x)
if x.components:
d += 1
visited.add(x)
if i < len(children):
x = children[i]
i += 1
if x.components:
stack.append((i, children))
children = list(x.components)
i = 0
else:
if stack:
i, children = stack.pop()
d -= 1
else:
done = True
return s.getvalue()
|
try:
from cStringIO import StringIO
except ImportError:
from StringIO import StringIO
def graph(x):
s = StringIO()
d = 0
i = 0
done = False
stack = []
visited = set()
children = list(x.components)
while not done:
if x not in visited:
if d:
s.write("%s%s\n" % (" " * d, "|"))
s.write("%s%s%s\n" % (" " * d, "|-", x))
else:
s.write(" .%s\n" % x)
if x.components:
d += 1
visited.add(x)
if i < len(children):
x = children[i]
i += 1
if x.components:
stack.append((i, d, children))
children = list(x.components)
i = 0
else:
if stack:
i, d, children = stack.pop()
else:
done = True
return s.getvalue()
| Store the depth (d) on the stack and restore when backtracking | tools: Store the depth (d) on the stack and restore when backtracking
| Python | mit | treemo/circuits,treemo/circuits,eriol/circuits,treemo/circuits,eriol/circuits,nizox/circuits,eriol/circuits |
try:
from cStringIO import StringIO
except ImportError:
from StringIO import StringIO
def graph(x):
s = StringIO()
d = 0
i = 0
done = False
stack = []
visited = set()
children = list(x.components)
while not done:
if x not in visited:
if d:
s.write("%s%s\n" % (" " * d, "|"))
s.write("%s%s%s\n" % (" " * d, "|-", x))
else:
s.write(" .%s\n" % x)
if x.components:
d += 1
visited.add(x)
if i < len(children):
x = children[i]
i += 1
if x.components:
- stack.append((i, children))
+ stack.append((i, d, children))
children = list(x.components)
i = 0
else:
if stack:
- i, children = stack.pop()
+ i, d, children = stack.pop()
- d -= 1
else:
done = True
return s.getvalue()
| Store the depth (d) on the stack and restore when backtracking | ## Code Before:
try:
from cStringIO import StringIO
except ImportError:
from StringIO import StringIO
def graph(x):
s = StringIO()
d = 0
i = 0
done = False
stack = []
visited = set()
children = list(x.components)
while not done:
if x not in visited:
if d:
s.write("%s%s\n" % (" " * d, "|"))
s.write("%s%s%s\n" % (" " * d, "|-", x))
else:
s.write(" .%s\n" % x)
if x.components:
d += 1
visited.add(x)
if i < len(children):
x = children[i]
i += 1
if x.components:
stack.append((i, children))
children = list(x.components)
i = 0
else:
if stack:
i, children = stack.pop()
d -= 1
else:
done = True
return s.getvalue()
## Instruction:
Store the depth (d) on the stack and restore when backtracking
## Code After:
try:
from cStringIO import StringIO
except ImportError:
from StringIO import StringIO
def graph(x):
s = StringIO()
d = 0
i = 0
done = False
stack = []
visited = set()
children = list(x.components)
while not done:
if x not in visited:
if d:
s.write("%s%s\n" % (" " * d, "|"))
s.write("%s%s%s\n" % (" " * d, "|-", x))
else:
s.write(" .%s\n" % x)
if x.components:
d += 1
visited.add(x)
if i < len(children):
x = children[i]
i += 1
if x.components:
stack.append((i, d, children))
children = list(x.components)
i = 0
else:
if stack:
i, d, children = stack.pop()
else:
done = True
return s.getvalue()
| # ... existing code ...
i += 1
if x.components:
stack.append((i, d, children))
children = list(x.components)
i = 0
# ... modified code ...
else:
if stack:
i, d, children = stack.pop()
else:
done = True
# ... rest of the code ... |
07b7efdc848deeb7634f23cdb878774a99c9c535 | ambassador/tests/t_grpc_bridge.py | ambassador/tests/t_grpc_bridge.py | import json
from kat.harness import Query
from abstract_tests import AmbassadorTest, ServiceType, EGRPC
class AcceptanceGrpcBridgeTest(AmbassadorTest):
target: ServiceType
def init(self):
self.target = EGRPC()
def config(self):
yield self, self.format("""
---
apiVersion: ambassador/v0
kind: Module
name: ambassador
config:
enable_grpc_http11_bridge: True
""")
yield self, self.format("""
---
apiVersion: ambassador/v0
kind: Mapping
grpc: True
prefix: /echo.EchoService/
rewrite: /echo.EchoService/
name: {self.target.path.k8s}
service: {self.target.path.k8s}
""")
def queries(self):
# [0]
yield Query(self.url("echo.EchoService/Echo"), headers={ "content-type": "application/grpc",
"requested-status": "0" }, expected=200)
# [1]
yield Query(self.url("echo.EchoService/Echo"), headers={ "content-type": "application/grpc",
"requested-status": "7" }, expected=200)
def check(self):
# [0]
assert self.results[0].status == 200
assert self.results[0].headers["Grpc-Status"] == ["0"]
# [0]
assert self.results[1].status == 200
assert self.results[1].headers["Grpc-Status"] == ["7"]
| from kat.harness import Query
from abstract_tests import AmbassadorTest, ServiceType, EGRPC
class AcceptanceGrpcBridgeTest(AmbassadorTest):
target: ServiceType
def init(self):
self.target = EGRPC()
def config(self):
yield self, self.format("""
---
apiVersion: ambassador/v0
kind: Module
name: ambassador
config:
enable_grpc_http11_bridge: True
""")
yield self, self.format("""
---
apiVersion: ambassador/v0
kind: Mapping
grpc: True
prefix: /echo.EchoService/
rewrite: /echo.EchoService/
name: {self.target.path.k8s}
service: {self.target.path.k8s}
""")
def queries(self):
# [0]
yield Query(self.url("echo.EchoService/Echo"),
headers={ "content-type": "application/grpc", "requested-status": "0" },
expected=200)
# [1]
yield Query(self.url("echo.EchoService/Echo"),
headers={ "content-type": "application/grpc", "requested-status": "7" },
expected=200)
def check(self):
# [0]
assert self.results[0].headers["Grpc-Status"] == ["0"]
# [1]
assert self.results[1].headers["Grpc-Status"] == ["7"]
| Clean up existing gRPC bridge test (make it like t_tcpmapping.py) | Clean up existing gRPC bridge test (make it like t_tcpmapping.py)
- Remove redundant HTTP status assertions
- Adjust formatting
- Remove unused import
| Python | apache-2.0 | datawire/ambassador,datawire/ambassador,datawire/ambassador,datawire/ambassador,datawire/ambassador | - import json
-
from kat.harness import Query
from abstract_tests import AmbassadorTest, ServiceType, EGRPC
class AcceptanceGrpcBridgeTest(AmbassadorTest):
target: ServiceType
def init(self):
self.target = EGRPC()
def config(self):
yield self, self.format("""
---
apiVersion: ambassador/v0
kind: Module
name: ambassador
config:
enable_grpc_http11_bridge: True
""")
yield self, self.format("""
---
apiVersion: ambassador/v0
kind: Mapping
grpc: True
prefix: /echo.EchoService/
rewrite: /echo.EchoService/
name: {self.target.path.k8s}
service: {self.target.path.k8s}
""")
def queries(self):
# [0]
- yield Query(self.url("echo.EchoService/Echo"), headers={ "content-type": "application/grpc",
- "requested-status": "0" }, expected=200)
+ yield Query(self.url("echo.EchoService/Echo"),
+ headers={ "content-type": "application/grpc", "requested-status": "0" },
+ expected=200)
- # [1]
+ # [1]
- yield Query(self.url("echo.EchoService/Echo"), headers={ "content-type": "application/grpc",
- "requested-status": "7" }, expected=200)
+ yield Query(self.url("echo.EchoService/Echo"),
+ headers={ "content-type": "application/grpc", "requested-status": "7" },
+ expected=200)
def check(self):
# [0]
- assert self.results[0].status == 200
assert self.results[0].headers["Grpc-Status"] == ["0"]
- # [0]
+ # [1]
- assert self.results[1].status == 200
assert self.results[1].headers["Grpc-Status"] == ["7"]
| Clean up existing gRPC bridge test (make it like t_tcpmapping.py) | ## Code Before:
import json
from kat.harness import Query
from abstract_tests import AmbassadorTest, ServiceType, EGRPC
class AcceptanceGrpcBridgeTest(AmbassadorTest):
target: ServiceType
def init(self):
self.target = EGRPC()
def config(self):
yield self, self.format("""
---
apiVersion: ambassador/v0
kind: Module
name: ambassador
config:
enable_grpc_http11_bridge: True
""")
yield self, self.format("""
---
apiVersion: ambassador/v0
kind: Mapping
grpc: True
prefix: /echo.EchoService/
rewrite: /echo.EchoService/
name: {self.target.path.k8s}
service: {self.target.path.k8s}
""")
def queries(self):
# [0]
yield Query(self.url("echo.EchoService/Echo"), headers={ "content-type": "application/grpc",
"requested-status": "0" }, expected=200)
# [1]
yield Query(self.url("echo.EchoService/Echo"), headers={ "content-type": "application/grpc",
"requested-status": "7" }, expected=200)
def check(self):
# [0]
assert self.results[0].status == 200
assert self.results[0].headers["Grpc-Status"] == ["0"]
# [0]
assert self.results[1].status == 200
assert self.results[1].headers["Grpc-Status"] == ["7"]
## Instruction:
Clean up existing gRPC bridge test (make it like t_tcpmapping.py)
## Code After:
from kat.harness import Query
from abstract_tests import AmbassadorTest, ServiceType, EGRPC
class AcceptanceGrpcBridgeTest(AmbassadorTest):
target: ServiceType
def init(self):
self.target = EGRPC()
def config(self):
yield self, self.format("""
---
apiVersion: ambassador/v0
kind: Module
name: ambassador
config:
enable_grpc_http11_bridge: True
""")
yield self, self.format("""
---
apiVersion: ambassador/v0
kind: Mapping
grpc: True
prefix: /echo.EchoService/
rewrite: /echo.EchoService/
name: {self.target.path.k8s}
service: {self.target.path.k8s}
""")
def queries(self):
# [0]
yield Query(self.url("echo.EchoService/Echo"),
headers={ "content-type": "application/grpc", "requested-status": "0" },
expected=200)
# [1]
yield Query(self.url("echo.EchoService/Echo"),
headers={ "content-type": "application/grpc", "requested-status": "7" },
expected=200)
def check(self):
# [0]
assert self.results[0].headers["Grpc-Status"] == ["0"]
# [1]
assert self.results[1].headers["Grpc-Status"] == ["7"]
| # ... existing code ...
from kat.harness import Query
# ... modified code ...
def queries(self):
# [0]
yield Query(self.url("echo.EchoService/Echo"),
headers={ "content-type": "application/grpc", "requested-status": "0" },
expected=200)
# [1]
yield Query(self.url("echo.EchoService/Echo"),
headers={ "content-type": "application/grpc", "requested-status": "7" },
expected=200)
def check(self):
# [0]
assert self.results[0].headers["Grpc-Status"] == ["0"]
# [1]
assert self.results[1].headers["Grpc-Status"] == ["7"]
# ... rest of the code ... |
e87490ea157f4882f644329e4b447f51c0a2acb3 | benchmarks/bench_vectorize.py | benchmarks/bench_vectorize.py |
import numpy as np
from numba import vectorize
@vectorize(["float32(float32, float32)",
"float64(float64, float64)",
"complex64(complex64, complex64)",
"complex128(complex128, complex128)"])
def mul(x, y):
return x * y
class TimeSuite:
n = 10000
dtypes = ('float32', 'float64', 'complex64', 'complex128')
def setup(self):
self.samples = {}
self.out = {}
for dtype in self.dtypes:
self.samples[dtype] = np.linspace(0, 1, self.n, dtype=dtype)
self.out[dtype] = np.zeros(self.n, dtype=dtype)
def _binary_func(func, dtype):
def f(self):
func(self.samples[dtype], self.samples[dtype], self.out[dtype])
return f
for dtype in dtypes:
locals()['time_mul_%s' % dtype] = _binary_func(mul, dtype)
del _binary_func
|
import numpy as np
from numba import vectorize
@vectorize(["float32(float32, float32)",
"float64(float64, float64)",
"complex64(complex64, complex64)",
"complex128(complex128, complex128)"])
def mul(x, y):
return x * y
@vectorize(["float32(float32, float32)",
"float64(float64, float64)"])
def rel_diff(x, y):
# XXX for float32 performance, we should write `np.float32(2)`, but
# that's not the natural way to write this code...
return 2 * (x - y) / (x + y)
class TimeSuite:
n = 10000
dtypes = ('float32', 'float64', 'complex64', 'complex128')
def setup(self):
self.samples = {}
self.out = {}
for dtype in self.dtypes:
self.samples[dtype] = np.linspace(0.1, 1, self.n, dtype=dtype)
self.out[dtype] = np.zeros(self.n, dtype=dtype)
def _binary_func(func, dtype):
def f(self):
func(self.samples[dtype], self.samples[dtype], self.out[dtype])
return f
for dtype in dtypes:
locals()['time_mul_%s' % dtype] = _binary_func(mul, dtype)
time_rel_diff_float32 = _binary_func(rel_diff, 'float32')
time_rel_diff_float64 = _binary_func(rel_diff, 'float64')
del _binary_func
| Add a relative difference vectorization benchmark | Add a relative difference vectorization benchmark
| Python | bsd-2-clause | gmarkall/numba-benchmark,numba/numba-benchmark |
import numpy as np
from numba import vectorize
@vectorize(["float32(float32, float32)",
"float64(float64, float64)",
"complex64(complex64, complex64)",
"complex128(complex128, complex128)"])
def mul(x, y):
return x * y
+ @vectorize(["float32(float32, float32)",
+ "float64(float64, float64)"])
+ def rel_diff(x, y):
+ # XXX for float32 performance, we should write `np.float32(2)`, but
+ # that's not the natural way to write this code...
+ return 2 * (x - y) / (x + y)
+
+
class TimeSuite:
n = 10000
dtypes = ('float32', 'float64', 'complex64', 'complex128')
def setup(self):
self.samples = {}
self.out = {}
for dtype in self.dtypes:
- self.samples[dtype] = np.linspace(0, 1, self.n, dtype=dtype)
+ self.samples[dtype] = np.linspace(0.1, 1, self.n, dtype=dtype)
self.out[dtype] = np.zeros(self.n, dtype=dtype)
def _binary_func(func, dtype):
def f(self):
func(self.samples[dtype], self.samples[dtype], self.out[dtype])
return f
for dtype in dtypes:
locals()['time_mul_%s' % dtype] = _binary_func(mul, dtype)
+ time_rel_diff_float32 = _binary_func(rel_diff, 'float32')
+ time_rel_diff_float64 = _binary_func(rel_diff, 'float64')
+
del _binary_func
| Add a relative difference vectorization benchmark | ## Code Before:
import numpy as np
from numba import vectorize
@vectorize(["float32(float32, float32)",
"float64(float64, float64)",
"complex64(complex64, complex64)",
"complex128(complex128, complex128)"])
def mul(x, y):
return x * y
class TimeSuite:
n = 10000
dtypes = ('float32', 'float64', 'complex64', 'complex128')
def setup(self):
self.samples = {}
self.out = {}
for dtype in self.dtypes:
self.samples[dtype] = np.linspace(0, 1, self.n, dtype=dtype)
self.out[dtype] = np.zeros(self.n, dtype=dtype)
def _binary_func(func, dtype):
def f(self):
func(self.samples[dtype], self.samples[dtype], self.out[dtype])
return f
for dtype in dtypes:
locals()['time_mul_%s' % dtype] = _binary_func(mul, dtype)
del _binary_func
## Instruction:
Add a relative difference vectorization benchmark
## Code After:
import numpy as np
from numba import vectorize
@vectorize(["float32(float32, float32)",
"float64(float64, float64)",
"complex64(complex64, complex64)",
"complex128(complex128, complex128)"])
def mul(x, y):
return x * y
@vectorize(["float32(float32, float32)",
"float64(float64, float64)"])
def rel_diff(x, y):
# XXX for float32 performance, we should write `np.float32(2)`, but
# that's not the natural way to write this code...
return 2 * (x - y) / (x + y)
class TimeSuite:
n = 10000
dtypes = ('float32', 'float64', 'complex64', 'complex128')
def setup(self):
self.samples = {}
self.out = {}
for dtype in self.dtypes:
self.samples[dtype] = np.linspace(0.1, 1, self.n, dtype=dtype)
self.out[dtype] = np.zeros(self.n, dtype=dtype)
def _binary_func(func, dtype):
def f(self):
func(self.samples[dtype], self.samples[dtype], self.out[dtype])
return f
for dtype in dtypes:
locals()['time_mul_%s' % dtype] = _binary_func(mul, dtype)
time_rel_diff_float32 = _binary_func(rel_diff, 'float32')
time_rel_diff_float64 = _binary_func(rel_diff, 'float64')
del _binary_func
| // ... existing code ...
@vectorize(["float32(float32, float32)",
"float64(float64, float64)"])
def rel_diff(x, y):
# XXX for float32 performance, we should write `np.float32(2)`, but
# that's not the natural way to write this code...
return 2 * (x - y) / (x + y)
class TimeSuite:
// ... modified code ...
self.out = {}
for dtype in self.dtypes:
self.samples[dtype] = np.linspace(0.1, 1, self.n, dtype=dtype)
self.out[dtype] = np.zeros(self.n, dtype=dtype)
...
locals()['time_mul_%s' % dtype] = _binary_func(mul, dtype)
time_rel_diff_float32 = _binary_func(rel_diff, 'float32')
time_rel_diff_float64 = _binary_func(rel_diff, 'float64')
del _binary_func
// ... rest of the code ... |
72045f86b25b396160e1a4c9237e977ed575afb2 | apps/catalogue/constants.py | apps/catalogue/constants.py | from django.utils.translation import ugettext_lazy as _
LICENSES = {
'http://creativecommons.org/licenses/by-sa/3.0/': {
'icon': 'cc-by-sa',
'description': _('Creative Commons Attribution-ShareAlike 3.0 Unported'),
},
}
# Those will be generated only for books with own HTML.
EBOOK_FORMATS_WITHOUT_CHILDREN = ['txt', 'fb2']
# Those will be generated for all books.
EBOOK_FORMATS_WITH_CHILDREN = ['pdf', 'epub', 'mobi']
# Those will be generated when inherited cover changes.
EBOOK_FORMATS_WITH_COVERS = ['pdf', 'epub', 'mobi']
EBOOK_FORMATS = EBOOK_FORMATS_WITHOUT_CHILDREN + EBOOK_FORMATS_WITH_CHILDREN
| from django.utils.translation import ugettext_lazy as _
LICENSES = {
'http://creativecommons.org/licenses/by-sa/3.0/': {
'icon': 'cc-by-sa',
'description': _('Creative Commons Attribution-ShareAlike 3.0 Unported'),
},
}
LICENSES['http://creativecommons.org/licenses/by-sa/3.0/deed.pl'] = \
LICENSES['http://creativecommons.org/licenses/by-sa/3.0/']
# Those will be generated only for books with own HTML.
EBOOK_FORMATS_WITHOUT_CHILDREN = ['txt', 'fb2']
# Those will be generated for all books.
EBOOK_FORMATS_WITH_CHILDREN = ['pdf', 'epub', 'mobi']
# Those will be generated when inherited cover changes.
EBOOK_FORMATS_WITH_COVERS = ['pdf', 'epub', 'mobi']
EBOOK_FORMATS = EBOOK_FORMATS_WITHOUT_CHILDREN + EBOOK_FORMATS_WITH_CHILDREN
| Support for 'deed.pl' license URL. | Support for 'deed.pl' license URL.
| Python | agpl-3.0 | fnp/wolnelektury,fnp/wolnelektury,fnp/wolnelektury,fnp/wolnelektury | from django.utils.translation import ugettext_lazy as _
LICENSES = {
'http://creativecommons.org/licenses/by-sa/3.0/': {
'icon': 'cc-by-sa',
'description': _('Creative Commons Attribution-ShareAlike 3.0 Unported'),
},
}
+ LICENSES['http://creativecommons.org/licenses/by-sa/3.0/deed.pl'] = \
+ LICENSES['http://creativecommons.org/licenses/by-sa/3.0/']
# Those will be generated only for books with own HTML.
EBOOK_FORMATS_WITHOUT_CHILDREN = ['txt', 'fb2']
# Those will be generated for all books.
EBOOK_FORMATS_WITH_CHILDREN = ['pdf', 'epub', 'mobi']
# Those will be generated when inherited cover changes.
EBOOK_FORMATS_WITH_COVERS = ['pdf', 'epub', 'mobi']
EBOOK_FORMATS = EBOOK_FORMATS_WITHOUT_CHILDREN + EBOOK_FORMATS_WITH_CHILDREN
| Support for 'deed.pl' license URL. | ## Code Before:
from django.utils.translation import ugettext_lazy as _
LICENSES = {
'http://creativecommons.org/licenses/by-sa/3.0/': {
'icon': 'cc-by-sa',
'description': _('Creative Commons Attribution-ShareAlike 3.0 Unported'),
},
}
# Those will be generated only for books with own HTML.
EBOOK_FORMATS_WITHOUT_CHILDREN = ['txt', 'fb2']
# Those will be generated for all books.
EBOOK_FORMATS_WITH_CHILDREN = ['pdf', 'epub', 'mobi']
# Those will be generated when inherited cover changes.
EBOOK_FORMATS_WITH_COVERS = ['pdf', 'epub', 'mobi']
EBOOK_FORMATS = EBOOK_FORMATS_WITHOUT_CHILDREN + EBOOK_FORMATS_WITH_CHILDREN
## Instruction:
Support for 'deed.pl' license URL.
## Code After:
from django.utils.translation import ugettext_lazy as _
LICENSES = {
'http://creativecommons.org/licenses/by-sa/3.0/': {
'icon': 'cc-by-sa',
'description': _('Creative Commons Attribution-ShareAlike 3.0 Unported'),
},
}
LICENSES['http://creativecommons.org/licenses/by-sa/3.0/deed.pl'] = \
LICENSES['http://creativecommons.org/licenses/by-sa/3.0/']
# Those will be generated only for books with own HTML.
EBOOK_FORMATS_WITHOUT_CHILDREN = ['txt', 'fb2']
# Those will be generated for all books.
EBOOK_FORMATS_WITH_CHILDREN = ['pdf', 'epub', 'mobi']
# Those will be generated when inherited cover changes.
EBOOK_FORMATS_WITH_COVERS = ['pdf', 'epub', 'mobi']
EBOOK_FORMATS = EBOOK_FORMATS_WITHOUT_CHILDREN + EBOOK_FORMATS_WITH_CHILDREN
| ...
},
}
LICENSES['http://creativecommons.org/licenses/by-sa/3.0/deed.pl'] = \
LICENSES['http://creativecommons.org/licenses/by-sa/3.0/']
# Those will be generated only for books with own HTML.
... |
8869eba1f74e677d1802aad0cc2592344ab81000 | podium/talks/models.py | podium/talks/models.py | from django.db import models
from django.urls import reverse
TALK_STATUS_CHOICES = (
('S', 'Submitted'),
('A', 'Approved'),
('R', 'Rejected'),
('C', 'Confirmed'),
)
class Talk(models.Model):
speaker_name = models.CharField(max_length=1000)
speaker_email = models.CharField(max_length=1000)
title = models.CharField(max_length=1000)
description = models.TextField()
sessions_available = models.ManyToManyField(
'Session', related_name='talks_available')
status = models.CharField(
max_length=1, choices=TALK_STATUS_CHOICES,
default='S')
def get_absolute_url(self):
return reverse('talks-talks-id', args=[self.id])
def __str__(self):
return self.speaker_name
class Session(models.Model):
date = models.DateField()
description = models.TextField(
blank=True, help_text='Any special theme or info about the session.')
def __str__(self):
return '{} - {} '.format(self.date, self.description)
def approved_talks(self):
sets = [
self.talks_available.filter(status=status) for status in ('A', 'C')
]
return sets[0].union(sets[1])
def get_absolute_url(self):
return reverse('talks-sessions-id', args=[self.id])
| from django.db import models
from django.urls import reverse
TALK_STATUS_CHOICES = (
('S', 'Submitted'),
('A', 'Approved'),
('R', 'Rejected'),
('C', 'Confirmed'),
)
class Talk(models.Model):
speaker_name = models.CharField(max_length=1000)
speaker_email = models.CharField(max_length=1000)
title = models.CharField(max_length=1000)
description = models.TextField()
sessions_available = models.ManyToManyField(
'Session', related_name='talks_available')
status = models.CharField(
max_length=1, choices=TALK_STATUS_CHOICES,
default='S')
def get_absolute_url(self):
return reverse('talks-talks-id', args=[self.id])
def __str__(self):
return self.speaker_name
class Session(models.Model):
date = models.DateField()
description = models.TextField(
blank=True, help_text='Any special theme or info about the session.')
def __str__(self):
return '{} - {} '.format(self.date, self.description)
def approved_talks(self):
return self.talks_available.filter(status__in=('A', 'C'))
def get_absolute_url(self):
return reverse('talks-sessions-id', args=[self.id])
| Use a filter field lookup | Use a filter field lookup
Looks like I forgot to do this when JR suggested it.
| Python | mit | pyatl/podium-django,pyatl/podium-django,pyatl/podium-django | from django.db import models
from django.urls import reverse
TALK_STATUS_CHOICES = (
('S', 'Submitted'),
('A', 'Approved'),
('R', 'Rejected'),
('C', 'Confirmed'),
)
class Talk(models.Model):
speaker_name = models.CharField(max_length=1000)
speaker_email = models.CharField(max_length=1000)
title = models.CharField(max_length=1000)
description = models.TextField()
sessions_available = models.ManyToManyField(
'Session', related_name='talks_available')
status = models.CharField(
max_length=1, choices=TALK_STATUS_CHOICES,
default='S')
def get_absolute_url(self):
return reverse('talks-talks-id', args=[self.id])
def __str__(self):
return self.speaker_name
class Session(models.Model):
date = models.DateField()
description = models.TextField(
blank=True, help_text='Any special theme or info about the session.')
def __str__(self):
return '{} - {} '.format(self.date, self.description)
def approved_talks(self):
- sets = [
- self.talks_available.filter(status=status) for status in ('A', 'C')
+ return self.talks_available.filter(status__in=('A', 'C'))
- ]
- return sets[0].union(sets[1])
def get_absolute_url(self):
return reverse('talks-sessions-id', args=[self.id])
| Use a filter field lookup | ## Code Before:
from django.db import models
from django.urls import reverse
TALK_STATUS_CHOICES = (
('S', 'Submitted'),
('A', 'Approved'),
('R', 'Rejected'),
('C', 'Confirmed'),
)
class Talk(models.Model):
speaker_name = models.CharField(max_length=1000)
speaker_email = models.CharField(max_length=1000)
title = models.CharField(max_length=1000)
description = models.TextField()
sessions_available = models.ManyToManyField(
'Session', related_name='talks_available')
status = models.CharField(
max_length=1, choices=TALK_STATUS_CHOICES,
default='S')
def get_absolute_url(self):
return reverse('talks-talks-id', args=[self.id])
def __str__(self):
return self.speaker_name
class Session(models.Model):
date = models.DateField()
description = models.TextField(
blank=True, help_text='Any special theme or info about the session.')
def __str__(self):
return '{} - {} '.format(self.date, self.description)
def approved_talks(self):
sets = [
self.talks_available.filter(status=status) for status in ('A', 'C')
]
return sets[0].union(sets[1])
def get_absolute_url(self):
return reverse('talks-sessions-id', args=[self.id])
## Instruction:
Use a filter field lookup
## Code After:
from django.db import models
from django.urls import reverse
TALK_STATUS_CHOICES = (
('S', 'Submitted'),
('A', 'Approved'),
('R', 'Rejected'),
('C', 'Confirmed'),
)
class Talk(models.Model):
speaker_name = models.CharField(max_length=1000)
speaker_email = models.CharField(max_length=1000)
title = models.CharField(max_length=1000)
description = models.TextField()
sessions_available = models.ManyToManyField(
'Session', related_name='talks_available')
status = models.CharField(
max_length=1, choices=TALK_STATUS_CHOICES,
default='S')
def get_absolute_url(self):
return reverse('talks-talks-id', args=[self.id])
def __str__(self):
return self.speaker_name
class Session(models.Model):
date = models.DateField()
description = models.TextField(
blank=True, help_text='Any special theme or info about the session.')
def __str__(self):
return '{} - {} '.format(self.date, self.description)
def approved_talks(self):
return self.talks_available.filter(status__in=('A', 'C'))
def get_absolute_url(self):
return reverse('talks-sessions-id', args=[self.id])
| # ... existing code ...
def approved_talks(self):
return self.talks_available.filter(status__in=('A', 'C'))
def get_absolute_url(self):
# ... rest of the code ... |
929909513e71282de388cf4e93476ba614e6c0c5 | Malcom/feeds/malwaredomains.py | Malcom/feeds/malwaredomains.py | import urllib2
import re
from Malcom.model.datatypes import Hostname, Evil
from feed import Feed
import Malcom.auxiliary.toolbox as toolbox
class MalwareDomains(Feed):
def __init__(self, name):
super(MalwareDomains, self).__init__(name)
self.source = "http://mirror1.malwaredomains.com/files/domains.txt"
self.description = "Malware domains blocklist"
self.confidence = 50
self.name = "MalwareDomains"
def update(self):
self.update_lines()
def analyze(self, line):
if line.startswith('#') or line.startswith('\n'):
return
splitted_mdl = line.split('\t')
# 20151201 agasi-story.info malicious blog.dynamoo.com 20131130 20121201 20120521 20110217
# Create the new hostname and store it in the DB
hostname = Hostname(hostname=splitted_mdl[2])
if hostname['value'] == None: return # hostname not found
evil = Evil()
evil['value'] = "Malware domain blocklist (%s)" % hostname['value']
evil['tags'] = ['malwaredomains', splitted_mdl[3]]
evil['reference'] = splitted_mdl[4]
return hostname, evil
| import urllib2
import re
from Malcom.model.datatypes import Hostname, Evil
from feed import Feed
import Malcom.auxiliary.toolbox as toolbox
class MalwareDomains(Feed):
def __init__(self, name):
super(MalwareDomains, self).__init__(name)
self.source = "http://mirror1.malwaredomains.com/files/domains.txt"
self.description = "Malware domains blocklist"
self.confidence = 50
self.name = "MalwareDomains"
def update(self):
self.update_lines()
def analyze(self, line):
if line.startswith('#') or line.startswith('\n'):
return
splitted_mdl = line.split('\t')
# 20151201 agasi-story.info malicious blog.dynamoo.com 20131130 20121201 20120521 20110217
# Create the new hostname and store it in the DB
hostname = Hostname(hostname=splitted_mdl[2])
if hostname['value'] == None: return # hostname not found
evil = Evil()
evil['value'] = "Malware domain blocklist (%s)" % hostname['value']
evil['tags'] = ['malwaredomains', re.sub(r'[^\w]', '', splitted_mdl[3])]
evil['reference'] = splitted_mdl[4]
return hostname, evil
| Deal with MalwareDomains non-ASCII characters | Deal with MalwareDomains non-ASCII characters
| Python | apache-2.0 | yeti-platform/yeti,yeti-platform/yeti,yeti-platform/yeti,yeti-platform/yeti | import urllib2
import re
from Malcom.model.datatypes import Hostname, Evil
from feed import Feed
import Malcom.auxiliary.toolbox as toolbox
class MalwareDomains(Feed):
def __init__(self, name):
super(MalwareDomains, self).__init__(name)
self.source = "http://mirror1.malwaredomains.com/files/domains.txt"
self.description = "Malware domains blocklist"
self.confidence = 50
self.name = "MalwareDomains"
def update(self):
self.update_lines()
def analyze(self, line):
if line.startswith('#') or line.startswith('\n'):
return
splitted_mdl = line.split('\t')
# 20151201 agasi-story.info malicious blog.dynamoo.com 20131130 20121201 20120521 20110217
# Create the new hostname and store it in the DB
hostname = Hostname(hostname=splitted_mdl[2])
if hostname['value'] == None: return # hostname not found
evil = Evil()
evil['value'] = "Malware domain blocklist (%s)" % hostname['value']
- evil['tags'] = ['malwaredomains', splitted_mdl[3]]
+ evil['tags'] = ['malwaredomains', re.sub(r'[^\w]', '', splitted_mdl[3])]
evil['reference'] = splitted_mdl[4]
return hostname, evil
| Deal with MalwareDomains non-ASCII characters | ## Code Before:
import urllib2
import re
from Malcom.model.datatypes import Hostname, Evil
from feed import Feed
import Malcom.auxiliary.toolbox as toolbox
class MalwareDomains(Feed):
def __init__(self, name):
super(MalwareDomains, self).__init__(name)
self.source = "http://mirror1.malwaredomains.com/files/domains.txt"
self.description = "Malware domains blocklist"
self.confidence = 50
self.name = "MalwareDomains"
def update(self):
self.update_lines()
def analyze(self, line):
if line.startswith('#') or line.startswith('\n'):
return
splitted_mdl = line.split('\t')
# 20151201 agasi-story.info malicious blog.dynamoo.com 20131130 20121201 20120521 20110217
# Create the new hostname and store it in the DB
hostname = Hostname(hostname=splitted_mdl[2])
if hostname['value'] == None: return # hostname not found
evil = Evil()
evil['value'] = "Malware domain blocklist (%s)" % hostname['value']
evil['tags'] = ['malwaredomains', splitted_mdl[3]]
evil['reference'] = splitted_mdl[4]
return hostname, evil
## Instruction:
Deal with MalwareDomains non-ASCII characters
## Code After:
import urllib2
import re
from Malcom.model.datatypes import Hostname, Evil
from feed import Feed
import Malcom.auxiliary.toolbox as toolbox
class MalwareDomains(Feed):
def __init__(self, name):
super(MalwareDomains, self).__init__(name)
self.source = "http://mirror1.malwaredomains.com/files/domains.txt"
self.description = "Malware domains blocklist"
self.confidence = 50
self.name = "MalwareDomains"
def update(self):
self.update_lines()
def analyze(self, line):
if line.startswith('#') or line.startswith('\n'):
return
splitted_mdl = line.split('\t')
# 20151201 agasi-story.info malicious blog.dynamoo.com 20131130 20121201 20120521 20110217
# Create the new hostname and store it in the DB
hostname = Hostname(hostname=splitted_mdl[2])
if hostname['value'] == None: return # hostname not found
evil = Evil()
evil['value'] = "Malware domain blocklist (%s)" % hostname['value']
evil['tags'] = ['malwaredomains', re.sub(r'[^\w]', '', splitted_mdl[3])]
evil['reference'] = splitted_mdl[4]
return hostname, evil
| # ... existing code ...
evil = Evil()
evil['value'] = "Malware domain blocklist (%s)" % hostname['value']
evil['tags'] = ['malwaredomains', re.sub(r'[^\w]', '', splitted_mdl[3])]
evil['reference'] = splitted_mdl[4]
# ... rest of the code ... |
c3745e7017c1788f4633d09ef4d29a37018b53d3 | populus/cli/main.py | populus/cli/main.py | import click
@click.group()
def main():
"""
Populus
"""
pass
| import click
CONTEXT_SETTINGS = dict(
# Support -h as a shortcut for --help
help_option_names=['-h', '--help'],
)
@click.group(context_settings=CONTEXT_SETTINGS)
def main():
"""
Populus
"""
pass
| Support -h as a shortcut for --help | CLI: Support -h as a shortcut for --help
| Python | mit | pipermerriam/populus,euri10/populus,euri10/populus,pipermerriam/populus,euri10/populus | import click
- @click.group()
+ CONTEXT_SETTINGS = dict(
+ # Support -h as a shortcut for --help
+ help_option_names=['-h', '--help'],
+ )
+
+
+ @click.group(context_settings=CONTEXT_SETTINGS)
def main():
"""
Populus
"""
pass
| Support -h as a shortcut for --help | ## Code Before:
import click
@click.group()
def main():
"""
Populus
"""
pass
## Instruction:
Support -h as a shortcut for --help
## Code After:
import click
CONTEXT_SETTINGS = dict(
# Support -h as a shortcut for --help
help_option_names=['-h', '--help'],
)
@click.group(context_settings=CONTEXT_SETTINGS)
def main():
"""
Populus
"""
pass
| // ... existing code ...
CONTEXT_SETTINGS = dict(
# Support -h as a shortcut for --help
help_option_names=['-h', '--help'],
)
@click.group(context_settings=CONTEXT_SETTINGS)
def main():
"""
// ... rest of the code ... |
ff725b4ae24c58cb126c1d49ce58a69d9b32d3b0 | app/soc/models/timeline.py | app/soc/models/timeline.py |
from google.appengine.ext import db
from django.utils.translation import ugettext
from soc.models import linkable
class Timeline(linkable.Linkable):
"""The Timeline Model, representing the timeline for a Program.
"""
program_start = db.DateTimeProperty(
verbose_name=ugettext('Program Start date'))
program_end = db.DateTimeProperty(
verbose_name=ugettext('Program End date'))
accepted_organization_announced_deadline = db.DateTimeProperty(
verbose_name=ugettext('Accepted Organizations Announced Deadline'))
student_signup_start = db.DateTimeProperty(
verbose_name=ugettext('Student Signup Start date'))
student_signup_end = db.DateTimeProperty(
verbose_name=ugettext('Student Signup End date'))
|
from google.appengine.ext import db
from django.utils.translation import ugettext
from soc.models import linkable
class Timeline(linkable.Linkable):
"""The Timeline Model, representing the timeline for a Program.
"""
program_start = db.DateTimeProperty(
verbose_name=ugettext('Program Start date'))
program_end = db.DateTimeProperty(
verbose_name=ugettext('Program End date'))
program_end.help_text = ugettext(
'After this date no data (such as profiles and forms) can be changed.')
accepted_organization_announced_deadline = db.DateTimeProperty(
verbose_name=ugettext('Accepted Organizations Announced Deadline'))
student_signup_start = db.DateTimeProperty(
verbose_name=ugettext('Student Signup Start date'))
student_signup_end = db.DateTimeProperty(
verbose_name=ugettext('Student Signup End date'))
| Add help text for program_end date. | Add help text for program_end date.
Fixes 1411.
| Python | apache-2.0 | rhyolight/nupic.son,rhyolight/nupic.son,rhyolight/nupic.son |
from google.appengine.ext import db
from django.utils.translation import ugettext
from soc.models import linkable
class Timeline(linkable.Linkable):
"""The Timeline Model, representing the timeline for a Program.
"""
program_start = db.DateTimeProperty(
verbose_name=ugettext('Program Start date'))
program_end = db.DateTimeProperty(
verbose_name=ugettext('Program End date'))
+ program_end.help_text = ugettext(
+ 'After this date no data (such as profiles and forms) can be changed.')
accepted_organization_announced_deadline = db.DateTimeProperty(
verbose_name=ugettext('Accepted Organizations Announced Deadline'))
student_signup_start = db.DateTimeProperty(
verbose_name=ugettext('Student Signup Start date'))
student_signup_end = db.DateTimeProperty(
verbose_name=ugettext('Student Signup End date'))
| Add help text for program_end date. | ## Code Before:
from google.appengine.ext import db
from django.utils.translation import ugettext
from soc.models import linkable
class Timeline(linkable.Linkable):
"""The Timeline Model, representing the timeline for a Program.
"""
program_start = db.DateTimeProperty(
verbose_name=ugettext('Program Start date'))
program_end = db.DateTimeProperty(
verbose_name=ugettext('Program End date'))
accepted_organization_announced_deadline = db.DateTimeProperty(
verbose_name=ugettext('Accepted Organizations Announced Deadline'))
student_signup_start = db.DateTimeProperty(
verbose_name=ugettext('Student Signup Start date'))
student_signup_end = db.DateTimeProperty(
verbose_name=ugettext('Student Signup End date'))
## Instruction:
Add help text for program_end date.
## Code After:
from google.appengine.ext import db
from django.utils.translation import ugettext
from soc.models import linkable
class Timeline(linkable.Linkable):
"""The Timeline Model, representing the timeline for a Program.
"""
program_start = db.DateTimeProperty(
verbose_name=ugettext('Program Start date'))
program_end = db.DateTimeProperty(
verbose_name=ugettext('Program End date'))
program_end.help_text = ugettext(
'After this date no data (such as profiles and forms) can be changed.')
accepted_organization_announced_deadline = db.DateTimeProperty(
verbose_name=ugettext('Accepted Organizations Announced Deadline'))
student_signup_start = db.DateTimeProperty(
verbose_name=ugettext('Student Signup Start date'))
student_signup_end = db.DateTimeProperty(
verbose_name=ugettext('Student Signup End date'))
| // ... existing code ...
program_end = db.DateTimeProperty(
verbose_name=ugettext('Program End date'))
program_end.help_text = ugettext(
'After this date no data (such as profiles and forms) can be changed.')
accepted_organization_announced_deadline = db.DateTimeProperty(
// ... rest of the code ... |
fc94ac89d2f602c381f4c882ec963995f3ce3043 | cla_frontend/apps/core/context_processors.py | cla_frontend/apps/core/context_processors.py | from django.conf import settings
def globals(request):
context = {
'app_title': 'Civil Legal Advice',
'proposition_title': 'Civil Legal Advice',
'phase': 'alpha',
'product_type': 'service',
'feedback_url': '#',
'ga_id': '',
'raven_config_site': settings.RAVEN_CONFIG['site'] or ''
}
if hasattr(request, 'zone') and request.zone:
context['app_base_template'] = '%s/base.html' % request.zone['name']
context['zone'] = request.zone
return context
| from django.conf import settings
def globals(request):
context = {
'app_title': 'Civil Legal Advice',
'proposition_title': 'Civil Legal Advice',
'phase': 'alpha',
'product_type': 'service',
'feedback_url': '#',
'ga_id': '',
'raven_config_site': settings.RAVEN_CONFIG['site'] or '',
'socketio_server_url': settings.SOCKETIO_SERVER_URL
}
if hasattr(request, 'zone') and request.zone:
context['app_base_template'] = '%s/base.html' % request.zone['name']
context['zone'] = request.zone
return context
| Make socketio server url a global context variable in Django | Make socketio server url a global context variable in Django
| Python | mit | ministryofjustice/cla_frontend,ministryofjustice/cla_frontend,ministryofjustice/cla_frontend,ministryofjustice/cla_frontend | from django.conf import settings
def globals(request):
context = {
'app_title': 'Civil Legal Advice',
'proposition_title': 'Civil Legal Advice',
'phase': 'alpha',
'product_type': 'service',
'feedback_url': '#',
'ga_id': '',
- 'raven_config_site': settings.RAVEN_CONFIG['site'] or ''
+ 'raven_config_site': settings.RAVEN_CONFIG['site'] or '',
+ 'socketio_server_url': settings.SOCKETIO_SERVER_URL
}
if hasattr(request, 'zone') and request.zone:
context['app_base_template'] = '%s/base.html' % request.zone['name']
context['zone'] = request.zone
return context
| Make socketio server url a global context variable in Django | ## Code Before:
from django.conf import settings
def globals(request):
context = {
'app_title': 'Civil Legal Advice',
'proposition_title': 'Civil Legal Advice',
'phase': 'alpha',
'product_type': 'service',
'feedback_url': '#',
'ga_id': '',
'raven_config_site': settings.RAVEN_CONFIG['site'] or ''
}
if hasattr(request, 'zone') and request.zone:
context['app_base_template'] = '%s/base.html' % request.zone['name']
context['zone'] = request.zone
return context
## Instruction:
Make socketio server url a global context variable in Django
## Code After:
from django.conf import settings
def globals(request):
context = {
'app_title': 'Civil Legal Advice',
'proposition_title': 'Civil Legal Advice',
'phase': 'alpha',
'product_type': 'service',
'feedback_url': '#',
'ga_id': '',
'raven_config_site': settings.RAVEN_CONFIG['site'] or '',
'socketio_server_url': settings.SOCKETIO_SERVER_URL
}
if hasattr(request, 'zone') and request.zone:
context['app_base_template'] = '%s/base.html' % request.zone['name']
context['zone'] = request.zone
return context
| # ... existing code ...
'feedback_url': '#',
'ga_id': '',
'raven_config_site': settings.RAVEN_CONFIG['site'] or '',
'socketio_server_url': settings.SOCKETIO_SERVER_URL
}
# ... rest of the code ... |
0bf00b40e84a5c5fbcdbeb7b81911998e3f1081a | src/idea/tests/smoke_tests.py | src/idea/tests/smoke_tests.py | import os
from django.utils import timezone
from django_webtest import WebTest
from exam.decorators import fixture
from exam.cases import Exam
from django.core.urlresolvers import reverse
class SmokeTest(Exam, WebTest):
csrf_checks = False
fixtures = ['state']
@fixture
def user(self):
try:
from collab.django_factories import UserF
return UserF(username="[email protected]", person__title='')
except ImportError:
from django.contrib.auth.models import User
user = User()
user.username = "[email protected]"
user.first_name = 'first'
user.last_name = 'last'
user.email = '"[email protected]"'
user.password = 'pbkdf2_sha256$10000$ggAKkiHobFL8$xQzwPeHNX1vWr9uNmZ/gKbd17uLGZVM8QNcgmaIEAUs='
user.is_staff = False
user.is_active = True
user.is_superuser = False
user.last_login = timezone.now()
user.date_joined = timezone.now()
user.save()
return user
def get(self, url):
return self.app.get(url, user=self.user)
def test_idea_home(self):
page = self.get(reverse('idea:idea_list'))
self.assertEquals(200, page.status_code)
| import os
from django.utils import timezone
from django_webtest import WebTest
from exam.decorators import fixture
from exam.cases import Exam
from django.core.urlresolvers import reverse
from django.contrib.auth.models import User
class SmokeTest(Exam, WebTest):
csrf_checks = False
fixtures = ['state', 'core-test-fixtures']
@fixture
def user(self):
user = User.objects.get(username="[email protected]")
return user
def get(self, url):
return self.app.get(url, user=self.user)
def test_idea_home(self):
page = self.get(reverse('idea:idea_list'))
self.assertEquals(200, page.status_code)
| Use fixtures for smoke tests | Use fixtures for smoke tests
| Python | cc0-1.0 | cmc333333/idea-box,m3brown/idea-box,18F/idea-box,cmc333333/idea-box,CapeSepias/idea-box,geomapdev/idea-box,CapeSepias/idea-box,cmc333333/idea-box,18F/idea-box,geomapdev/idea-box,geomapdev/idea-box,18F/idea-box,CapeSepias/idea-box,m3brown/idea-box | import os
from django.utils import timezone
from django_webtest import WebTest
from exam.decorators import fixture
from exam.cases import Exam
from django.core.urlresolvers import reverse
+ from django.contrib.auth.models import User
class SmokeTest(Exam, WebTest):
csrf_checks = False
- fixtures = ['state']
+ fixtures = ['state', 'core-test-fixtures']
@fixture
def user(self):
- try:
- from collab.django_factories import UserF
- return UserF(username="[email protected]", person__title='')
- except ImportError:
- from django.contrib.auth.models import User
- user = User()
- user.username = "[email protected]"
+ user = User.objects.get(username="[email protected]")
- user.first_name = 'first'
- user.last_name = 'last'
- user.email = '"[email protected]"'
- user.password = 'pbkdf2_sha256$10000$ggAKkiHobFL8$xQzwPeHNX1vWr9uNmZ/gKbd17uLGZVM8QNcgmaIEAUs='
- user.is_staff = False
- user.is_active = True
- user.is_superuser = False
- user.last_login = timezone.now()
- user.date_joined = timezone.now()
- user.save()
- return user
+ return user
def get(self, url):
return self.app.get(url, user=self.user)
def test_idea_home(self):
page = self.get(reverse('idea:idea_list'))
self.assertEquals(200, page.status_code)
| Use fixtures for smoke tests | ## Code Before:
import os
from django.utils import timezone
from django_webtest import WebTest
from exam.decorators import fixture
from exam.cases import Exam
from django.core.urlresolvers import reverse
class SmokeTest(Exam, WebTest):
csrf_checks = False
fixtures = ['state']
@fixture
def user(self):
try:
from collab.django_factories import UserF
return UserF(username="[email protected]", person__title='')
except ImportError:
from django.contrib.auth.models import User
user = User()
user.username = "[email protected]"
user.first_name = 'first'
user.last_name = 'last'
user.email = '"[email protected]"'
user.password = 'pbkdf2_sha256$10000$ggAKkiHobFL8$xQzwPeHNX1vWr9uNmZ/gKbd17uLGZVM8QNcgmaIEAUs='
user.is_staff = False
user.is_active = True
user.is_superuser = False
user.last_login = timezone.now()
user.date_joined = timezone.now()
user.save()
return user
def get(self, url):
return self.app.get(url, user=self.user)
def test_idea_home(self):
page = self.get(reverse('idea:idea_list'))
self.assertEquals(200, page.status_code)
## Instruction:
Use fixtures for smoke tests
## Code After:
import os
from django.utils import timezone
from django_webtest import WebTest
from exam.decorators import fixture
from exam.cases import Exam
from django.core.urlresolvers import reverse
from django.contrib.auth.models import User
class SmokeTest(Exam, WebTest):
csrf_checks = False
fixtures = ['state', 'core-test-fixtures']
@fixture
def user(self):
user = User.objects.get(username="[email protected]")
return user
def get(self, url):
return self.app.get(url, user=self.user)
def test_idea_home(self):
page = self.get(reverse('idea:idea_list'))
self.assertEquals(200, page.status_code)
| ...
from exam.cases import Exam
from django.core.urlresolvers import reverse
from django.contrib.auth.models import User
...
class SmokeTest(Exam, WebTest):
csrf_checks = False
fixtures = ['state', 'core-test-fixtures']
@fixture
def user(self):
user = User.objects.get(username="[email protected]")
return user
def get(self, url):
... |
bf2ace8bd6cb0c492ff4347f9c2fe10a003abaff | sqlalchemy_redshift/__init__.py | sqlalchemy_redshift/__init__.py | from pkg_resources import get_distribution, parse_version
try:
import psycopg2 # noqa: F401
if get_distribution('psycopg2').parsed_version < parse_version('2.5'):
raise ImportError('Minimum required version for psycopg2 is 2.5')
except ImportError:
raise ImportError(
'No module named psycopg2. Please install either '
'psycopg2 or psycopg2-binary package for CPython '
'or psycopg2cffi for Pypy.'
)
__version__ = get_distribution('sqlalchemy-redshift').version
from sqlalchemy.dialects import registry
registry.register("redshift", "sqlalchemy_redshift.dialect", "RedshiftDialect")
registry.register(
"redshift.psycopg2", "sqlalchemy_redshift.dialect", "RedshiftDialect"
)
| from pkg_resources import DistributionNotFound, get_distribution, parse_version
try:
import psycopg2 # noqa: F401
except ImportError:
raise ImportError(
'No module named psycopg2. Please install either '
'psycopg2 or psycopg2-binary package for CPython '
'or psycopg2cffi for Pypy.'
) from None
for package in ['psycopg2', 'psycopg2-binary', 'psycopg2cffi']:
try:
if get_distribution(package).parsed_version < parse_version('2.5'):
raise ImportError('Minimum required version for psycopg2 is 2.5')
break
except DistributionNotFound:
pass
else:
raise ImportError(
'A module was found named psycopg2, '
'but the version of it could not be checked '
'as it was neither the Python package psycopg2, '
'psycopg2-binary or psycopg2cffi.'
)
__version__ = get_distribution('sqlalchemy-redshift').version
from sqlalchemy.dialects import registry
registry.register("redshift", "sqlalchemy_redshift.dialect", "RedshiftDialect")
registry.register(
"redshift.psycopg2", "sqlalchemy_redshift.dialect", "RedshiftDialect"
)
| Check the version of any of the supported Psycopg2 packages | Check the version of any of the supported Psycopg2 packages
A check was introduced in commit 8e0c4857a1c08f257b95d3b1ee5f6eb795d55cdc which
would check what version of the 'psycopg2' Python (pip) package was installed
as the dependency was removed from setup.py.
The check would however only check the 'psycopg2' package and not the other two
supported providers of the psycopg2 module, which meant importing the
sqlalchemy_redshift module would throw an exception, even though they were
installed.
This changes the check to check for either of the three supported psycopg2
packages and throws an exception if any of them fail to validate.
| Python | mit | sqlalchemy-redshift/sqlalchemy-redshift,graingert/redshift_sqlalchemy,sqlalchemy-redshift/sqlalchemy-redshift | - from pkg_resources import get_distribution, parse_version
+ from pkg_resources import DistributionNotFound, get_distribution, parse_version
try:
import psycopg2 # noqa: F401
- if get_distribution('psycopg2').parsed_version < parse_version('2.5'):
- raise ImportError('Minimum required version for psycopg2 is 2.5')
except ImportError:
raise ImportError(
'No module named psycopg2. Please install either '
'psycopg2 or psycopg2-binary package for CPython '
'or psycopg2cffi for Pypy.'
+ ) from None
+
+ for package in ['psycopg2', 'psycopg2-binary', 'psycopg2cffi']:
+ try:
+ if get_distribution(package).parsed_version < parse_version('2.5'):
+ raise ImportError('Minimum required version for psycopg2 is 2.5')
+ break
+ except DistributionNotFound:
+ pass
+ else:
+ raise ImportError(
+ 'A module was found named psycopg2, '
+ 'but the version of it could not be checked '
+ 'as it was neither the Python package psycopg2, '
+ 'psycopg2-binary or psycopg2cffi.'
)
__version__ = get_distribution('sqlalchemy-redshift').version
from sqlalchemy.dialects import registry
registry.register("redshift", "sqlalchemy_redshift.dialect", "RedshiftDialect")
registry.register(
"redshift.psycopg2", "sqlalchemy_redshift.dialect", "RedshiftDialect"
)
| Check the version of any of the supported Psycopg2 packages | ## Code Before:
from pkg_resources import get_distribution, parse_version
try:
import psycopg2 # noqa: F401
if get_distribution('psycopg2').parsed_version < parse_version('2.5'):
raise ImportError('Minimum required version for psycopg2 is 2.5')
except ImportError:
raise ImportError(
'No module named psycopg2. Please install either '
'psycopg2 or psycopg2-binary package for CPython '
'or psycopg2cffi for Pypy.'
)
__version__ = get_distribution('sqlalchemy-redshift').version
from sqlalchemy.dialects import registry
registry.register("redshift", "sqlalchemy_redshift.dialect", "RedshiftDialect")
registry.register(
"redshift.psycopg2", "sqlalchemy_redshift.dialect", "RedshiftDialect"
)
## Instruction:
Check the version of any of the supported Psycopg2 packages
## Code After:
from pkg_resources import DistributionNotFound, get_distribution, parse_version
try:
import psycopg2 # noqa: F401
except ImportError:
raise ImportError(
'No module named psycopg2. Please install either '
'psycopg2 or psycopg2-binary package for CPython '
'or psycopg2cffi for Pypy.'
) from None
for package in ['psycopg2', 'psycopg2-binary', 'psycopg2cffi']:
try:
if get_distribution(package).parsed_version < parse_version('2.5'):
raise ImportError('Minimum required version for psycopg2 is 2.5')
break
except DistributionNotFound:
pass
else:
raise ImportError(
'A module was found named psycopg2, '
'but the version of it could not be checked '
'as it was neither the Python package psycopg2, '
'psycopg2-binary or psycopg2cffi.'
)
__version__ = get_distribution('sqlalchemy-redshift').version
from sqlalchemy.dialects import registry
registry.register("redshift", "sqlalchemy_redshift.dialect", "RedshiftDialect")
registry.register(
"redshift.psycopg2", "sqlalchemy_redshift.dialect", "RedshiftDialect"
)
| // ... existing code ...
from pkg_resources import DistributionNotFound, get_distribution, parse_version
try:
import psycopg2 # noqa: F401
except ImportError:
raise ImportError(
// ... modified code ...
'psycopg2 or psycopg2-binary package for CPython '
'or psycopg2cffi for Pypy.'
) from None
for package in ['psycopg2', 'psycopg2-binary', 'psycopg2cffi']:
try:
if get_distribution(package).parsed_version < parse_version('2.5'):
raise ImportError('Minimum required version for psycopg2 is 2.5')
break
except DistributionNotFound:
pass
else:
raise ImportError(
'A module was found named psycopg2, '
'but the version of it could not be checked '
'as it was neither the Python package psycopg2, '
'psycopg2-binary or psycopg2cffi.'
)
// ... rest of the code ... |
156d62f15963bc95f52db7eb1493fad6890e2fc7 | dadi/__init__.py | dadi/__init__.py | import numpy
# This gives a nicer printout for masked arrays.
numpy.ma.default_real_fill_value = numpy.nan
import IO
import Integration
import PhiManip
import SFS
import ms
try:
import os
__DIRECTORY__ = os.path.dirname(IO.__file__)
__svn_file__ = os.path.join(__DIRECTORY__, 'svnversion')
__SVNVERSION__ = file(__svn_file__).read().strip()
except:
__SVNVERSION__ = 'Unknown'
| import numpy
# This gives a nicer printout for masked arrays.
numpy.ma.default_real_fill_value = numpy.nan
import IO
import Integration
import PhiManip
import SFS
import ms
import Plotting
try:
import os
__DIRECTORY__ = os.path.dirname(IO.__file__)
__svn_file__ = os.path.join(__DIRECTORY__, 'svnversion')
__SVNVERSION__ = file(__svn_file__).read().strip()
except:
__SVNVERSION__ = 'Unknown'
| Add Plotting to default imports. | Add Plotting to default imports.
git-svn-id: 4c7b13231a96299fde701bb5dec4bd2aaf383fc6@43 979d6bd5-6d4d-0410-bece-f567c23bd345
| Python | bsd-3-clause | beni55/dadi,beni55/dadi,ChenHsiang/dadi,RyanGutenkunst/dadi,paulirish/dadi,yangjl/dadi,ChenHsiang/dadi,cheese1213/dadi,yangjl/dadi,cheese1213/dadi,paulirish/dadi,niuhuifei/dadi,niuhuifei/dadi,RyanGutenkunst/dadi | import numpy
# This gives a nicer printout for masked arrays.
numpy.ma.default_real_fill_value = numpy.nan
import IO
import Integration
import PhiManip
import SFS
import ms
+ import Plotting
try:
import os
__DIRECTORY__ = os.path.dirname(IO.__file__)
__svn_file__ = os.path.join(__DIRECTORY__, 'svnversion')
__SVNVERSION__ = file(__svn_file__).read().strip()
except:
__SVNVERSION__ = 'Unknown'
| Add Plotting to default imports. | ## Code Before:
import numpy
# This gives a nicer printout for masked arrays.
numpy.ma.default_real_fill_value = numpy.nan
import IO
import Integration
import PhiManip
import SFS
import ms
try:
import os
__DIRECTORY__ = os.path.dirname(IO.__file__)
__svn_file__ = os.path.join(__DIRECTORY__, 'svnversion')
__SVNVERSION__ = file(__svn_file__).read().strip()
except:
__SVNVERSION__ = 'Unknown'
## Instruction:
Add Plotting to default imports.
## Code After:
import numpy
# This gives a nicer printout for masked arrays.
numpy.ma.default_real_fill_value = numpy.nan
import IO
import Integration
import PhiManip
import SFS
import ms
import Plotting
try:
import os
__DIRECTORY__ = os.path.dirname(IO.__file__)
__svn_file__ = os.path.join(__DIRECTORY__, 'svnversion')
__SVNVERSION__ = file(__svn_file__).read().strip()
except:
__SVNVERSION__ = 'Unknown'
| // ... existing code ...
import SFS
import ms
import Plotting
try:
// ... rest of the code ... |
7ed3ba20aae568d0c12ec361210d1189ecd534cf | lazysignup/backends.py | lazysignup/backends.py | from django.contrib.auth.backends import ModelBackend
from django.contrib.auth.models import User
class LazySignupBackend(ModelBackend):
def authenticate(self, username=None):
users = [u for u in User.objects.filter(username=username)
if not u.has_usable_password()]
if len(users) != 1:
return None
return users[0]
def get_user(self, user_id):
# Annotate the user with our backend so it's always available,
# not just when authenticate() has been called. This will be
# used by the is_lazy_user filter.
user = super(LazySignupBackend, self).get_user(user_id)
if user:
user.backend = 'lazysignup.backends.LazySignupBackend'
return user
| from django.contrib.auth.backends import ModelBackend
from lazysignup.models import LazyUser
class LazySignupBackend(ModelBackend):
def authenticate(self, username=None):
lazy_users = LazyUser.objects.filter(
user__username=username
).select_related('user')
try:
return lazy_users[0].user
except IndexError:
return None
def get_user(self, user_id):
# Annotate the user with our backend so it's always available,
# not just when authenticate() has been called. This will be
# used by the is_lazy_user filter.
user = super(LazySignupBackend, self).get_user(user_id)
if user:
user.backend = 'lazysignup.backends.LazySignupBackend'
return user
| Remove the lazy signup backend's hard dependency on django.contrib.auth.user (and remove the inconsistency in checking for whether a user is lazy or not). | Remove the lazy signup backend's hard dependency on django.contrib.auth.user (and remove the inconsistency in checking for whether a user is lazy or not). | Python | bsd-3-clause | stefanklug/django-lazysignup,rwillmer/django-lazysignup,rwillmer/django-lazysignup,danfairs/django-lazysignup,stefanklug/django-lazysignup,danfairs/django-lazysignup | from django.contrib.auth.backends import ModelBackend
- from django.contrib.auth.models import User
+ from lazysignup.models import LazyUser
class LazySignupBackend(ModelBackend):
def authenticate(self, username=None):
- users = [u for u in User.objects.filter(username=username)
- if not u.has_usable_password()]
- if len(users) != 1:
+ lazy_users = LazyUser.objects.filter(
+ user__username=username
+ ).select_related('user')
+ try:
+ return lazy_users[0].user
+ except IndexError:
return None
- return users[0]
def get_user(self, user_id):
# Annotate the user with our backend so it's always available,
# not just when authenticate() has been called. This will be
# used by the is_lazy_user filter.
user = super(LazySignupBackend, self).get_user(user_id)
if user:
user.backend = 'lazysignup.backends.LazySignupBackend'
return user
+ | Remove the lazy signup backend's hard dependency on django.contrib.auth.user (and remove the inconsistency in checking for whether a user is lazy or not). | ## Code Before:
from django.contrib.auth.backends import ModelBackend
from django.contrib.auth.models import User
class LazySignupBackend(ModelBackend):
def authenticate(self, username=None):
users = [u for u in User.objects.filter(username=username)
if not u.has_usable_password()]
if len(users) != 1:
return None
return users[0]
def get_user(self, user_id):
# Annotate the user with our backend so it's always available,
# not just when authenticate() has been called. This will be
# used by the is_lazy_user filter.
user = super(LazySignupBackend, self).get_user(user_id)
if user:
user.backend = 'lazysignup.backends.LazySignupBackend'
return user
## Instruction:
Remove the lazy signup backend's hard dependency on django.contrib.auth.user (and remove the inconsistency in checking for whether a user is lazy or not).
## Code After:
from django.contrib.auth.backends import ModelBackend
from lazysignup.models import LazyUser
class LazySignupBackend(ModelBackend):
def authenticate(self, username=None):
lazy_users = LazyUser.objects.filter(
user__username=username
).select_related('user')
try:
return lazy_users[0].user
except IndexError:
return None
def get_user(self, user_id):
# Annotate the user with our backend so it's always available,
# not just when authenticate() has been called. This will be
# used by the is_lazy_user filter.
user = super(LazySignupBackend, self).get_user(user_id)
if user:
user.backend = 'lazysignup.backends.LazySignupBackend'
return user
| # ... existing code ...
from django.contrib.auth.backends import ModelBackend
from lazysignup.models import LazyUser
class LazySignupBackend(ModelBackend):
# ... modified code ...
def authenticate(self, username=None):
lazy_users = LazyUser.objects.filter(
user__username=username
).select_related('user')
try:
return lazy_users[0].user
except IndexError:
return None
def get_user(self, user_id):
...
user.backend = 'lazysignup.backends.LazySignupBackend'
return user
# ... rest of the code ... |
a17933c7806634391137244e2c17327898187146 | djstripe/__init__.py | djstripe/__init__.py | from __future__ import absolute_import, division, print_function, unicode_literals
import pkg_resources
from . import checks # noqa: Register the checks
__version__ = pkg_resources.require("dj-stripe")[0].version
| from __future__ import absolute_import, division, print_function, unicode_literals
import pkg_resources
import stripe
from . import checks # noqa: Register the checks
__version__ = pkg_resources.require("dj-stripe")[0].version
# Set app info
# https://stripe.com/docs/building-plugins#setappinfo
stripe.set_app_info(
"dj-stripe",
version=__version__,
url="https://github.com/dj-stripe/dj-stripe"
)
| Set dj-stripe as stripe app info | Set dj-stripe as stripe app info
https://stripe.com/docs/building-plugins#setappinfo
| Python | mit | pydanny/dj-stripe,pydanny/dj-stripe,jleclanche/dj-stripe,dj-stripe/dj-stripe,kavdev/dj-stripe,jleclanche/dj-stripe,dj-stripe/dj-stripe,kavdev/dj-stripe | from __future__ import absolute_import, division, print_function, unicode_literals
import pkg_resources
+
+ import stripe
+
from . import checks # noqa: Register the checks
__version__ = pkg_resources.require("dj-stripe")[0].version
+
+ # Set app info
+ # https://stripe.com/docs/building-plugins#setappinfo
+ stripe.set_app_info(
+ "dj-stripe",
+ version=__version__,
+ url="https://github.com/dj-stripe/dj-stripe"
+ )
+ | Set dj-stripe as stripe app info | ## Code Before:
from __future__ import absolute_import, division, print_function, unicode_literals
import pkg_resources
from . import checks # noqa: Register the checks
__version__ = pkg_resources.require("dj-stripe")[0].version
## Instruction:
Set dj-stripe as stripe app info
## Code After:
from __future__ import absolute_import, division, print_function, unicode_literals
import pkg_resources
import stripe
from . import checks # noqa: Register the checks
__version__ = pkg_resources.require("dj-stripe")[0].version
# Set app info
# https://stripe.com/docs/building-plugins#setappinfo
stripe.set_app_info(
"dj-stripe",
version=__version__,
url="https://github.com/dj-stripe/dj-stripe"
)
| # ... existing code ...
from __future__ import absolute_import, division, print_function, unicode_literals
import pkg_resources
import stripe
from . import checks # noqa: Register the checks
# ... modified code ...
__version__ = pkg_resources.require("dj-stripe")[0].version
# Set app info
# https://stripe.com/docs/building-plugins#setappinfo
stripe.set_app_info(
"dj-stripe",
version=__version__,
url="https://github.com/dj-stripe/dj-stripe"
)
# ... rest of the code ... |
2dfe6e78088f974310c1e7fc309f008310be0080 | dask_ndmeasure/_utils.py | dask_ndmeasure/_utils.py |
import operator
import numpy
import dask.array
from . import _compat
def _norm_input_labels_index(input, labels=None, index=None):
"""
Normalize arguments to a standard form.
"""
input = _compat._asarray(input)
if labels is None:
labels = (input != 0).astype(numpy.int64)
index = None
if index is None:
labels = (labels > 0).astype(numpy.int64)
index = dask.array.ones(tuple(), dtype=numpy.int64, chunks=tuple())
labels = _compat._asarray(labels)
index = _compat._asarray(index)
# SciPy transposes these for some reason.
# So we do the same thing here.
# This only matters if index is some array.
index = index.T
if input.shape != labels.shape:
raise ValueError("The input and labels arrays must be the same shape.")
return (input, labels, index)
def _get_label_matches(input, labels, index):
input_i = _compat._indices(
input.shape, dtype=numpy.int64, chunks=input.chunks
)
lbl_mtch = operator.eq(
index[(Ellipsis,) + labels.ndim * (None,)],
labels[index.ndim * (None,)]
)
input_i_mtch = dask.array.where(
lbl_mtch[index.ndim * (slice(None),) + (None,)],
input_i[index.ndim * (None,)],
input_i.dtype.type(0)
)
input_mtch = dask.array.where(
lbl_mtch, input[index.ndim * (None,)], input.dtype.type(0)
)
return (lbl_mtch, input_i_mtch, input_mtch)
|
import operator
import numpy
import dask.array
from . import _compat
def _norm_input_labels_index(input, labels=None, index=None):
"""
Normalize arguments to a standard form.
"""
input = _compat._asarray(input)
if labels is None:
labels = (input != 0).astype(numpy.int64)
index = None
if index is None:
labels = (labels > 0).astype(numpy.int64)
index = dask.array.ones(tuple(), dtype=numpy.int64, chunks=tuple())
labels = _compat._asarray(labels)
index = _compat._asarray(index)
# SciPy transposes these for some reason.
# So we do the same thing here.
# This only matters if index is some array.
index = index.T
if input.shape != labels.shape:
raise ValueError("The input and labels arrays must be the same shape.")
return (input, labels, index)
def _get_label_matches(labels, index):
lbl_mtch = operator.eq(
index[(Ellipsis,) + labels.ndim * (None,)],
labels[index.ndim * (None,)]
)
return lbl_mtch
| Simplify the label matches function | Simplify the label matches function
Focus only creating a mask of selected labels and none of the other
products that this function was creating before.
| Python | bsd-3-clause | dask-image/dask-ndmeasure |
import operator
import numpy
import dask.array
from . import _compat
def _norm_input_labels_index(input, labels=None, index=None):
"""
Normalize arguments to a standard form.
"""
input = _compat._asarray(input)
if labels is None:
labels = (input != 0).astype(numpy.int64)
index = None
if index is None:
labels = (labels > 0).astype(numpy.int64)
index = dask.array.ones(tuple(), dtype=numpy.int64, chunks=tuple())
labels = _compat._asarray(labels)
index = _compat._asarray(index)
# SciPy transposes these for some reason.
# So we do the same thing here.
# This only matters if index is some array.
index = index.T
if input.shape != labels.shape:
raise ValueError("The input and labels arrays must be the same shape.")
return (input, labels, index)
- def _get_label_matches(input, labels, index):
+ def _get_label_matches(labels, index):
- input_i = _compat._indices(
- input.shape, dtype=numpy.int64, chunks=input.chunks
- )
-
lbl_mtch = operator.eq(
index[(Ellipsis,) + labels.ndim * (None,)],
labels[index.ndim * (None,)]
)
+ return lbl_mtch
- input_i_mtch = dask.array.where(
- lbl_mtch[index.ndim * (slice(None),) + (None,)],
- input_i[index.ndim * (None,)],
- input_i.dtype.type(0)
- )
- input_mtch = dask.array.where(
- lbl_mtch, input[index.ndim * (None,)], input.dtype.type(0)
- )
-
- return (lbl_mtch, input_i_mtch, input_mtch)
- | Simplify the label matches function | ## Code Before:
import operator
import numpy
import dask.array
from . import _compat
def _norm_input_labels_index(input, labels=None, index=None):
"""
Normalize arguments to a standard form.
"""
input = _compat._asarray(input)
if labels is None:
labels = (input != 0).astype(numpy.int64)
index = None
if index is None:
labels = (labels > 0).astype(numpy.int64)
index = dask.array.ones(tuple(), dtype=numpy.int64, chunks=tuple())
labels = _compat._asarray(labels)
index = _compat._asarray(index)
# SciPy transposes these for some reason.
# So we do the same thing here.
# This only matters if index is some array.
index = index.T
if input.shape != labels.shape:
raise ValueError("The input and labels arrays must be the same shape.")
return (input, labels, index)
def _get_label_matches(input, labels, index):
input_i = _compat._indices(
input.shape, dtype=numpy.int64, chunks=input.chunks
)
lbl_mtch = operator.eq(
index[(Ellipsis,) + labels.ndim * (None,)],
labels[index.ndim * (None,)]
)
input_i_mtch = dask.array.where(
lbl_mtch[index.ndim * (slice(None),) + (None,)],
input_i[index.ndim * (None,)],
input_i.dtype.type(0)
)
input_mtch = dask.array.where(
lbl_mtch, input[index.ndim * (None,)], input.dtype.type(0)
)
return (lbl_mtch, input_i_mtch, input_mtch)
## Instruction:
Simplify the label matches function
## Code After:
import operator
import numpy
import dask.array
from . import _compat
def _norm_input_labels_index(input, labels=None, index=None):
"""
Normalize arguments to a standard form.
"""
input = _compat._asarray(input)
if labels is None:
labels = (input != 0).astype(numpy.int64)
index = None
if index is None:
labels = (labels > 0).astype(numpy.int64)
index = dask.array.ones(tuple(), dtype=numpy.int64, chunks=tuple())
labels = _compat._asarray(labels)
index = _compat._asarray(index)
# SciPy transposes these for some reason.
# So we do the same thing here.
# This only matters if index is some array.
index = index.T
if input.shape != labels.shape:
raise ValueError("The input and labels arrays must be the same shape.")
return (input, labels, index)
def _get_label_matches(labels, index):
lbl_mtch = operator.eq(
index[(Ellipsis,) + labels.ndim * (None,)],
labels[index.ndim * (None,)]
)
return lbl_mtch
| # ... existing code ...
def _get_label_matches(labels, index):
lbl_mtch = operator.eq(
index[(Ellipsis,) + labels.ndim * (None,)],
# ... modified code ...
)
return lbl_mtch
# ... rest of the code ... |
56aa0448fb3cd1df1a0fd43abc9a0e37e8ddf55b | trans_sync/management/commands/save_trans.py | trans_sync/management/commands/save_trans.py | from __future__ import unicode_literals
from optparse import make_option
from django.core.management.base import NoArgsCommand
class Command(NoArgsCommand):
option_list = NoArgsCommand.option_list + (
make_option(
'--dry-run',
action='store_true',
dest='dry_run',
default=False,
help='Do not actually send signals (and all connected stuff).'
),
)
def handle(self, *args, **options):
if not options['dry_run']:
pass | from __future__ import unicode_literals
import os
from os.path import join, isdir
from optparse import make_option
from django.core.management.base import NoArgsCommand
from django.conf import settings
from modeltranslation.translator import translator
from babel.messages.catalog import Catalog
from babel.messages.pofile import write_po
class Command(NoArgsCommand):
option_list = NoArgsCommand.option_list + (
make_option(
'--dry-run',
action='store_true',
dest='dry_run',
default=False,
help='Do not actually save files.'
),
)
def handle(self, *args, **options):
if not options['dry_run']:
pass
locale_path = settings.LOCALE_MODEL_TRANS
if not isdir(locale_path):
os.mkdir(locale_path)
for lang in [l[0] for l in list(settings.LANGUAGES)]:
catalog = Catalog(locale=lang)
for model in translator.get_registered_models():
opts = translator.get_options_for_model(model)
for field in opts.get_field_names():
tr_field = "%s_%s" % (field, lang)
for item in model.objects.all():
msgid = "%s.%s.%s" % (item._meta, item.pk, field)
msgstr = "%s" % getattr(item, tr_field)
catalog.add(id=msgid, string=msgstr)
# write catalog to file
lang_path = os.path.join(locale_path, lang)
if not isdir(lang_path):
os.mkdir(lang_path)
f = open(join(lang_path, "LC_MESSAGES", "modeltranslation.po"), "w")
write_po(f, catalog)
f.close() | Save trans to .po files | Save trans to .po files
| Python | mit | djentlemen/django-modeltranslation-sync | from __future__ import unicode_literals
-
+ import os
+ from os.path import join, isdir
from optparse import make_option
from django.core.management.base import NoArgsCommand
+ from django.conf import settings
+ from modeltranslation.translator import translator
+
+ from babel.messages.catalog import Catalog
+ from babel.messages.pofile import write_po
class Command(NoArgsCommand):
option_list = NoArgsCommand.option_list + (
make_option(
'--dry-run',
action='store_true',
dest='dry_run',
default=False,
- help='Do not actually send signals (and all connected stuff).'
+ help='Do not actually save files.'
),
)
def handle(self, *args, **options):
if not options['dry_run']:
pass
+
+ locale_path = settings.LOCALE_MODEL_TRANS
+ if not isdir(locale_path):
+ os.mkdir(locale_path)
+
+ for lang in [l[0] for l in list(settings.LANGUAGES)]:
+
+ catalog = Catalog(locale=lang)
+
+ for model in translator.get_registered_models():
+ opts = translator.get_options_for_model(model)
+
+ for field in opts.get_field_names():
+ tr_field = "%s_%s" % (field, lang)
+ for item in model.objects.all():
+ msgid = "%s.%s.%s" % (item._meta, item.pk, field)
+ msgstr = "%s" % getattr(item, tr_field)
+ catalog.add(id=msgid, string=msgstr)
+
+ # write catalog to file
+ lang_path = os.path.join(locale_path, lang)
+ if not isdir(lang_path):
+ os.mkdir(lang_path)
+ f = open(join(lang_path, "LC_MESSAGES", "modeltranslation.po"), "w")
+ write_po(f, catalog)
+ f.close() | Save trans to .po files | ## Code Before:
from __future__ import unicode_literals
from optparse import make_option
from django.core.management.base import NoArgsCommand
class Command(NoArgsCommand):
option_list = NoArgsCommand.option_list + (
make_option(
'--dry-run',
action='store_true',
dest='dry_run',
default=False,
help='Do not actually send signals (and all connected stuff).'
),
)
def handle(self, *args, **options):
if not options['dry_run']:
pass
## Instruction:
Save trans to .po files
## Code After:
from __future__ import unicode_literals
import os
from os.path import join, isdir
from optparse import make_option
from django.core.management.base import NoArgsCommand
from django.conf import settings
from modeltranslation.translator import translator
from babel.messages.catalog import Catalog
from babel.messages.pofile import write_po
class Command(NoArgsCommand):
option_list = NoArgsCommand.option_list + (
make_option(
'--dry-run',
action='store_true',
dest='dry_run',
default=False,
help='Do not actually save files.'
),
)
def handle(self, *args, **options):
if not options['dry_run']:
pass
locale_path = settings.LOCALE_MODEL_TRANS
if not isdir(locale_path):
os.mkdir(locale_path)
for lang in [l[0] for l in list(settings.LANGUAGES)]:
catalog = Catalog(locale=lang)
for model in translator.get_registered_models():
opts = translator.get_options_for_model(model)
for field in opts.get_field_names():
tr_field = "%s_%s" % (field, lang)
for item in model.objects.all():
msgid = "%s.%s.%s" % (item._meta, item.pk, field)
msgstr = "%s" % getattr(item, tr_field)
catalog.add(id=msgid, string=msgstr)
# write catalog to file
lang_path = os.path.join(locale_path, lang)
if not isdir(lang_path):
os.mkdir(lang_path)
f = open(join(lang_path, "LC_MESSAGES", "modeltranslation.po"), "w")
write_po(f, catalog)
f.close() | # ... existing code ...
from __future__ import unicode_literals
import os
from os.path import join, isdir
from optparse import make_option
from django.core.management.base import NoArgsCommand
from django.conf import settings
from modeltranslation.translator import translator
from babel.messages.catalog import Catalog
from babel.messages.pofile import write_po
# ... modified code ...
dest='dry_run',
default=False,
help='Do not actually save files.'
),
)
...
if not options['dry_run']:
pass
locale_path = settings.LOCALE_MODEL_TRANS
if not isdir(locale_path):
os.mkdir(locale_path)
for lang in [l[0] for l in list(settings.LANGUAGES)]:
catalog = Catalog(locale=lang)
for model in translator.get_registered_models():
opts = translator.get_options_for_model(model)
for field in opts.get_field_names():
tr_field = "%s_%s" % (field, lang)
for item in model.objects.all():
msgid = "%s.%s.%s" % (item._meta, item.pk, field)
msgstr = "%s" % getattr(item, tr_field)
catalog.add(id=msgid, string=msgstr)
# write catalog to file
lang_path = os.path.join(locale_path, lang)
if not isdir(lang_path):
os.mkdir(lang_path)
f = open(join(lang_path, "LC_MESSAGES", "modeltranslation.po"), "w")
write_po(f, catalog)
f.close()
# ... rest of the code ... |
35b45fd793ac695f6ec6a792534fdde77a3023aa | napalm_yang/supported_models.py | napalm_yang/supported_models.py | SUPPORTED_MODELS = (
# module_name, models
("openconfig-interfaces", ["interfaces"]),
("openconfig-network-instance", ["network_instances"]),
("openconfig-platform", ["components"]),
("openconfig-vlan", ["vlans"]),
)
| SUPPORTED_MODELS = (
# module_name, models
("openconfig-interfaces", ["interfaces"]),
("openconfig-network-instance", ["network_instances"]),
("openconfig-platform", ["components"]),
("openconfig-vlan", ["vlans"]),
('openconfig-system', ['system'])
)
| Add system as supported models | Add system as supported models
| Python | apache-2.0 | napalm-automation/napalm-yang,napalm-automation/napalm-yang | SUPPORTED_MODELS = (
# module_name, models
("openconfig-interfaces", ["interfaces"]),
("openconfig-network-instance", ["network_instances"]),
("openconfig-platform", ["components"]),
("openconfig-vlan", ["vlans"]),
+ ('openconfig-system', ['system'])
)
| Add system as supported models | ## Code Before:
SUPPORTED_MODELS = (
# module_name, models
("openconfig-interfaces", ["interfaces"]),
("openconfig-network-instance", ["network_instances"]),
("openconfig-platform", ["components"]),
("openconfig-vlan", ["vlans"]),
)
## Instruction:
Add system as supported models
## Code After:
SUPPORTED_MODELS = (
# module_name, models
("openconfig-interfaces", ["interfaces"]),
("openconfig-network-instance", ["network_instances"]),
("openconfig-platform", ["components"]),
("openconfig-vlan", ["vlans"]),
('openconfig-system', ['system'])
)
| ...
("openconfig-platform", ["components"]),
("openconfig-vlan", ["vlans"]),
('openconfig-system', ['system'])
)
... |
53b9eff3ffc1768d3503021e7248351e24d59af7 | tests/httpd.py | tests/httpd.py | import SimpleHTTPServer
import BaseHTTPServer
class Handler(SimpleHTTPServer.SimpleHTTPRequestHandler):
def do_POST(s):
s.send_response(200)
s.end_headers()
if __name__ == '__main__':
server_class = BaseHTTPServer.HTTPServer
httpd = server_class(('0.0.0.0', 8328), Handler)
try:
httpd.serve_forever()
except KeyboardInterrupt:
httpd.server_close()
| import BaseHTTPServer
class Handler(BaseHTTPServer.BaseHTTPRequestHandler):
def do_POST(self):
content_type = self.headers.getheader('content-type')
content_length = int(self.headers.getheader('content-length'))
self.send_response(200)
self.send_header('Content-Type', content_type)
self.send_header('Content-Length', str(content_length))
self.end_headers()
self.wfile.write(self.rfile.read(content_length))
if __name__ == '__main__':
server_class = BaseHTTPServer.HTTPServer
httpd = server_class(('0.0.0.0', 8328), Handler)
try:
httpd.serve_forever()
except KeyboardInterrupt:
httpd.server_close()
| Fix test http server, change to echo back request body | Fix test http server, change to echo back request body | Python | bsd-2-clause | chop-dbhi/django-webhooks,pombredanne/django-webhooks,pombredanne/django-webhooks,chop-dbhi/django-webhooks | - import SimpleHTTPServer
import BaseHTTPServer
- class Handler(SimpleHTTPServer.SimpleHTTPRequestHandler):
+ class Handler(BaseHTTPServer.BaseHTTPRequestHandler):
- def do_POST(s):
+ def do_POST(self):
+ content_type = self.headers.getheader('content-type')
+ content_length = int(self.headers.getheader('content-length'))
- s.send_response(200)
+ self.send_response(200)
+ self.send_header('Content-Type', content_type)
+ self.send_header('Content-Length', str(content_length))
- s.end_headers()
+ self.end_headers()
+ self.wfile.write(self.rfile.read(content_length))
if __name__ == '__main__':
server_class = BaseHTTPServer.HTTPServer
httpd = server_class(('0.0.0.0', 8328), Handler)
try:
httpd.serve_forever()
except KeyboardInterrupt:
httpd.server_close()
| Fix test http server, change to echo back request body | ## Code Before:
import SimpleHTTPServer
import BaseHTTPServer
class Handler(SimpleHTTPServer.SimpleHTTPRequestHandler):
def do_POST(s):
s.send_response(200)
s.end_headers()
if __name__ == '__main__':
server_class = BaseHTTPServer.HTTPServer
httpd = server_class(('0.0.0.0', 8328), Handler)
try:
httpd.serve_forever()
except KeyboardInterrupt:
httpd.server_close()
## Instruction:
Fix test http server, change to echo back request body
## Code After:
import BaseHTTPServer
class Handler(BaseHTTPServer.BaseHTTPRequestHandler):
def do_POST(self):
content_type = self.headers.getheader('content-type')
content_length = int(self.headers.getheader('content-length'))
self.send_response(200)
self.send_header('Content-Type', content_type)
self.send_header('Content-Length', str(content_length))
self.end_headers()
self.wfile.write(self.rfile.read(content_length))
if __name__ == '__main__':
server_class = BaseHTTPServer.HTTPServer
httpd = server_class(('0.0.0.0', 8328), Handler)
try:
httpd.serve_forever()
except KeyboardInterrupt:
httpd.server_close()
| # ... existing code ...
import BaseHTTPServer
class Handler(BaseHTTPServer.BaseHTTPRequestHandler):
def do_POST(self):
content_type = self.headers.getheader('content-type')
content_length = int(self.headers.getheader('content-length'))
self.send_response(200)
self.send_header('Content-Type', content_type)
self.send_header('Content-Length', str(content_length))
self.end_headers()
self.wfile.write(self.rfile.read(content_length))
if __name__ == '__main__':
# ... rest of the code ... |
b1feed0ced6d1328cc39bc9bba36331ec6da7803 | pre_commit_hooks/detect_private_key.py | pre_commit_hooks/detect_private_key.py | from __future__ import print_function
import argparse
import sys
BLACKLIST = [
b'BEGIN RSA PRIVATE KEY',
b'BEGIN DSA PRIVATE KEY',
b'BEGIN EC PRIVATE KEY',
b'BEGIN OPENSSH PRIVATE KEY',
b'BEGIN PRIVATE KEY',
b'PuTTY-User-Key-File-2',
b'BEGIN SSH2 ENCRYPTED PRIVATE KEY',
]
def detect_private_key(argv=None):
parser = argparse.ArgumentParser()
parser.add_argument('filenames', nargs='*', help='Filenames to check')
args = parser.parse_args(argv)
private_key_files = []
for filename in args.filenames:
with open(filename, 'rb') as f:
content = f.read()
if any(line in content for line in BLACKLIST):
private_key_files.append(filename)
if private_key_files:
for private_key_file in private_key_files:
print('Private key found: {}'.format(private_key_file))
return 1
else:
return 0
if __name__ == '__main__':
sys.exit(detect_private_key())
| from __future__ import print_function
import argparse
import sys
BLACKLIST = [
b'BEGIN RSA PRIVATE KEY',
b'BEGIN DSA PRIVATE KEY',
b'BEGIN EC PRIVATE KEY',
b'BEGIN OPENSSH PRIVATE KEY',
b'BEGIN PRIVATE KEY',
b'PuTTY-User-Key-File-2',
b'BEGIN SSH2 ENCRYPTED PRIVATE KEY',
b'BEGIN PGP PRIVATE KEY BLOCK',
]
def detect_private_key(argv=None):
parser = argparse.ArgumentParser()
parser.add_argument('filenames', nargs='*', help='Filenames to check')
args = parser.parse_args(argv)
private_key_files = []
for filename in args.filenames:
with open(filename, 'rb') as f:
content = f.read()
if any(line in content for line in BLACKLIST):
private_key_files.append(filename)
if private_key_files:
for private_key_file in private_key_files:
print('Private key found: {}'.format(private_key_file))
return 1
else:
return 0
if __name__ == '__main__':
sys.exit(detect_private_key())
| Add ban for pgp/gpg private key blocks | Add ban for pgp/gpg private key blocks
| Python | mit | pre-commit/pre-commit-hooks,Harwood/pre-commit-hooks | from __future__ import print_function
import argparse
import sys
BLACKLIST = [
b'BEGIN RSA PRIVATE KEY',
b'BEGIN DSA PRIVATE KEY',
b'BEGIN EC PRIVATE KEY',
b'BEGIN OPENSSH PRIVATE KEY',
b'BEGIN PRIVATE KEY',
b'PuTTY-User-Key-File-2',
b'BEGIN SSH2 ENCRYPTED PRIVATE KEY',
+ b'BEGIN PGP PRIVATE KEY BLOCK',
]
def detect_private_key(argv=None):
parser = argparse.ArgumentParser()
parser.add_argument('filenames', nargs='*', help='Filenames to check')
args = parser.parse_args(argv)
private_key_files = []
for filename in args.filenames:
with open(filename, 'rb') as f:
content = f.read()
if any(line in content for line in BLACKLIST):
private_key_files.append(filename)
if private_key_files:
for private_key_file in private_key_files:
print('Private key found: {}'.format(private_key_file))
return 1
else:
return 0
if __name__ == '__main__':
sys.exit(detect_private_key())
| Add ban for pgp/gpg private key blocks | ## Code Before:
from __future__ import print_function
import argparse
import sys
BLACKLIST = [
b'BEGIN RSA PRIVATE KEY',
b'BEGIN DSA PRIVATE KEY',
b'BEGIN EC PRIVATE KEY',
b'BEGIN OPENSSH PRIVATE KEY',
b'BEGIN PRIVATE KEY',
b'PuTTY-User-Key-File-2',
b'BEGIN SSH2 ENCRYPTED PRIVATE KEY',
]
def detect_private_key(argv=None):
parser = argparse.ArgumentParser()
parser.add_argument('filenames', nargs='*', help='Filenames to check')
args = parser.parse_args(argv)
private_key_files = []
for filename in args.filenames:
with open(filename, 'rb') as f:
content = f.read()
if any(line in content for line in BLACKLIST):
private_key_files.append(filename)
if private_key_files:
for private_key_file in private_key_files:
print('Private key found: {}'.format(private_key_file))
return 1
else:
return 0
if __name__ == '__main__':
sys.exit(detect_private_key())
## Instruction:
Add ban for pgp/gpg private key blocks
## Code After:
from __future__ import print_function
import argparse
import sys
BLACKLIST = [
b'BEGIN RSA PRIVATE KEY',
b'BEGIN DSA PRIVATE KEY',
b'BEGIN EC PRIVATE KEY',
b'BEGIN OPENSSH PRIVATE KEY',
b'BEGIN PRIVATE KEY',
b'PuTTY-User-Key-File-2',
b'BEGIN SSH2 ENCRYPTED PRIVATE KEY',
b'BEGIN PGP PRIVATE KEY BLOCK',
]
def detect_private_key(argv=None):
parser = argparse.ArgumentParser()
parser.add_argument('filenames', nargs='*', help='Filenames to check')
args = parser.parse_args(argv)
private_key_files = []
for filename in args.filenames:
with open(filename, 'rb') as f:
content = f.read()
if any(line in content for line in BLACKLIST):
private_key_files.append(filename)
if private_key_files:
for private_key_file in private_key_files:
print('Private key found: {}'.format(private_key_file))
return 1
else:
return 0
if __name__ == '__main__':
sys.exit(detect_private_key())
| ...
b'PuTTY-User-Key-File-2',
b'BEGIN SSH2 ENCRYPTED PRIVATE KEY',
b'BEGIN PGP PRIVATE KEY BLOCK',
]
... |
1914bd9a9fc2f1ce7a557dd2134d98796866283d | rootpy/userdata.py | rootpy/userdata.py | import os
import tempfile
import atexit
DATA_ROOT = None
if os.getenv('ROOTPY_GRIDMODE') not in ('1', 'true'):
DATA_ROOT = os.getenv('ROOTPY_DATA')
if DATA_ROOT is None:
DATA_ROOT = os.path.expanduser('~/.rootpy')
else:
DATA_ROOT = os.path.expandvars(os.path.expanduser(DATA_ROOT))
# check if expanduser failed:
if DATA_ROOT.startswith('~'):
DATA_ROOT = None
elif not os.path.exists(DATA_ROOT):
os.mkdir(DATA_ROOT)
elif not os.path.isdir(DATA_ROOT):
# A file at DATA_ROOT already exists
DATA_ROOT = None
__is_tmp = False
if DATA_ROOT is None:
print "Warning: placing user data in /tmp.\n" \
"Make sure ~/.rootpy or $ROOTPY_DATA\n" \
"is a writable directory so that I don't need to\n" \
"recreate all user data each time"
DATA_ROOT = tempfile.mkdtemp()
__is_tmp = True
@atexit.register
def __cleanup():
if __is_tmp:
import shutil
shutil.rmtree(DATA_ROOT)
| import os
import tempfile
import atexit
from os.path import expanduser, expandvars, exists, isdir
if "XDG_CONFIG_HOME" not in os.environ:
os.environ["XDG_CONFIG_HOME"] = expanduser('~/.config')
if "XDG_CACHE_HOME" not in os.environ:
os.environ["XDG_CACHE_HOME"] = expanduser('~/.cache')
def ensure_directory(variable, default):
path = os.getenv(variable)
if path is None:
path = expandvars(default)
else:
path = expandvars(expanduser(path))
# check if expanduser failed:
if path.startswith('~'):
path = None
elif not exists(path):
os.makedirs(path)
elif not isdir(path):
# A file at path already exists
path = None
return path
DATA_ROOT = CONFIG_ROOT = None
if os.getenv('ROOTPY_GRIDMODE') not in ('1', 'true'):
DATA_ROOT = ensure_directory('ROOTPY_DATA', '${XDG_CACHE_HOME}/rootpy')
CONFIG_ROOT = ensure_directory('ROOTPY_CONFIG', '${XDG_CONFIG_HOME}/rootpy')
if DATA_ROOT is None:
log.info("Placing user data in /tmp.")
log.warning("Make sure '~/.cache/rootpy' or $ROOTPY_DATA is a writable "
"directory so that it isn't necessary to recreate all user data"
" each time")
DATA_ROOT = tempfile.mkdtemp()
@atexit.register
def __cleanup():
import shutil
shutil.rmtree(DATA_ROOT)
| Use XDG base directory specificiation | Use XDG base directory specificiation
| Python | bsd-3-clause | rootpy/rootpy,rootpy/rootpy,ndawe/rootpy,kreczko/rootpy,ndawe/rootpy,kreczko/rootpy,ndawe/rootpy,kreczko/rootpy,rootpy/rootpy | import os
import tempfile
import atexit
+ from os.path import expanduser, expandvars, exists, isdir
-
- DATA_ROOT = None
- if os.getenv('ROOTPY_GRIDMODE') not in ('1', 'true'):
- DATA_ROOT = os.getenv('ROOTPY_DATA')
- if DATA_ROOT is None:
- DATA_ROOT = os.path.expanduser('~/.rootpy')
- else:
- DATA_ROOT = os.path.expandvars(os.path.expanduser(DATA_ROOT))
- # check if expanduser failed:
- if DATA_ROOT.startswith('~'):
- DATA_ROOT = None
- elif not os.path.exists(DATA_ROOT):
- os.mkdir(DATA_ROOT)
- elif not os.path.isdir(DATA_ROOT):
- # A file at DATA_ROOT already exists
- DATA_ROOT = None
-
- __is_tmp = False
- if DATA_ROOT is None:
- print "Warning: placing user data in /tmp.\n" \
- "Make sure ~/.rootpy or $ROOTPY_DATA\n" \
- "is a writable directory so that I don't need to\n" \
- "recreate all user data each time"
- DATA_ROOT = tempfile.mkdtemp()
- __is_tmp = True
+ if "XDG_CONFIG_HOME" not in os.environ:
+ os.environ["XDG_CONFIG_HOME"] = expanduser('~/.config')
+ if "XDG_CACHE_HOME" not in os.environ:
+ os.environ["XDG_CACHE_HOME"] = expanduser('~/.cache')
+
+ def ensure_directory(variable, default):
+ path = os.getenv(variable)
+ if path is None:
+ path = expandvars(default)
+ else:
+ path = expandvars(expanduser(path))
+
+ # check if expanduser failed:
+ if path.startswith('~'):
+ path = None
+ elif not exists(path):
+ os.makedirs(path)
+ elif not isdir(path):
+ # A file at path already exists
+ path = None
+ return path
+
+ DATA_ROOT = CONFIG_ROOT = None
+ if os.getenv('ROOTPY_GRIDMODE') not in ('1', 'true'):
+ DATA_ROOT = ensure_directory('ROOTPY_DATA', '${XDG_CACHE_HOME}/rootpy')
+ CONFIG_ROOT = ensure_directory('ROOTPY_CONFIG', '${XDG_CONFIG_HOME}/rootpy')
+
+
+ if DATA_ROOT is None:
+ log.info("Placing user data in /tmp.")
+ log.warning("Make sure '~/.cache/rootpy' or $ROOTPY_DATA is a writable "
+ "directory so that it isn't necessary to recreate all user data"
+ " each time")
+
+ DATA_ROOT = tempfile.mkdtemp()
+
- @atexit.register
+ @atexit.register
- def __cleanup():
+ def __cleanup():
- if __is_tmp:
import shutil
shutil.rmtree(DATA_ROOT)
+ | Use XDG base directory specificiation | ## Code Before:
import os
import tempfile
import atexit
DATA_ROOT = None
if os.getenv('ROOTPY_GRIDMODE') not in ('1', 'true'):
DATA_ROOT = os.getenv('ROOTPY_DATA')
if DATA_ROOT is None:
DATA_ROOT = os.path.expanduser('~/.rootpy')
else:
DATA_ROOT = os.path.expandvars(os.path.expanduser(DATA_ROOT))
# check if expanduser failed:
if DATA_ROOT.startswith('~'):
DATA_ROOT = None
elif not os.path.exists(DATA_ROOT):
os.mkdir(DATA_ROOT)
elif not os.path.isdir(DATA_ROOT):
# A file at DATA_ROOT already exists
DATA_ROOT = None
__is_tmp = False
if DATA_ROOT is None:
print "Warning: placing user data in /tmp.\n" \
"Make sure ~/.rootpy or $ROOTPY_DATA\n" \
"is a writable directory so that I don't need to\n" \
"recreate all user data each time"
DATA_ROOT = tempfile.mkdtemp()
__is_tmp = True
@atexit.register
def __cleanup():
if __is_tmp:
import shutil
shutil.rmtree(DATA_ROOT)
## Instruction:
Use XDG base directory specificiation
## Code After:
import os
import tempfile
import atexit
from os.path import expanduser, expandvars, exists, isdir
if "XDG_CONFIG_HOME" not in os.environ:
os.environ["XDG_CONFIG_HOME"] = expanduser('~/.config')
if "XDG_CACHE_HOME" not in os.environ:
os.environ["XDG_CACHE_HOME"] = expanduser('~/.cache')
def ensure_directory(variable, default):
path = os.getenv(variable)
if path is None:
path = expandvars(default)
else:
path = expandvars(expanduser(path))
# check if expanduser failed:
if path.startswith('~'):
path = None
elif not exists(path):
os.makedirs(path)
elif not isdir(path):
# A file at path already exists
path = None
return path
DATA_ROOT = CONFIG_ROOT = None
if os.getenv('ROOTPY_GRIDMODE') not in ('1', 'true'):
DATA_ROOT = ensure_directory('ROOTPY_DATA', '${XDG_CACHE_HOME}/rootpy')
CONFIG_ROOT = ensure_directory('ROOTPY_CONFIG', '${XDG_CONFIG_HOME}/rootpy')
if DATA_ROOT is None:
log.info("Placing user data in /tmp.")
log.warning("Make sure '~/.cache/rootpy' or $ROOTPY_DATA is a writable "
"directory so that it isn't necessary to recreate all user data"
" each time")
DATA_ROOT = tempfile.mkdtemp()
@atexit.register
def __cleanup():
import shutil
shutil.rmtree(DATA_ROOT)
| # ... existing code ...
import atexit
from os.path import expanduser, expandvars, exists, isdir
if "XDG_CONFIG_HOME" not in os.environ:
os.environ["XDG_CONFIG_HOME"] = expanduser('~/.config')
if "XDG_CACHE_HOME" not in os.environ:
os.environ["XDG_CACHE_HOME"] = expanduser('~/.cache')
def ensure_directory(variable, default):
path = os.getenv(variable)
if path is None:
path = expandvars(default)
else:
path = expandvars(expanduser(path))
# check if expanduser failed:
if path.startswith('~'):
path = None
elif not exists(path):
os.makedirs(path)
elif not isdir(path):
# A file at path already exists
path = None
return path
DATA_ROOT = CONFIG_ROOT = None
if os.getenv('ROOTPY_GRIDMODE') not in ('1', 'true'):
DATA_ROOT = ensure_directory('ROOTPY_DATA', '${XDG_CACHE_HOME}/rootpy')
CONFIG_ROOT = ensure_directory('ROOTPY_CONFIG', '${XDG_CONFIG_HOME}/rootpy')
if DATA_ROOT is None:
log.info("Placing user data in /tmp.")
log.warning("Make sure '~/.cache/rootpy' or $ROOTPY_DATA is a writable "
"directory so that it isn't necessary to recreate all user data"
" each time")
DATA_ROOT = tempfile.mkdtemp()
@atexit.register
def __cleanup():
import shutil
shutil.rmtree(DATA_ROOT)
# ... rest of the code ... |
cd9048f64c6a2184e148daf0baa7bb3be51b3268 | vol/__init__.py | vol/__init__.py | from __future__ import unicode_literals, print_function
from sys import platform
if platform == 'darwin':
from .osx import OSXVolumeController as VolumeController
else:
raise NotImplementedError(
'VolumeController for {} platform has not been implemented yet'.format(platform)
)
| '''
A cross platform implementation of volume control
'''
from __future__ import unicode_literals, print_function
from sys import platform
if platform == 'darwin':
from .osx import OSXVolumeController as VolumeController
else:
raise NotImplementedError(
'VolumeController for {} platform has not been implemented yet'.format(platform)
)
| Update doc for vol pkg | Update doc for vol pkg
| Python | bsd-3-clause | Microcore/AGT,Microcore/YAS | + '''
+ A cross platform implementation of volume control
+ '''
from __future__ import unicode_literals, print_function
from sys import platform
if platform == 'darwin':
from .osx import OSXVolumeController as VolumeController
else:
raise NotImplementedError(
'VolumeController for {} platform has not been implemented yet'.format(platform)
)
| Update doc for vol pkg | ## Code Before:
from __future__ import unicode_literals, print_function
from sys import platform
if platform == 'darwin':
from .osx import OSXVolumeController as VolumeController
else:
raise NotImplementedError(
'VolumeController for {} platform has not been implemented yet'.format(platform)
)
## Instruction:
Update doc for vol pkg
## Code After:
'''
A cross platform implementation of volume control
'''
from __future__ import unicode_literals, print_function
from sys import platform
if platform == 'darwin':
from .osx import OSXVolumeController as VolumeController
else:
raise NotImplementedError(
'VolumeController for {} platform has not been implemented yet'.format(platform)
)
| // ... existing code ...
'''
A cross platform implementation of volume control
'''
from __future__ import unicode_literals, print_function
from sys import platform
// ... rest of the code ... |
0ea1153438c1d98232a921c8d14d401a541e95fd | examples/regex/regex_parser.py | examples/regex/regex_parser.py | from __future__ import print_function, division, absolute_import, unicode_literals
from parser_base import RegexParser
import model
class RegexSemantics(object):
def __init__(self):
super(RegexSemantics, self).__init__()
self._count = 0
def START(self, ast):
return model.Regex(ast)
def CHOICE(self, ast):
return model.Choice(ast.opts)
def SEQUENCE(self, ast):
if not ast.terms:
return model.Empty()
elif len(ast.terms) < 2:
return ast.terms[0]
else:
return model.Sequence(ast.terms)
def CLOSURE(self, ast):
return model.Closure(ast)
def SUBEXP(self, ast):
return ast
def LITERAL(self, ast):
return model.Literal(ast)
def translate(regex, trace=False):
parser = RegexParser(trace=trace, semantics=RegexSemantics())
model = parser.parse(regex, 'START')
model.set_rule_numbers()
return model.render()
| from __future__ import print_function, division, absolute_import, unicode_literals
from parser_base import RegexParser
import model
class RegexSemantics(object):
def __init__(self):
super(RegexSemantics, self).__init__()
self._count = 0
def START(self, ast):
return model.Regex(ast)
def CHOICE(self, ast):
return model.Choice(ast.opts)
def SEQUENCE(self, ast):
if not ast.terms:
return model.Empty()
elif len(ast.terms) < 2:
return ast.terms[0]
else:
return model.Sequence(ast.terms)
def CLOSURE(self, ast):
return model.Closure(ast)
def SUBEXP(self, ast):
return ast
def LITERAL(self, ast):
return model.Literal(ast)
def translate(regex, trace=False):
parser = RegexParser(trace=trace, semantics=RegexSemantics())
model = parser.parse(regex, 'START')
model.set_rule_numbers()
return model.render().encode("ascii")
| Fix regex example, the model must not be a unicode string. | Fix regex example, the model must not be a unicode string.
| Python | bsd-2-clause | vmuriart/grako,frnknglrt/grako | from __future__ import print_function, division, absolute_import, unicode_literals
from parser_base import RegexParser
import model
class RegexSemantics(object):
def __init__(self):
super(RegexSemantics, self).__init__()
self._count = 0
def START(self, ast):
return model.Regex(ast)
def CHOICE(self, ast):
return model.Choice(ast.opts)
def SEQUENCE(self, ast):
if not ast.terms:
return model.Empty()
elif len(ast.terms) < 2:
return ast.terms[0]
else:
return model.Sequence(ast.terms)
def CLOSURE(self, ast):
return model.Closure(ast)
def SUBEXP(self, ast):
return ast
def LITERAL(self, ast):
return model.Literal(ast)
def translate(regex, trace=False):
parser = RegexParser(trace=trace, semantics=RegexSemantics())
model = parser.parse(regex, 'START')
model.set_rule_numbers()
- return model.render()
+ return model.render().encode("ascii")
| Fix regex example, the model must not be a unicode string. | ## Code Before:
from __future__ import print_function, division, absolute_import, unicode_literals
from parser_base import RegexParser
import model
class RegexSemantics(object):
def __init__(self):
super(RegexSemantics, self).__init__()
self._count = 0
def START(self, ast):
return model.Regex(ast)
def CHOICE(self, ast):
return model.Choice(ast.opts)
def SEQUENCE(self, ast):
if not ast.terms:
return model.Empty()
elif len(ast.terms) < 2:
return ast.terms[0]
else:
return model.Sequence(ast.terms)
def CLOSURE(self, ast):
return model.Closure(ast)
def SUBEXP(self, ast):
return ast
def LITERAL(self, ast):
return model.Literal(ast)
def translate(regex, trace=False):
parser = RegexParser(trace=trace, semantics=RegexSemantics())
model = parser.parse(regex, 'START')
model.set_rule_numbers()
return model.render()
## Instruction:
Fix regex example, the model must not be a unicode string.
## Code After:
from __future__ import print_function, division, absolute_import, unicode_literals
from parser_base import RegexParser
import model
class RegexSemantics(object):
def __init__(self):
super(RegexSemantics, self).__init__()
self._count = 0
def START(self, ast):
return model.Regex(ast)
def CHOICE(self, ast):
return model.Choice(ast.opts)
def SEQUENCE(self, ast):
if not ast.terms:
return model.Empty()
elif len(ast.terms) < 2:
return ast.terms[0]
else:
return model.Sequence(ast.terms)
def CLOSURE(self, ast):
return model.Closure(ast)
def SUBEXP(self, ast):
return ast
def LITERAL(self, ast):
return model.Literal(ast)
def translate(regex, trace=False):
parser = RegexParser(trace=trace, semantics=RegexSemantics())
model = parser.parse(regex, 'START')
model.set_rule_numbers()
return model.render().encode("ascii")
| # ... existing code ...
model = parser.parse(regex, 'START')
model.set_rule_numbers()
return model.render().encode("ascii")
# ... rest of the code ... |
03d628abc4711bb0de4a7a0ef13cc4c0ecb92032 | opps/articles/tests/models.py | opps/articles/tests/models.py | from django.test import TestCase
from opps.articles.models import Post
class PostModelTest(TestCase):
fixtures = ['tests/initial_data.json']
def test_basic_post_exist(self):
post = Post.objects.all()
self.assertTrue(post)
self.assertEqual(len(post), 1)
self.assertEqual(post[0].slug, u'test-post-application')
self.assertEqual(post[0].title, u'test post application')
self.assertTrue(post[0].short_url)
| from django.test import TestCase
from opps.articles.models import Post
class PostModelTest(TestCase):
fixtures = ['tests/initial_data.json']
def test_basic_post_exist(self):
post = Post.objects.all()
self.assertTrue(post)
self.assertEqual(len(post), 1)
self.assertEqual(post[0].slug, u'test-post-application')
self.assertEqual(post[0].title, u'test post application')
self.assertTrue(post[0].short_url)
def test_child_class(self):
post = Post.objects.get(id=1)
self.assertTrue(post.child_class)
self.assertEqual(post.child_class, 'Post')
| Add test articles (post), check child_class | Add test articles (post), check child_class
| Python | mit | YACOWS/opps,jeanmask/opps,williamroot/opps,opps/opps,YACOWS/opps,opps/opps,jeanmask/opps,opps/opps,williamroot/opps,YACOWS/opps,jeanmask/opps,williamroot/opps,YACOWS/opps,jeanmask/opps,opps/opps,williamroot/opps | from django.test import TestCase
from opps.articles.models import Post
class PostModelTest(TestCase):
fixtures = ['tests/initial_data.json']
def test_basic_post_exist(self):
post = Post.objects.all()
self.assertTrue(post)
self.assertEqual(len(post), 1)
self.assertEqual(post[0].slug, u'test-post-application')
self.assertEqual(post[0].title, u'test post application')
self.assertTrue(post[0].short_url)
+ def test_child_class(self):
+ post = Post.objects.get(id=1)
+
+ self.assertTrue(post.child_class)
+ self.assertEqual(post.child_class, 'Post')
+ | Add test articles (post), check child_class | ## Code Before:
from django.test import TestCase
from opps.articles.models import Post
class PostModelTest(TestCase):
fixtures = ['tests/initial_data.json']
def test_basic_post_exist(self):
post = Post.objects.all()
self.assertTrue(post)
self.assertEqual(len(post), 1)
self.assertEqual(post[0].slug, u'test-post-application')
self.assertEqual(post[0].title, u'test post application')
self.assertTrue(post[0].short_url)
## Instruction:
Add test articles (post), check child_class
## Code After:
from django.test import TestCase
from opps.articles.models import Post
class PostModelTest(TestCase):
fixtures = ['tests/initial_data.json']
def test_basic_post_exist(self):
post = Post.objects.all()
self.assertTrue(post)
self.assertEqual(len(post), 1)
self.assertEqual(post[0].slug, u'test-post-application')
self.assertEqual(post[0].title, u'test post application')
self.assertTrue(post[0].short_url)
def test_child_class(self):
post = Post.objects.get(id=1)
self.assertTrue(post.child_class)
self.assertEqual(post.child_class, 'Post')
| ...
self.assertEqual(post[0].title, u'test post application')
self.assertTrue(post[0].short_url)
def test_child_class(self):
post = Post.objects.get(id=1)
self.assertTrue(post.child_class)
self.assertEqual(post.child_class, 'Post')
... |
6af05b8af7bb284388af4960bbf240122f7f3dae | plugins/PerObjectSettingsTool/__init__.py | plugins/PerObjectSettingsTool/__init__.py |
from . import PerObjectSettingsTool
from UM.i18n import i18nCatalog
i18n_catalog = i18nCatalog("uranium")
def getMetaData():
return {
"plugin": {
"name": i18n_catalog.i18nc("@label", "Per Object Settings Tool"),
"author": "Ultimaker",
"version": "1.0",
"description": i18n_catalog.i18nc("@info:whatsthis", "Provides the Per Object Settings."),
"api": 2
},
"tool": {
"name": i18n_catalog.i18nc("@label", "Per Object Settings"),
"description": i18n_catalog.i18nc("@info:tooltip", "Configure Per Object Settings"),
"icon": "setting_per_object",
"tool_panel": "PerObjectSettingsPanel.qml"
},
}
def register(app):
return { "tool": PerObjectSettingsTool.PerObjectSettingsTool() }
|
from . import PerObjectSettingsTool
from UM.i18n import i18nCatalog
i18n_catalog = i18nCatalog("uranium")
def getMetaData():
return {
"plugin": {
"name": i18n_catalog.i18nc("@label", "Per Object Settings Tool"),
"author": "Ultimaker",
"version": "1.0",
"description": i18n_catalog.i18nc("@info:whatsthis", "Provides the Per Object Settings."),
"api": 2
},
"tool": {
"name": i18n_catalog.i18nc("@label", "Per Object Settings"),
"description": i18n_catalog.i18nc("@info:tooltip", "Configure Per Object Settings"),
"icon": "setting_per_object",
"tool_panel": "PerObjectSettingsPanel.qml",
"weight": 3
},
}
def register(app):
return { "tool": PerObjectSettingsTool.PerObjectSettingsTool() }
| Add order to PerObjectSettings tool | Add order to PerObjectSettings tool
| Python | agpl-3.0 | ynotstartups/Wanhao,hmflash/Cura,fieldOfView/Cura,senttech/Cura,hmflash/Cura,fieldOfView/Cura,Curahelper/Cura,senttech/Cura,Curahelper/Cura,totalretribution/Cura,totalretribution/Cura,ynotstartups/Wanhao |
from . import PerObjectSettingsTool
from UM.i18n import i18nCatalog
i18n_catalog = i18nCatalog("uranium")
def getMetaData():
return {
"plugin": {
"name": i18n_catalog.i18nc("@label", "Per Object Settings Tool"),
"author": "Ultimaker",
"version": "1.0",
"description": i18n_catalog.i18nc("@info:whatsthis", "Provides the Per Object Settings."),
"api": 2
},
"tool": {
"name": i18n_catalog.i18nc("@label", "Per Object Settings"),
"description": i18n_catalog.i18nc("@info:tooltip", "Configure Per Object Settings"),
"icon": "setting_per_object",
- "tool_panel": "PerObjectSettingsPanel.qml"
+ "tool_panel": "PerObjectSettingsPanel.qml",
+ "weight": 3
},
}
def register(app):
return { "tool": PerObjectSettingsTool.PerObjectSettingsTool() }
| Add order to PerObjectSettings tool | ## Code Before:
from . import PerObjectSettingsTool
from UM.i18n import i18nCatalog
i18n_catalog = i18nCatalog("uranium")
def getMetaData():
return {
"plugin": {
"name": i18n_catalog.i18nc("@label", "Per Object Settings Tool"),
"author": "Ultimaker",
"version": "1.0",
"description": i18n_catalog.i18nc("@info:whatsthis", "Provides the Per Object Settings."),
"api": 2
},
"tool": {
"name": i18n_catalog.i18nc("@label", "Per Object Settings"),
"description": i18n_catalog.i18nc("@info:tooltip", "Configure Per Object Settings"),
"icon": "setting_per_object",
"tool_panel": "PerObjectSettingsPanel.qml"
},
}
def register(app):
return { "tool": PerObjectSettingsTool.PerObjectSettingsTool() }
## Instruction:
Add order to PerObjectSettings tool
## Code After:
from . import PerObjectSettingsTool
from UM.i18n import i18nCatalog
i18n_catalog = i18nCatalog("uranium")
def getMetaData():
return {
"plugin": {
"name": i18n_catalog.i18nc("@label", "Per Object Settings Tool"),
"author": "Ultimaker",
"version": "1.0",
"description": i18n_catalog.i18nc("@info:whatsthis", "Provides the Per Object Settings."),
"api": 2
},
"tool": {
"name": i18n_catalog.i18nc("@label", "Per Object Settings"),
"description": i18n_catalog.i18nc("@info:tooltip", "Configure Per Object Settings"),
"icon": "setting_per_object",
"tool_panel": "PerObjectSettingsPanel.qml",
"weight": 3
},
}
def register(app):
return { "tool": PerObjectSettingsTool.PerObjectSettingsTool() }
| # ... existing code ...
"description": i18n_catalog.i18nc("@info:tooltip", "Configure Per Object Settings"),
"icon": "setting_per_object",
"tool_panel": "PerObjectSettingsPanel.qml",
"weight": 3
},
}
# ... rest of the code ... |
52e9390d88062e9442b18a7793e6696a36f5b9c3 | testinfra/functional/test_tor_interfaces.py | testinfra/functional/test_tor_interfaces.py | import os
import re
import pytest
sdvars = pytest.securedrop_test_vars
@pytest.mark.xfail
@pytest.mark.parametrize('site', sdvars.tor_url_files)
@pytest.mark.skipif(os.environ.get('FPF_CI', 'false') == "false",
reason="Can only assure Tor is configured in CI atm")
def test_www(Command, site):
"""
Ensure tor interface is reachable and returns expected content.
"""
# Extract Onion URL from saved onion file, fetched back from app-staging.
onion_url_filepath = os.path.join(
os.path.dirname(__file__),
"../../install_files/ansible-base/{}".format(site['file'])
)
onion_url_raw = open(onion_url_filepath, 'ro').read()
onion_url = re.search("\w+\.onion", onion_url_raw).group()
# Fetch Onion URL via curl to confirm interface is rendered correctly.
curl_tor = 'curl -s --socks5-hostname "${{TOR_PROXY}}":9050 {}'.format(
onion_url)
curl_tor_status = '{} -o /dev/null -w "%{{http_code}}"'.format(curl_tor)
site_scrape = Command.check_output(curl_tor)
assert Command.check_output(curl_tor_status) == "200"
assert site['check_string'] in site_scrape
assert site['error_string'] not in site_scrape
| import os
import re
import pytest
sdvars = pytest.securedrop_test_vars
@pytest.mark.parametrize('site', sdvars.tor_url_files)
@pytest.mark.skipif(os.environ.get('FPF_CI', 'false') == "false",
reason="Can only assure Tor is configured in CI atm")
def test_www(Command, site):
"""
Ensure tor interface is reachable and returns expected content.
"""
# Extract Onion URL from saved onion file, fetched back from app-staging.
onion_url_filepath = os.path.join(
os.path.dirname(__file__),
"../../install_files/ansible-base/{}".format(site['file'])
)
onion_url_raw = open(onion_url_filepath, 'ro').read()
onion_url = re.search("\w+\.onion", onion_url_raw).group()
# Fetch Onion URL via curl to confirm interface is rendered correctly.
curl_tor = 'curl -s --socks5-hostname "${{TOR_PROXY}}":9050 {}'.format(
onion_url)
curl_tor_status = '{} -o /dev/null -w "%{{http_code}}"'.format(curl_tor)
site_scrape = Command.check_output(curl_tor)
assert Command.check_output(curl_tor_status) == "200"
assert site['check_string'] in site_scrape
assert site['error_string'] not in site_scrape
| Remove XFAIL on functional tor test | Remove XFAIL on functional tor test
| Python | agpl-3.0 | conorsch/securedrop,ehartsuyker/securedrop,garrettr/securedrop,ehartsuyker/securedrop,conorsch/securedrop,heartsucker/securedrop,garrettr/securedrop,ehartsuyker/securedrop,ehartsuyker/securedrop,conorsch/securedrop,ehartsuyker/securedrop,heartsucker/securedrop,conorsch/securedrop,heartsucker/securedrop,ehartsuyker/securedrop,heartsucker/securedrop,garrettr/securedrop,conorsch/securedrop,garrettr/securedrop,heartsucker/securedrop | import os
import re
import pytest
sdvars = pytest.securedrop_test_vars
- @pytest.mark.xfail
@pytest.mark.parametrize('site', sdvars.tor_url_files)
@pytest.mark.skipif(os.environ.get('FPF_CI', 'false') == "false",
reason="Can only assure Tor is configured in CI atm")
def test_www(Command, site):
"""
Ensure tor interface is reachable and returns expected content.
"""
# Extract Onion URL from saved onion file, fetched back from app-staging.
onion_url_filepath = os.path.join(
os.path.dirname(__file__),
"../../install_files/ansible-base/{}".format(site['file'])
)
onion_url_raw = open(onion_url_filepath, 'ro').read()
onion_url = re.search("\w+\.onion", onion_url_raw).group()
# Fetch Onion URL via curl to confirm interface is rendered correctly.
curl_tor = 'curl -s --socks5-hostname "${{TOR_PROXY}}":9050 {}'.format(
onion_url)
curl_tor_status = '{} -o /dev/null -w "%{{http_code}}"'.format(curl_tor)
site_scrape = Command.check_output(curl_tor)
assert Command.check_output(curl_tor_status) == "200"
assert site['check_string'] in site_scrape
assert site['error_string'] not in site_scrape
| Remove XFAIL on functional tor test | ## Code Before:
import os
import re
import pytest
sdvars = pytest.securedrop_test_vars
@pytest.mark.xfail
@pytest.mark.parametrize('site', sdvars.tor_url_files)
@pytest.mark.skipif(os.environ.get('FPF_CI', 'false') == "false",
reason="Can only assure Tor is configured in CI atm")
def test_www(Command, site):
"""
Ensure tor interface is reachable and returns expected content.
"""
# Extract Onion URL from saved onion file, fetched back from app-staging.
onion_url_filepath = os.path.join(
os.path.dirname(__file__),
"../../install_files/ansible-base/{}".format(site['file'])
)
onion_url_raw = open(onion_url_filepath, 'ro').read()
onion_url = re.search("\w+\.onion", onion_url_raw).group()
# Fetch Onion URL via curl to confirm interface is rendered correctly.
curl_tor = 'curl -s --socks5-hostname "${{TOR_PROXY}}":9050 {}'.format(
onion_url)
curl_tor_status = '{} -o /dev/null -w "%{{http_code}}"'.format(curl_tor)
site_scrape = Command.check_output(curl_tor)
assert Command.check_output(curl_tor_status) == "200"
assert site['check_string'] in site_scrape
assert site['error_string'] not in site_scrape
## Instruction:
Remove XFAIL on functional tor test
## Code After:
import os
import re
import pytest
sdvars = pytest.securedrop_test_vars
@pytest.mark.parametrize('site', sdvars.tor_url_files)
@pytest.mark.skipif(os.environ.get('FPF_CI', 'false') == "false",
reason="Can only assure Tor is configured in CI atm")
def test_www(Command, site):
"""
Ensure tor interface is reachable and returns expected content.
"""
# Extract Onion URL from saved onion file, fetched back from app-staging.
onion_url_filepath = os.path.join(
os.path.dirname(__file__),
"../../install_files/ansible-base/{}".format(site['file'])
)
onion_url_raw = open(onion_url_filepath, 'ro').read()
onion_url = re.search("\w+\.onion", onion_url_raw).group()
# Fetch Onion URL via curl to confirm interface is rendered correctly.
curl_tor = 'curl -s --socks5-hostname "${{TOR_PROXY}}":9050 {}'.format(
onion_url)
curl_tor_status = '{} -o /dev/null -w "%{{http_code}}"'.format(curl_tor)
site_scrape = Command.check_output(curl_tor)
assert Command.check_output(curl_tor_status) == "200"
assert site['check_string'] in site_scrape
assert site['error_string'] not in site_scrape
| // ... existing code ...
@pytest.mark.parametrize('site', sdvars.tor_url_files)
@pytest.mark.skipif(os.environ.get('FPF_CI', 'false') == "false",
// ... rest of the code ... |
548cfea821bf1b0b92ce09c54405554d264b5395 | tests/integration/session/test_timeout.py | tests/integration/session/test_timeout.py | import time
from app import settings
from tests.integration.integration_test_case import IntegrationTestCase
class TestTimeout(IntegrationTestCase):
def setUp(self):
settings.EQ_SESSION_TIMEOUT_SECONDS = 1
settings.EQ_SESSION_TIMEOUT_GRACE_PERIOD_SECONDS = 0
super().setUp()
def test_timeout_continue_returns_200(self):
self.launchSurvey('test', 'timeout')
self.get('/timeout-continue')
self.assertStatusOK()
def test_when_session_times_out_server_side_401_is_returned(self):
self.launchSurvey('test', 'timeout')
time.sleep(2)
self.get(self.last_url)
self.assertStatusUnauthorised()
def test_schema_defined_timeout_is_used(self):
self.launchSurvey('test', 'timeout')
self.assertInPage('window.__EQ_SESSION_TIMEOUT__ = 1')
def test_schema_defined_timeout_cant_be_higher_than_server(self):
self._application.config['EQ_SESSION_TIMEOUT_SECONDS'] = 10
self.launchSurvey('test', 'timeout')
self.assertInPage('window.__EQ_SESSION_TIMEOUT__ = 6')
| import time
from app import settings
from tests.integration.integration_test_case import IntegrationTestCase
class TestTimeout(IntegrationTestCase):
def setUp(self):
settings.EQ_SESSION_TIMEOUT_SECONDS = 1
settings.EQ_SESSION_TIMEOUT_GRACE_PERIOD_SECONDS = 0
super().setUp()
def tearDown(self):
settings.EQ_SESSION_TIMEOUT_SECONDS = 45 * 60
settings.EQ_SESSION_TIMEOUT_GRACE_PERIOD_SECONDS = 30
super().tearDown()
def test_timeout_continue_returns_200(self):
self.launchSurvey('test', 'timeout')
self.get('/timeout-continue')
self.assertStatusOK()
def test_when_session_times_out_server_side_401_is_returned(self):
self.launchSurvey('test', 'timeout')
time.sleep(2)
self.get(self.last_url)
self.assertStatusUnauthorised()
def test_schema_defined_timeout_is_used(self):
self.launchSurvey('test', 'timeout')
self.assertInPage('window.__EQ_SESSION_TIMEOUT__ = 1')
def test_schema_defined_timeout_cant_be_higher_than_server(self):
self._application.config['EQ_SESSION_TIMEOUT_SECONDS'] = 10
self.launchSurvey('test', 'timeout')
self.assertInPage('window.__EQ_SESSION_TIMEOUT__ = 6')
| Fix CSRF missing errors that happen occasionally in tests | Fix CSRF missing errors that happen occasionally in tests
| Python | mit | ONSdigital/eq-survey-runner,ONSdigital/eq-survey-runner,ONSdigital/eq-survey-runner,ONSdigital/eq-survey-runner | import time
from app import settings
from tests.integration.integration_test_case import IntegrationTestCase
class TestTimeout(IntegrationTestCase):
def setUp(self):
settings.EQ_SESSION_TIMEOUT_SECONDS = 1
settings.EQ_SESSION_TIMEOUT_GRACE_PERIOD_SECONDS = 0
super().setUp()
+
+ def tearDown(self):
+ settings.EQ_SESSION_TIMEOUT_SECONDS = 45 * 60
+ settings.EQ_SESSION_TIMEOUT_GRACE_PERIOD_SECONDS = 30
+ super().tearDown()
def test_timeout_continue_returns_200(self):
self.launchSurvey('test', 'timeout')
self.get('/timeout-continue')
self.assertStatusOK()
def test_when_session_times_out_server_side_401_is_returned(self):
self.launchSurvey('test', 'timeout')
time.sleep(2)
self.get(self.last_url)
self.assertStatusUnauthorised()
def test_schema_defined_timeout_is_used(self):
self.launchSurvey('test', 'timeout')
self.assertInPage('window.__EQ_SESSION_TIMEOUT__ = 1')
def test_schema_defined_timeout_cant_be_higher_than_server(self):
self._application.config['EQ_SESSION_TIMEOUT_SECONDS'] = 10
self.launchSurvey('test', 'timeout')
self.assertInPage('window.__EQ_SESSION_TIMEOUT__ = 6')
| Fix CSRF missing errors that happen occasionally in tests | ## Code Before:
import time
from app import settings
from tests.integration.integration_test_case import IntegrationTestCase
class TestTimeout(IntegrationTestCase):
def setUp(self):
settings.EQ_SESSION_TIMEOUT_SECONDS = 1
settings.EQ_SESSION_TIMEOUT_GRACE_PERIOD_SECONDS = 0
super().setUp()
def test_timeout_continue_returns_200(self):
self.launchSurvey('test', 'timeout')
self.get('/timeout-continue')
self.assertStatusOK()
def test_when_session_times_out_server_side_401_is_returned(self):
self.launchSurvey('test', 'timeout')
time.sleep(2)
self.get(self.last_url)
self.assertStatusUnauthorised()
def test_schema_defined_timeout_is_used(self):
self.launchSurvey('test', 'timeout')
self.assertInPage('window.__EQ_SESSION_TIMEOUT__ = 1')
def test_schema_defined_timeout_cant_be_higher_than_server(self):
self._application.config['EQ_SESSION_TIMEOUT_SECONDS'] = 10
self.launchSurvey('test', 'timeout')
self.assertInPage('window.__EQ_SESSION_TIMEOUT__ = 6')
## Instruction:
Fix CSRF missing errors that happen occasionally in tests
## Code After:
import time
from app import settings
from tests.integration.integration_test_case import IntegrationTestCase
class TestTimeout(IntegrationTestCase):
def setUp(self):
settings.EQ_SESSION_TIMEOUT_SECONDS = 1
settings.EQ_SESSION_TIMEOUT_GRACE_PERIOD_SECONDS = 0
super().setUp()
def tearDown(self):
settings.EQ_SESSION_TIMEOUT_SECONDS = 45 * 60
settings.EQ_SESSION_TIMEOUT_GRACE_PERIOD_SECONDS = 30
super().tearDown()
def test_timeout_continue_returns_200(self):
self.launchSurvey('test', 'timeout')
self.get('/timeout-continue')
self.assertStatusOK()
def test_when_session_times_out_server_side_401_is_returned(self):
self.launchSurvey('test', 'timeout')
time.sleep(2)
self.get(self.last_url)
self.assertStatusUnauthorised()
def test_schema_defined_timeout_is_used(self):
self.launchSurvey('test', 'timeout')
self.assertInPage('window.__EQ_SESSION_TIMEOUT__ = 1')
def test_schema_defined_timeout_cant_be_higher_than_server(self):
self._application.config['EQ_SESSION_TIMEOUT_SECONDS'] = 10
self.launchSurvey('test', 'timeout')
self.assertInPage('window.__EQ_SESSION_TIMEOUT__ = 6')
| ...
settings.EQ_SESSION_TIMEOUT_GRACE_PERIOD_SECONDS = 0
super().setUp()
def tearDown(self):
settings.EQ_SESSION_TIMEOUT_SECONDS = 45 * 60
settings.EQ_SESSION_TIMEOUT_GRACE_PERIOD_SECONDS = 30
super().tearDown()
def test_timeout_continue_returns_200(self):
... |
2664e9124af6b0d8f6b2eacd50f4d7e93b91e931 | examples/GoBot/gobot.py | examples/GoBot/gobot.py | from minibot.bot import Bot
from minibot.hardware.rpi.gpio import PWM
from minibot.interface.servo import Servo
import math
import time
L_MOTOR_PIN = 12
R_MOTOR_PIN = 18
class GoBot(Bot):
def __init__(self):
Bot.__init__(self, "GoBot")
self.l_motor = Servo(PWM(L_MOTOR_PIN, 2, 15))
self.r_motor = Servo(PWM(R_MOTOR_PIN, 2, 15))
self.l_motor.set(17)
self.r_motor.set(13)
def run(self):
pass
if __name__ == "__main__":
bot = GoBot()
while True:
bot.run()
|
from minibot.bot import Bot
from minibot.hardware.rpi.gpio import PWM
from minibot.interface.servo import Servo
L_MOTOR_PIN = 12
R_MOTOR_PIN = 18
class GoBot(Bot):
"""
GoBot
"""
def __init__(self):
Bot.__init__(self, "GoBot")
self.l_motor = Servo(PWM(L_MOTOR_PIN, 2, 15))
self.r_motor = Servo(PWM(R_MOTOR_PIN, 2, 15))
self.l_motor.set(17)
self.r_motor.set(13)
def run(self):
pass
if __name__ == "__main__":
bot = GoBot()
while True:
bot.run()
| Fix linting errors in GoBot | Fix linting errors in GoBot
| Python | apache-2.0 | cornell-cup/cs-minibot,cornell-cup/cs-minibot,cornell-cup/cs-minibot,cornell-cup/cs-minibot,cornell-cup/cs-minibot | +
from minibot.bot import Bot
from minibot.hardware.rpi.gpio import PWM
from minibot.interface.servo import Servo
-
- import math
- import time
L_MOTOR_PIN = 12
R_MOTOR_PIN = 18
class GoBot(Bot):
+ """
+ GoBot
+ """
def __init__(self):
Bot.__init__(self, "GoBot")
self.l_motor = Servo(PWM(L_MOTOR_PIN, 2, 15))
self.r_motor = Servo(PWM(R_MOTOR_PIN, 2, 15))
self.l_motor.set(17)
self.r_motor.set(13)
def run(self):
pass
if __name__ == "__main__":
bot = GoBot()
while True:
bot.run()
| Fix linting errors in GoBot | ## Code Before:
from minibot.bot import Bot
from minibot.hardware.rpi.gpio import PWM
from minibot.interface.servo import Servo
import math
import time
L_MOTOR_PIN = 12
R_MOTOR_PIN = 18
class GoBot(Bot):
def __init__(self):
Bot.__init__(self, "GoBot")
self.l_motor = Servo(PWM(L_MOTOR_PIN, 2, 15))
self.r_motor = Servo(PWM(R_MOTOR_PIN, 2, 15))
self.l_motor.set(17)
self.r_motor.set(13)
def run(self):
pass
if __name__ == "__main__":
bot = GoBot()
while True:
bot.run()
## Instruction:
Fix linting errors in GoBot
## Code After:
from minibot.bot import Bot
from minibot.hardware.rpi.gpio import PWM
from minibot.interface.servo import Servo
L_MOTOR_PIN = 12
R_MOTOR_PIN = 18
class GoBot(Bot):
"""
GoBot
"""
def __init__(self):
Bot.__init__(self, "GoBot")
self.l_motor = Servo(PWM(L_MOTOR_PIN, 2, 15))
self.r_motor = Servo(PWM(R_MOTOR_PIN, 2, 15))
self.l_motor.set(17)
self.r_motor.set(13)
def run(self):
pass
if __name__ == "__main__":
bot = GoBot()
while True:
bot.run()
| ...
from minibot.bot import Bot
from minibot.hardware.rpi.gpio import PWM
from minibot.interface.servo import Servo
L_MOTOR_PIN = 12
...
class GoBot(Bot):
"""
GoBot
"""
def __init__(self):
... |
dcecdbae798e0a83afb17911ec459224790e51cd | launch_control/dashboard_app/tests.py | launch_control/dashboard_app/tests.py |
from django.test import TestCase
from django.db import IntegrityError
from launch_control.dashboard_app.models import (
SoftwarePackage,
)
class SoftwarePackageTestCase(TestCase):
def test_creation_1(self):
sw_package = SoftwarePackage.objects.create(name='libfoo', version='1.2.0')
self.assertEqual(sw_package.name, 'libfoo')
self.assertEqual(sw_package.version, '1.2.0')
def test_uniqueness(self):
SoftwarePackage.objects.create(name='a', version='0')
self.assertRaises(IntegrityError, SoftwarePackage.objects.create,
name='a', version='0')
|
from django.test import TestCase
from django.db import IntegrityError
from launch_control.utils.call_helper import ObjectFactoryMixIn
from launch_control.dashboard_app.models import (
SoftwarePackage,
)
class SoftwarePackageTestCase(TestCase, ObjectFactoryMixIn):
class Dummy:
class SoftwarePackage:
name = 'libfoo'
version = '1.2.0'
def test_creation_1(self):
dummy, sw_package = self.make_and_get_dummy(SoftwarePackage)
self.assertEqual(sw_package.name, dummy.name)
self.assertEqual(sw_package.version, dummy.version)
def test_uniqueness(self):
pkg1 = self.make(SoftwarePackage)
pkg1.save()
pkg2 = self.make(SoftwarePackage)
self.assertRaises(IntegrityError, pkg2.save)
| Update SoftwarePackageTestCase to use ObjectFactoryMixIn | Update SoftwarePackageTestCase to use ObjectFactoryMixIn
| Python | agpl-3.0 | OSSystems/lava-server,Linaro/lava-server,OSSystems/lava-server,Linaro/lava-server,Linaro/lava-server,Linaro/lava-server,OSSystems/lava-server |
from django.test import TestCase
from django.db import IntegrityError
+ from launch_control.utils.call_helper import ObjectFactoryMixIn
from launch_control.dashboard_app.models import (
SoftwarePackage,
)
- class SoftwarePackageTestCase(TestCase):
+ class SoftwarePackageTestCase(TestCase, ObjectFactoryMixIn):
+
+ class Dummy:
+ class SoftwarePackage:
+ name = 'libfoo'
+ version = '1.2.0'
def test_creation_1(self):
- sw_package = SoftwarePackage.objects.create(name='libfoo', version='1.2.0')
+ dummy, sw_package = self.make_and_get_dummy(SoftwarePackage)
- self.assertEqual(sw_package.name, 'libfoo')
+ self.assertEqual(sw_package.name, dummy.name)
- self.assertEqual(sw_package.version, '1.2.0')
+ self.assertEqual(sw_package.version, dummy.version)
def test_uniqueness(self):
- SoftwarePackage.objects.create(name='a', version='0')
- self.assertRaises(IntegrityError, SoftwarePackage.objects.create,
- name='a', version='0')
+ pkg1 = self.make(SoftwarePackage)
+ pkg1.save()
+ pkg2 = self.make(SoftwarePackage)
+ self.assertRaises(IntegrityError, pkg2.save)
| Update SoftwarePackageTestCase to use ObjectFactoryMixIn | ## Code Before:
from django.test import TestCase
from django.db import IntegrityError
from launch_control.dashboard_app.models import (
SoftwarePackage,
)
class SoftwarePackageTestCase(TestCase):
def test_creation_1(self):
sw_package = SoftwarePackage.objects.create(name='libfoo', version='1.2.0')
self.assertEqual(sw_package.name, 'libfoo')
self.assertEqual(sw_package.version, '1.2.0')
def test_uniqueness(self):
SoftwarePackage.objects.create(name='a', version='0')
self.assertRaises(IntegrityError, SoftwarePackage.objects.create,
name='a', version='0')
## Instruction:
Update SoftwarePackageTestCase to use ObjectFactoryMixIn
## Code After:
from django.test import TestCase
from django.db import IntegrityError
from launch_control.utils.call_helper import ObjectFactoryMixIn
from launch_control.dashboard_app.models import (
SoftwarePackage,
)
class SoftwarePackageTestCase(TestCase, ObjectFactoryMixIn):
class Dummy:
class SoftwarePackage:
name = 'libfoo'
version = '1.2.0'
def test_creation_1(self):
dummy, sw_package = self.make_and_get_dummy(SoftwarePackage)
self.assertEqual(sw_package.name, dummy.name)
self.assertEqual(sw_package.version, dummy.version)
def test_uniqueness(self):
pkg1 = self.make(SoftwarePackage)
pkg1.save()
pkg2 = self.make(SoftwarePackage)
self.assertRaises(IntegrityError, pkg2.save)
| # ... existing code ...
from django.db import IntegrityError
from launch_control.utils.call_helper import ObjectFactoryMixIn
from launch_control.dashboard_app.models import (
SoftwarePackage,
# ... modified code ...
class SoftwarePackageTestCase(TestCase, ObjectFactoryMixIn):
class Dummy:
class SoftwarePackage:
name = 'libfoo'
version = '1.2.0'
def test_creation_1(self):
dummy, sw_package = self.make_and_get_dummy(SoftwarePackage)
self.assertEqual(sw_package.name, dummy.name)
self.assertEqual(sw_package.version, dummy.version)
def test_uniqueness(self):
pkg1 = self.make(SoftwarePackage)
pkg1.save()
pkg2 = self.make(SoftwarePackage)
self.assertRaises(IntegrityError, pkg2.save)
# ... rest of the code ... |
b1685dc4a0a2036378d47f07d7315e5b1935a4ad | hyrodactil/tests/openings/models.py | hyrodactil/tests/openings/models.py | from django.test import TestCase
from ..factories._applications import ApplicationFactory
from ..factories._companysettings import InterviewStageFactory
from ..factories._openings import OpeningFactory
class OpeningsModelsTests(TestCase):
def test_applicants_stats(self):
opening = OpeningFactory()
self.assertEqual(opening.applicants_stats(), [])
s1 = InterviewStageFactory(name="L220", initial=True,
company=opening.company)
s2 = InterviewStageFactory(name="L33",
company=opening.company)
self.assertEqual(opening.applicants_stats(),
[[s1.name, 0], [s2.name, 0]])
application = ApplicationFactory.create(opening=opening)
application.stage_transitions.create(stage=s1)
self.assertEqual(opening.applicants_stats(),
[[s1.name, 1], [s2.name, 0]])
| from django.test import TestCase
from ..factories._applications import ApplicationFactory
from ..factories._companysettings import InterviewStageFactory
from ..factories._openings import OpeningFactory
class OpeningsModelsTests(TestCase):
def test_applicants_stats(self):
opening = OpeningFactory()
self.assertEqual(opening.applicants_stats(), [])
s1 = InterviewStageFactory(name="L220",
company=opening.company)
s2 = InterviewStageFactory(name="L33",
company=opening.company)
self.assertEqual(opening.applicants_stats(),
[[s1.name, 0], [s2.name, 0]])
application = ApplicationFactory.create(opening=opening)
application.stage_transitions.create(stage=s1)
self.assertEqual(opening.applicants_stats(),
[[s1.name, 1], [s2.name, 0]])
| Fix tests for positioned stages | openings: Fix tests for positioned stages
| Python | mit | hizardapp/Hizard,hizardapp/Hizard,hizardapp/Hizard | from django.test import TestCase
from ..factories._applications import ApplicationFactory
from ..factories._companysettings import InterviewStageFactory
from ..factories._openings import OpeningFactory
class OpeningsModelsTests(TestCase):
def test_applicants_stats(self):
opening = OpeningFactory()
self.assertEqual(opening.applicants_stats(), [])
- s1 = InterviewStageFactory(name="L220", initial=True,
+ s1 = InterviewStageFactory(name="L220",
company=opening.company)
s2 = InterviewStageFactory(name="L33",
company=opening.company)
self.assertEqual(opening.applicants_stats(),
[[s1.name, 0], [s2.name, 0]])
application = ApplicationFactory.create(opening=opening)
application.stage_transitions.create(stage=s1)
self.assertEqual(opening.applicants_stats(),
[[s1.name, 1], [s2.name, 0]])
| Fix tests for positioned stages | ## Code Before:
from django.test import TestCase
from ..factories._applications import ApplicationFactory
from ..factories._companysettings import InterviewStageFactory
from ..factories._openings import OpeningFactory
class OpeningsModelsTests(TestCase):
def test_applicants_stats(self):
opening = OpeningFactory()
self.assertEqual(opening.applicants_stats(), [])
s1 = InterviewStageFactory(name="L220", initial=True,
company=opening.company)
s2 = InterviewStageFactory(name="L33",
company=opening.company)
self.assertEqual(opening.applicants_stats(),
[[s1.name, 0], [s2.name, 0]])
application = ApplicationFactory.create(opening=opening)
application.stage_transitions.create(stage=s1)
self.assertEqual(opening.applicants_stats(),
[[s1.name, 1], [s2.name, 0]])
## Instruction:
Fix tests for positioned stages
## Code After:
from django.test import TestCase
from ..factories._applications import ApplicationFactory
from ..factories._companysettings import InterviewStageFactory
from ..factories._openings import OpeningFactory
class OpeningsModelsTests(TestCase):
def test_applicants_stats(self):
opening = OpeningFactory()
self.assertEqual(opening.applicants_stats(), [])
s1 = InterviewStageFactory(name="L220",
company=opening.company)
s2 = InterviewStageFactory(name="L33",
company=opening.company)
self.assertEqual(opening.applicants_stats(),
[[s1.name, 0], [s2.name, 0]])
application = ApplicationFactory.create(opening=opening)
application.stage_transitions.create(stage=s1)
self.assertEqual(opening.applicants_stats(),
[[s1.name, 1], [s2.name, 0]])
| # ... existing code ...
opening = OpeningFactory()
self.assertEqual(opening.applicants_stats(), [])
s1 = InterviewStageFactory(name="L220",
company=opening.company)
s2 = InterviewStageFactory(name="L33",
# ... rest of the code ... |
20696d6f236afc1bc0e2b3db570363540e70ca84 | test/test_serve.py | test/test_serve.py | import unittest
import asyncio
import io
import multiprocessing
import urllib.request
import time
import grole
def simple_server():
app = grole.Grole()
@app.route('/')
def hello(env, req):
return 'Hello, World!'
app.run()
class TestServe(unittest.TestCase):
def test_simple(self):
p = multiprocessing.Process(target=simple_server)
p.start()
time.sleep(0.1)
with urllib.request.urlopen('http://localhost:1234') as response:
html = response.read()
self.assertEqual(html, b'Hello, World!')
p.terminate()
def test_fileserver(self):
p = multiprocessing.Process(target=grole.main, args=[[]])
p.start()
time.sleep(0.1)
with urllib.request.urlopen('http://localhost:1234/test/test.dat') as response:
html = response.read()
self.assertEqual(html, b'foo\n')
p.terminate()
| import unittest
import asyncio
import io
import multiprocessing
import urllib.request
import time
import grole
def simple_server():
app = grole.Grole()
@app.route('/')
def hello(env, req):
return 'Hello, World!'
app.run(host='127.0.0.1')
class TestServe(unittest.TestCase):
def test_simple(self):
p = multiprocessing.Process(target=simple_server)
p.start()
time.sleep(0.1)
with urllib.request.urlopen('http://127.0.0.1:1234') as response:
html = response.read()
self.assertEqual(html, b'Hello, World!')
p.terminate()
def test_fileserver(self):
p = multiprocessing.Process(target=grole.main, args=[['-a', '127.0.0.1']])
p.start()
time.sleep(0.1)
with urllib.request.urlopen('http://127.0.0.1:1234/test/test.dat') as response:
html = response.read()
self.assertEqual(html, b'foo\n')
p.terminate()
| Use ip instead of localhost for travis | Use ip instead of localhost for travis
| Python | mit | witchard/grole | import unittest
import asyncio
import io
import multiprocessing
import urllib.request
import time
import grole
def simple_server():
app = grole.Grole()
@app.route('/')
def hello(env, req):
return 'Hello, World!'
- app.run()
+ app.run(host='127.0.0.1')
class TestServe(unittest.TestCase):
def test_simple(self):
p = multiprocessing.Process(target=simple_server)
p.start()
time.sleep(0.1)
- with urllib.request.urlopen('http://localhost:1234') as response:
+ with urllib.request.urlopen('http://127.0.0.1:1234') as response:
html = response.read()
self.assertEqual(html, b'Hello, World!')
p.terminate()
def test_fileserver(self):
- p = multiprocessing.Process(target=grole.main, args=[[]])
+ p = multiprocessing.Process(target=grole.main, args=[['-a', '127.0.0.1']])
p.start()
time.sleep(0.1)
- with urllib.request.urlopen('http://localhost:1234/test/test.dat') as response:
+ with urllib.request.urlopen('http://127.0.0.1:1234/test/test.dat') as response:
html = response.read()
self.assertEqual(html, b'foo\n')
p.terminate()
| Use ip instead of localhost for travis | ## Code Before:
import unittest
import asyncio
import io
import multiprocessing
import urllib.request
import time
import grole
def simple_server():
app = grole.Grole()
@app.route('/')
def hello(env, req):
return 'Hello, World!'
app.run()
class TestServe(unittest.TestCase):
def test_simple(self):
p = multiprocessing.Process(target=simple_server)
p.start()
time.sleep(0.1)
with urllib.request.urlopen('http://localhost:1234') as response:
html = response.read()
self.assertEqual(html, b'Hello, World!')
p.terminate()
def test_fileserver(self):
p = multiprocessing.Process(target=grole.main, args=[[]])
p.start()
time.sleep(0.1)
with urllib.request.urlopen('http://localhost:1234/test/test.dat') as response:
html = response.read()
self.assertEqual(html, b'foo\n')
p.terminate()
## Instruction:
Use ip instead of localhost for travis
## Code After:
import unittest
import asyncio
import io
import multiprocessing
import urllib.request
import time
import grole
def simple_server():
app = grole.Grole()
@app.route('/')
def hello(env, req):
return 'Hello, World!'
app.run(host='127.0.0.1')
class TestServe(unittest.TestCase):
def test_simple(self):
p = multiprocessing.Process(target=simple_server)
p.start()
time.sleep(0.1)
with urllib.request.urlopen('http://127.0.0.1:1234') as response:
html = response.read()
self.assertEqual(html, b'Hello, World!')
p.terminate()
def test_fileserver(self):
p = multiprocessing.Process(target=grole.main, args=[['-a', '127.0.0.1']])
p.start()
time.sleep(0.1)
with urllib.request.urlopen('http://127.0.0.1:1234/test/test.dat') as response:
html = response.read()
self.assertEqual(html, b'foo\n')
p.terminate()
| ...
return 'Hello, World!'
app.run(host='127.0.0.1')
class TestServe(unittest.TestCase):
...
p.start()
time.sleep(0.1)
with urllib.request.urlopen('http://127.0.0.1:1234') as response:
html = response.read()
self.assertEqual(html, b'Hello, World!')
...
def test_fileserver(self):
p = multiprocessing.Process(target=grole.main, args=[['-a', '127.0.0.1']])
p.start()
time.sleep(0.1)
with urllib.request.urlopen('http://127.0.0.1:1234/test/test.dat') as response:
html = response.read()
self.assertEqual(html, b'foo\n')
... |
c07234bb3142df96dc9e02a236975bc3de2415cc | nailgun/nailgun/test/test_plugin.py | nailgun/nailgun/test/test_plugin.py | from nailgun.test.base import BaseHandlers
class TestPluginStateMachine(BaseHandlers):
def test_attrs_creation(self):
pass
| from nailgun.test.base import BaseHandlers
from nailgun.plugin.process import get_queue, PluginProcessor
from nailgun.api.models import Task
class TestPluginProcess(BaseHandlers):
def setUp(self):
super(TestPluginProcess, self).setUp()
self.plugin_processor = PluginProcessor()
self.plugin_processor.start()
def tearDown(self):
super(TestPluginProcess, self).tearDown()
self.plugin_processor.terminate()
def test_task_set_to_error_when_exception_raised(self):
queue = get_queue()
task = Task(name='install_plugin', cache={'plugin_id': -1})
self.env.db.add(task)
self.env.db.commit()
queue.put(task.uuid)
def check_task_status_is_error():
self.env.db.refresh(task)
return task.status == 'error'
self.env.wait_for_true(check_task_status_is_error, timeout=2)
self.assertEquals(task.progress, 100)
| Implement plugin test on exception handling | Implement plugin test on exception handling
| Python | apache-2.0 | SmartInfrastructures/fuel-main-dev,ddepaoli3/fuel-main-dev,zhaochao/fuel-main,zhaochao/fuel-main,huntxu/fuel-main,prmtl/fuel-web,huntxu/fuel-web,huntxu/fuel-main,SmartInfrastructures/fuel-main-dev,huntxu/fuel-web,teselkin/fuel-main,ddepaoli3/fuel-main-dev,teselkin/fuel-main,SmartInfrastructures/fuel-web-dev,SergK/fuel-main,dancn/fuel-main-dev,SmartInfrastructures/fuel-web-dev,nebril/fuel-web,SergK/fuel-main,zhaochao/fuel-web,dancn/fuel-main-dev,nebril/fuel-web,dancn/fuel-main-dev,AnselZhangGit/fuel-main,Fiware/ops.Fuel-main-dev,AnselZhangGit/fuel-main,nebril/fuel-web,SmartInfrastructures/fuel-main-dev,eayunstack/fuel-web,AnselZhangGit/fuel-main,eayunstack/fuel-web,SmartInfrastructures/fuel-web-dev,eayunstack/fuel-main,SergK/fuel-main,prmtl/fuel-web,zhaochao/fuel-web,eayunstack/fuel-main,huntxu/fuel-web,zhaochao/fuel-web,SmartInfrastructures/fuel-web-dev,koder-ua/nailgun-fcert,koder-ua/nailgun-fcert,ddepaoli3/fuel-main-dev,SmartInfrastructures/fuel-web-dev,zhaochao/fuel-web,zhaochao/fuel-web,teselkin/fuel-main,stackforge/fuel-main,prmtl/fuel-web,AnselZhangGit/fuel-main,zhaochao/fuel-main,eayunstack/fuel-web,eayunstack/fuel-main,koder-ua/nailgun-fcert,stackforge/fuel-web,nebril/fuel-web,stackforge/fuel-web,zhaochao/fuel-main,ddepaoli3/fuel-main-dev,Fiware/ops.Fuel-main-dev,prmtl/fuel-web,SmartInfrastructures/fuel-main-dev,Fiware/ops.Fuel-main-dev,zhaochao/fuel-main,koder-ua/nailgun-fcert,huntxu/fuel-web,dancn/fuel-main-dev,stackforge/fuel-main,eayunstack/fuel-web,prmtl/fuel-web,stackforge/fuel-web,huntxu/fuel-web,huntxu/fuel-main,Fiware/ops.Fuel-main-dev,teselkin/fuel-main,nebril/fuel-web,eayunstack/fuel-web,stackforge/fuel-main | from nailgun.test.base import BaseHandlers
+ from nailgun.plugin.process import get_queue, PluginProcessor
+ from nailgun.api.models import Task
+ class TestPluginProcess(BaseHandlers):
+ def setUp(self):
+ super(TestPluginProcess, self).setUp()
+ self.plugin_processor = PluginProcessor()
+ self.plugin_processor.start()
- class TestPluginStateMachine(BaseHandlers):
+ def tearDown(self):
+ super(TestPluginProcess, self).tearDown()
+ self.plugin_processor.terminate()
- def test_attrs_creation(self):
- pass
+ def test_task_set_to_error_when_exception_raised(self):
+ queue = get_queue()
+ task = Task(name='install_plugin', cache={'plugin_id': -1})
+ self.env.db.add(task)
+ self.env.db.commit()
+ queue.put(task.uuid)
+
+ def check_task_status_is_error():
+ self.env.db.refresh(task)
+ return task.status == 'error'
+
+ self.env.wait_for_true(check_task_status_is_error, timeout=2)
+ self.assertEquals(task.progress, 100)
+ | Implement plugin test on exception handling | ## Code Before:
from nailgun.test.base import BaseHandlers
class TestPluginStateMachine(BaseHandlers):
def test_attrs_creation(self):
pass
## Instruction:
Implement plugin test on exception handling
## Code After:
from nailgun.test.base import BaseHandlers
from nailgun.plugin.process import get_queue, PluginProcessor
from nailgun.api.models import Task
class TestPluginProcess(BaseHandlers):
def setUp(self):
super(TestPluginProcess, self).setUp()
self.plugin_processor = PluginProcessor()
self.plugin_processor.start()
def tearDown(self):
super(TestPluginProcess, self).tearDown()
self.plugin_processor.terminate()
def test_task_set_to_error_when_exception_raised(self):
queue = get_queue()
task = Task(name='install_plugin', cache={'plugin_id': -1})
self.env.db.add(task)
self.env.db.commit()
queue.put(task.uuid)
def check_task_status_is_error():
self.env.db.refresh(task)
return task.status == 'error'
self.env.wait_for_true(check_task_status_is_error, timeout=2)
self.assertEquals(task.progress, 100)
| // ... existing code ...
from nailgun.test.base import BaseHandlers
from nailgun.plugin.process import get_queue, PluginProcessor
from nailgun.api.models import Task
class TestPluginProcess(BaseHandlers):
def setUp(self):
super(TestPluginProcess, self).setUp()
self.plugin_processor = PluginProcessor()
self.plugin_processor.start()
def tearDown(self):
super(TestPluginProcess, self).tearDown()
self.plugin_processor.terminate()
def test_task_set_to_error_when_exception_raised(self):
queue = get_queue()
task = Task(name='install_plugin', cache={'plugin_id': -1})
self.env.db.add(task)
self.env.db.commit()
queue.put(task.uuid)
def check_task_status_is_error():
self.env.db.refresh(task)
return task.status == 'error'
self.env.wait_for_true(check_task_status_is_error, timeout=2)
self.assertEquals(task.progress, 100)
// ... rest of the code ... |
f37d26541d6baf3da47a8f373a8c7a65177067db | push/modules/push_notification.py | push/modules/push_notification.py |
import time, os, json
from apns import APNs, Frame, Payload
from push.models import DevelopFileModel, ProductFileModel
from django.conf import settings
PEM_FILE_DIR = settings.BASE_DIR + '/push/files/'
def execute(device_token_lists, notification):
if notification.is_production:
pem_file_name = ProductFileModel.objects.all()[0].production_file_name
apns = APNs(use_sandbox = False, cert_file = PEM_FILE_DIR + pem_file_name, enhanced = True)
else:
pem_file_name = DevelopFileModel.objects.all()[0].development_file_name
apns = APNs(use_sandbox = True, cert_file = PEM_FILE_DIR + pem_file_name, enhanced = True)
token_hex = []
for token in device_token_lists:
token_hex.append(token)
json_data = ''
if notification.json != '':
json_data = json.loads(notification.json)
payload = Payload(alert = notification.message,
sound = notification.sound,
badge = notification.badge,
custom = json_data)
frame = Frame()
identifier = 1
expiry = time.time() + 3600
priority = 10
for token in token_hex:
frame.add_item(token, payload, identifier, expiry, priority)
apns.gateway_server.send_notification_multiple(frame)
|
import time, os, json
from apns import APNs, Frame, Payload
from push.models import DevelopFileModel, ProductFileModel
from django.conf import settings
PEM_FILE_DIR = settings.BASE_DIR + '/push/files/'
def execute(device_token_lists, notification):
if notification.is_production:
pem_file_name = ProductFileModel.objects.all()[0].production_file_name
apns = APNs(use_sandbox = False, cert_file = PEM_FILE_DIR + pem_file_name, enhanced = True)
else:
pem_file_name = DevelopFileModel.objects.all()[0].development_file_name
apns = APNs(use_sandbox = True, cert_file = PEM_FILE_DIR + pem_file_name, enhanced = True)
token_hex = []
for token in device_token_lists:
token_hex.append(token)
json_data = ''
if notification.json != '':
json_data = json.loads(notification.json)
payload = Payload(alert = notification.message,
sound = notification.sound,
badge = notification.badge,
custom = json_data)
frame = Frame()
identifier = 1
expiry = time.time() + 3600
priority = 10
for token in token_hex:
frame.add_item(token, payload, identifier, expiry, priority)
apns.gateway_server.send_notification_multiple(frame)
notification.is_sent = True
notification.save()
| Send flag when success push notifications | Send flag when success push notifications
| Python | apache-2.0 | nnsnodnb/django-mbaas,nnsnodnb/django-mbaas,nnsnodnb/django-mbaas |
import time, os, json
from apns import APNs, Frame, Payload
from push.models import DevelopFileModel, ProductFileModel
from django.conf import settings
PEM_FILE_DIR = settings.BASE_DIR + '/push/files/'
def execute(device_token_lists, notification):
if notification.is_production:
pem_file_name = ProductFileModel.objects.all()[0].production_file_name
apns = APNs(use_sandbox = False, cert_file = PEM_FILE_DIR + pem_file_name, enhanced = True)
else:
pem_file_name = DevelopFileModel.objects.all()[0].development_file_name
apns = APNs(use_sandbox = True, cert_file = PEM_FILE_DIR + pem_file_name, enhanced = True)
token_hex = []
for token in device_token_lists:
token_hex.append(token)
json_data = ''
if notification.json != '':
json_data = json.loads(notification.json)
payload = Payload(alert = notification.message,
sound = notification.sound,
badge = notification.badge,
custom = json_data)
frame = Frame()
identifier = 1
expiry = time.time() + 3600
priority = 10
for token in token_hex:
frame.add_item(token, payload, identifier, expiry, priority)
apns.gateway_server.send_notification_multiple(frame)
+ notification.is_sent = True
+ notification.save()
+ | Send flag when success push notifications | ## Code Before:
import time, os, json
from apns import APNs, Frame, Payload
from push.models import DevelopFileModel, ProductFileModel
from django.conf import settings
PEM_FILE_DIR = settings.BASE_DIR + '/push/files/'
def execute(device_token_lists, notification):
if notification.is_production:
pem_file_name = ProductFileModel.objects.all()[0].production_file_name
apns = APNs(use_sandbox = False, cert_file = PEM_FILE_DIR + pem_file_name, enhanced = True)
else:
pem_file_name = DevelopFileModel.objects.all()[0].development_file_name
apns = APNs(use_sandbox = True, cert_file = PEM_FILE_DIR + pem_file_name, enhanced = True)
token_hex = []
for token in device_token_lists:
token_hex.append(token)
json_data = ''
if notification.json != '':
json_data = json.loads(notification.json)
payload = Payload(alert = notification.message,
sound = notification.sound,
badge = notification.badge,
custom = json_data)
frame = Frame()
identifier = 1
expiry = time.time() + 3600
priority = 10
for token in token_hex:
frame.add_item(token, payload, identifier, expiry, priority)
apns.gateway_server.send_notification_multiple(frame)
## Instruction:
Send flag when success push notifications
## Code After:
import time, os, json
from apns import APNs, Frame, Payload
from push.models import DevelopFileModel, ProductFileModel
from django.conf import settings
PEM_FILE_DIR = settings.BASE_DIR + '/push/files/'
def execute(device_token_lists, notification):
if notification.is_production:
pem_file_name = ProductFileModel.objects.all()[0].production_file_name
apns = APNs(use_sandbox = False, cert_file = PEM_FILE_DIR + pem_file_name, enhanced = True)
else:
pem_file_name = DevelopFileModel.objects.all()[0].development_file_name
apns = APNs(use_sandbox = True, cert_file = PEM_FILE_DIR + pem_file_name, enhanced = True)
token_hex = []
for token in device_token_lists:
token_hex.append(token)
json_data = ''
if notification.json != '':
json_data = json.loads(notification.json)
payload = Payload(alert = notification.message,
sound = notification.sound,
badge = notification.badge,
custom = json_data)
frame = Frame()
identifier = 1
expiry = time.time() + 3600
priority = 10
for token in token_hex:
frame.add_item(token, payload, identifier, expiry, priority)
apns.gateway_server.send_notification_multiple(frame)
notification.is_sent = True
notification.save()
| ...
apns.gateway_server.send_notification_multiple(frame)
notification.is_sent = True
notification.save()
... |
8623aae8778307648e4a0380d84ca7dc7a63f3f2 | oneflow/core/context_processors.py | oneflow/core/context_processors.py |
from .models.nonrel import User
def mongodb_user(request):
if request.user.is_anonymous():
return {u'mongodb_user': None}
try:
mongodb_user = User.objects.get(id=request.session[u'mongodb_user_id'])
except KeyError:
mongodb_user = User.objects.get(django_user=request.user.id)
# Cache it for next time.
request.session[u'mongodb_user_id'] = mongodb_user.id
return {u'mongodb_user': mongodb_user}
|
def mongodb_user(request):
""" not the most usefull context manager in the world. """
if request.user.is_anonymous():
return {u'mongodb_user': None}
return {u'mongodb_user': request.user.mongo}
| Simplify the context processor. Not very useful anymore, in fact. | Simplify the context processor. Not very useful anymore, in fact.
| Python | agpl-3.0 | 1flow/1flow,1flow/1flow,WillianPaiva/1flow,WillianPaiva/1flow,1flow/1flow,WillianPaiva/1flow,WillianPaiva/1flow,1flow/1flow,1flow/1flow,WillianPaiva/1flow | -
- from .models.nonrel import User
def mongodb_user(request):
+ """ not the most usefull context manager in the world. """
if request.user.is_anonymous():
return {u'mongodb_user': None}
+ return {u'mongodb_user': request.user.mongo}
- try:
- mongodb_user = User.objects.get(id=request.session[u'mongodb_user_id'])
- except KeyError:
- mongodb_user = User.objects.get(django_user=request.user.id)
-
- # Cache it for next time.
- request.session[u'mongodb_user_id'] = mongodb_user.id
-
- return {u'mongodb_user': mongodb_user}
- | Simplify the context processor. Not very useful anymore, in fact. | ## Code Before:
from .models.nonrel import User
def mongodb_user(request):
if request.user.is_anonymous():
return {u'mongodb_user': None}
try:
mongodb_user = User.objects.get(id=request.session[u'mongodb_user_id'])
except KeyError:
mongodb_user = User.objects.get(django_user=request.user.id)
# Cache it for next time.
request.session[u'mongodb_user_id'] = mongodb_user.id
return {u'mongodb_user': mongodb_user}
## Instruction:
Simplify the context processor. Not very useful anymore, in fact.
## Code After:
def mongodb_user(request):
""" not the most usefull context manager in the world. """
if request.user.is_anonymous():
return {u'mongodb_user': None}
return {u'mongodb_user': request.user.mongo}
| ...
def mongodb_user(request):
""" not the most usefull context manager in the world. """
if request.user.is_anonymous():
...
return {u'mongodb_user': None}
return {u'mongodb_user': request.user.mongo}
... |
f5cc3275a11c809bb6f5ab097414d0a5ccda2341 | main.py | main.py | def main():
website = input("Input website(cnn, nytimes, bbc, nzherald): ")
url = input("Input url: ")
scraper(website, url)
def scraper(website, url):
print("%s, %s" % (website, url))
if __name__ == '__main__':
main() | def main():
website = input("Input website(cnn, nytimes, bbc, nzherald): ")
url = input("Input url: ")
scraper(website, url)
def scraper(website, url):
if ".com" not in url:
print("Invalid url")
exit()
print("%s, %s" % (website, url))
if __name__ == '__main__':
main() | Check for .com in url | Check for .com in url
| Python | mit | Alex-Gurung/ScrapeTheNews | def main():
website = input("Input website(cnn, nytimes, bbc, nzherald): ")
url = input("Input url: ")
scraper(website, url)
def scraper(website, url):
+ if ".com" not in url:
+ print("Invalid url")
+ exit()
print("%s, %s" % (website, url))
if __name__ == '__main__':
main() | Check for .com in url | ## Code Before:
def main():
website = input("Input website(cnn, nytimes, bbc, nzherald): ")
url = input("Input url: ")
scraper(website, url)
def scraper(website, url):
print("%s, %s" % (website, url))
if __name__ == '__main__':
main()
## Instruction:
Check for .com in url
## Code After:
def main():
website = input("Input website(cnn, nytimes, bbc, nzherald): ")
url = input("Input url: ")
scraper(website, url)
def scraper(website, url):
if ".com" not in url:
print("Invalid url")
exit()
print("%s, %s" % (website, url))
if __name__ == '__main__':
main() | // ... existing code ...
scraper(website, url)
def scraper(website, url):
if ".com" not in url:
print("Invalid url")
exit()
print("%s, %s" % (website, url))
if __name__ == '__main__':
// ... rest of the code ... |
c496be720461722ce482c981b4915365dd0df8ab | events/views.py | events/views.py | from django.contrib import messages
from django.contrib.auth.decorators import login_required
from django.views.generic.list import ListView
from django.views.generic.detail import DetailView
from base.util import class_view_decorator
from base.views import RedirectBackView
from .models import Event, EventUserRegistration
class EventListView(ListView):
model = Event
context_object_name = 'events'
class EventDetailView(DetailView):
model = Event
context_object_name = 'event'
@class_view_decorator(login_required)
class EventUserRegisterView(RedirectBackView):
default_return_view = 'events_event_list'
def dispatch(self, request, *args, **kwargs):
event = Event.objects.get(pk=kwargs['event_id'])
if event.registration_open():
registration = EventUserRegistration(user=request.user, event=event)
registration.save()
message = 'Successfully registered to the %s' % event
messages.add_message(request, messages.INFO, message)
else:
message = 'Registration to the %s is not open.' % event
messages.add_message(request, messages.ERROR, message)
return super(EventUserRegisterView, self).dispatch(request,
*args, **kwargs)
| from django.contrib import messages
from django.contrib.auth.decorators import login_required
from django.utils.translation import ugettext_lazy as _
from django.views.generic.list import ListView
from django.views.generic.detail import DetailView
from base.util import class_view_decorator
from base.views import RedirectBackView
from .models import Event, EventUserRegistration
class EventListView(ListView):
model = Event
context_object_name = 'events'
class EventDetailView(DetailView):
model = Event
context_object_name = 'event'
@class_view_decorator(login_required)
class EventUserRegisterView(RedirectBackView):
default_return_view = 'events_event_list'
def dispatch(self, request, *args, **kwargs):
event = Event.objects.get(pk=kwargs['event_id'])
# Check if user is not already registered
registrations = EventUserRegistration.objects.filter(
user=request.user,
event=event).count()
if registrations:
message = _('You are already registered to the %s') % event
messages.add_message(request, messages.ERROR, message)
return super(EventUserRegisterView, self).dispatch(request,
*args,
**kwargs)
if event.registration_open():
registration = EventUserRegistration(user=request.user, event=event)
registration.save()
message = _('Successfully registered to the %s') % event
messages.add_message(request, messages.INFO, message)
else:
message = _('Registration to the %s is not open.') % event
messages.add_message(request, messages.ERROR, message)
return super(EventUserRegisterView, self).dispatch(request,
*args, **kwargs)
| Raise error when user is registering to the event multiple times | events: Raise error when user is registering to the event multiple times
| Python | mit | matus-stehlik/roots,rtrembecky/roots,tbabej/roots,rtrembecky/roots,matus-stehlik/roots,rtrembecky/roots,tbabej/roots,tbabej/roots,matus-stehlik/roots | from django.contrib import messages
from django.contrib.auth.decorators import login_required
+ from django.utils.translation import ugettext_lazy as _
from django.views.generic.list import ListView
from django.views.generic.detail import DetailView
from base.util import class_view_decorator
from base.views import RedirectBackView
from .models import Event, EventUserRegistration
class EventListView(ListView):
model = Event
context_object_name = 'events'
class EventDetailView(DetailView):
model = Event
context_object_name = 'event'
@class_view_decorator(login_required)
class EventUserRegisterView(RedirectBackView):
default_return_view = 'events_event_list'
def dispatch(self, request, *args, **kwargs):
event = Event.objects.get(pk=kwargs['event_id'])
+ # Check if user is not already registered
+ registrations = EventUserRegistration.objects.filter(
+ user=request.user,
+ event=event).count()
+
+ if registrations:
+ message = _('You are already registered to the %s') % event
+ messages.add_message(request, messages.ERROR, message)
+ return super(EventUserRegisterView, self).dispatch(request,
+ *args,
+ **kwargs)
+
if event.registration_open():
registration = EventUserRegistration(user=request.user, event=event)
registration.save()
- message = 'Successfully registered to the %s' % event
+ message = _('Successfully registered to the %s') % event
messages.add_message(request, messages.INFO, message)
else:
- message = 'Registration to the %s is not open.' % event
+ message = _('Registration to the %s is not open.') % event
messages.add_message(request, messages.ERROR, message)
return super(EventUserRegisterView, self).dispatch(request,
*args, **kwargs)
| Raise error when user is registering to the event multiple times | ## Code Before:
from django.contrib import messages
from django.contrib.auth.decorators import login_required
from django.views.generic.list import ListView
from django.views.generic.detail import DetailView
from base.util import class_view_decorator
from base.views import RedirectBackView
from .models import Event, EventUserRegistration
class EventListView(ListView):
model = Event
context_object_name = 'events'
class EventDetailView(DetailView):
model = Event
context_object_name = 'event'
@class_view_decorator(login_required)
class EventUserRegisterView(RedirectBackView):
default_return_view = 'events_event_list'
def dispatch(self, request, *args, **kwargs):
event = Event.objects.get(pk=kwargs['event_id'])
if event.registration_open():
registration = EventUserRegistration(user=request.user, event=event)
registration.save()
message = 'Successfully registered to the %s' % event
messages.add_message(request, messages.INFO, message)
else:
message = 'Registration to the %s is not open.' % event
messages.add_message(request, messages.ERROR, message)
return super(EventUserRegisterView, self).dispatch(request,
*args, **kwargs)
## Instruction:
Raise error when user is registering to the event multiple times
## Code After:
from django.contrib import messages
from django.contrib.auth.decorators import login_required
from django.utils.translation import ugettext_lazy as _
from django.views.generic.list import ListView
from django.views.generic.detail import DetailView
from base.util import class_view_decorator
from base.views import RedirectBackView
from .models import Event, EventUserRegistration
class EventListView(ListView):
model = Event
context_object_name = 'events'
class EventDetailView(DetailView):
model = Event
context_object_name = 'event'
@class_view_decorator(login_required)
class EventUserRegisterView(RedirectBackView):
default_return_view = 'events_event_list'
def dispatch(self, request, *args, **kwargs):
event = Event.objects.get(pk=kwargs['event_id'])
# Check if user is not already registered
registrations = EventUserRegistration.objects.filter(
user=request.user,
event=event).count()
if registrations:
message = _('You are already registered to the %s') % event
messages.add_message(request, messages.ERROR, message)
return super(EventUserRegisterView, self).dispatch(request,
*args,
**kwargs)
if event.registration_open():
registration = EventUserRegistration(user=request.user, event=event)
registration.save()
message = _('Successfully registered to the %s') % event
messages.add_message(request, messages.INFO, message)
else:
message = _('Registration to the %s is not open.') % event
messages.add_message(request, messages.ERROR, message)
return super(EventUserRegisterView, self).dispatch(request,
*args, **kwargs)
| ...
from django.contrib import messages
from django.contrib.auth.decorators import login_required
from django.utils.translation import ugettext_lazy as _
from django.views.generic.list import ListView
from django.views.generic.detail import DetailView
...
event = Event.objects.get(pk=kwargs['event_id'])
# Check if user is not already registered
registrations = EventUserRegistration.objects.filter(
user=request.user,
event=event).count()
if registrations:
message = _('You are already registered to the %s') % event
messages.add_message(request, messages.ERROR, message)
return super(EventUserRegisterView, self).dispatch(request,
*args,
**kwargs)
if event.registration_open():
registration = EventUserRegistration(user=request.user, event=event)
...
registration.save()
message = _('Successfully registered to the %s') % event
messages.add_message(request, messages.INFO, message)
else:
message = _('Registration to the %s is not open.') % event
messages.add_message(request, messages.ERROR, message)
... |
60743b33e5034776576073b151c7a02dc0a40b7e | tests/unit_project/test_fields.py | tests/unit_project/test_fields.py | from djangosanetesting.cases import DatabaseTestCase
from djangomarkup.fields import RichTextField
from djangomarkup.models import SourceText
from exampleapp.models import Article
class TestRichTextField(DatabaseTestCase):
def setUp(self):
super(TestRichTextField, self).setUp()
self.field = RichTextField(
instance = Article(),
model = Article,
syntax_processor_name = "markdown",
field_name = "text",
required = True,
label = "Text"
)
def test_retrieve_empty_source_for_empty_article(self):
self.assert_equals(u'', self.field.get_source().content)
def test_source_available_for_empty_article(self):
self.assert_equals(u'', self.field.get_source_text())
def test_render_available_for_empty_article(self):
self.assert_equals(u'<p></p>', self.field.get_rendered_text().strip())
| from djangosanetesting.cases import UnitTestCase
from djangomarkup.fields import RichTextField
from exampleapp.models import Article
class TestRichTextField(UnitTestCase):
def setUp(self):
super(TestRichTextField, self).setUp()
self.field = RichTextField(
instance = Article(),
model = Article,
syntax_processor_name = "markdown",
field_name = "text",
required = True,
label = "Text"
)
def test_retrieve_empty_source_for_empty_article(self):
self.assert_equals(u'', self.field.get_source().content)
def test_source_available_for_empty_article(self):
self.assert_equals(u'', self.field.get_source_text())
def test_render_available_for_empty_article(self):
self.assert_equals(u'<p></p>', self.field.get_rendered_text().strip())
def test_value_error_raised_when_accessing_source_without_instance(self):
field = RichTextField(
instance = None,
model = Article,
syntax_processor_name = "markdown",
field_name = "text",
required = True,
label = "Text"
)
self.assert_raises(ValueError, field.get_source) | Check proper error when accessing source without instance | Check proper error when accessing source without instance
| Python | bsd-3-clause | ella/django-markup | - from djangosanetesting.cases import DatabaseTestCase
+ from djangosanetesting.cases import UnitTestCase
from djangomarkup.fields import RichTextField
- from djangomarkup.models import SourceText
from exampleapp.models import Article
- class TestRichTextField(DatabaseTestCase):
+ class TestRichTextField(UnitTestCase):
def setUp(self):
super(TestRichTextField, self).setUp()
self.field = RichTextField(
instance = Article(),
model = Article,
syntax_processor_name = "markdown",
field_name = "text",
required = True,
label = "Text"
)
def test_retrieve_empty_source_for_empty_article(self):
self.assert_equals(u'', self.field.get_source().content)
def test_source_available_for_empty_article(self):
self.assert_equals(u'', self.field.get_source_text())
def test_render_available_for_empty_article(self):
self.assert_equals(u'<p></p>', self.field.get_rendered_text().strip())
+ def test_value_error_raised_when_accessing_source_without_instance(self):
+ field = RichTextField(
+ instance = None,
+ model = Article,
+ syntax_processor_name = "markdown",
+ field_name = "text",
+ required = True,
+ label = "Text"
+ )
+ self.assert_raises(ValueError, field.get_source) | Check proper error when accessing source without instance | ## Code Before:
from djangosanetesting.cases import DatabaseTestCase
from djangomarkup.fields import RichTextField
from djangomarkup.models import SourceText
from exampleapp.models import Article
class TestRichTextField(DatabaseTestCase):
def setUp(self):
super(TestRichTextField, self).setUp()
self.field = RichTextField(
instance = Article(),
model = Article,
syntax_processor_name = "markdown",
field_name = "text",
required = True,
label = "Text"
)
def test_retrieve_empty_source_for_empty_article(self):
self.assert_equals(u'', self.field.get_source().content)
def test_source_available_for_empty_article(self):
self.assert_equals(u'', self.field.get_source_text())
def test_render_available_for_empty_article(self):
self.assert_equals(u'<p></p>', self.field.get_rendered_text().strip())
## Instruction:
Check proper error when accessing source without instance
## Code After:
from djangosanetesting.cases import UnitTestCase
from djangomarkup.fields import RichTextField
from exampleapp.models import Article
class TestRichTextField(UnitTestCase):
def setUp(self):
super(TestRichTextField, self).setUp()
self.field = RichTextField(
instance = Article(),
model = Article,
syntax_processor_name = "markdown",
field_name = "text",
required = True,
label = "Text"
)
def test_retrieve_empty_source_for_empty_article(self):
self.assert_equals(u'', self.field.get_source().content)
def test_source_available_for_empty_article(self):
self.assert_equals(u'', self.field.get_source_text())
def test_render_available_for_empty_article(self):
self.assert_equals(u'<p></p>', self.field.get_rendered_text().strip())
def test_value_error_raised_when_accessing_source_without_instance(self):
field = RichTextField(
instance = None,
model = Article,
syntax_processor_name = "markdown",
field_name = "text",
required = True,
label = "Text"
)
self.assert_raises(ValueError, field.get_source) | # ... existing code ...
from djangosanetesting.cases import UnitTestCase
from djangomarkup.fields import RichTextField
from exampleapp.models import Article
class TestRichTextField(UnitTestCase):
def setUp(self):
# ... modified code ...
self.assert_equals(u'<p></p>', self.field.get_rendered_text().strip())
def test_value_error_raised_when_accessing_source_without_instance(self):
field = RichTextField(
instance = None,
model = Article,
syntax_processor_name = "markdown",
field_name = "text",
required = True,
label = "Text"
)
self.assert_raises(ValueError, field.get_source)
# ... rest of the code ... |
252cfa3baa7973a923952ecb3c83cdfb9f28ab67 | l10n_br_account/models/fiscal_document.py | l10n_br_account/models/fiscal_document.py |
from odoo import api, models
class FiscalDocument(models.Model):
_inherit = 'l10n_br_fiscal.document'
@api.multi
def unlink(self):
invoices = self.env['account.invoice'].search(
[('fiscal_document_id', 'in', self.ids)])
invoices.unlink()
return super().unlink()
|
from odoo import _, api, models
from odoo.exceptions import UserError
from odoo.addons.l10n_br_fiscal.constants.fiscal import (
SITUACAO_EDOC_EM_DIGITACAO,
)
class FiscalDocument(models.Model):
_inherit = 'l10n_br_fiscal.document'
@api.multi
def unlink(self):
draft_documents = self.filtered(
lambda d: d.state == SITUACAO_EDOC_EM_DIGITACAO)
if draft_documents:
UserError(_("You cannot delete a fiscal document "
"which is not draft state."))
invoices = self.env['account.invoice'].search(
[('fiscal_document_id', 'in', self.ids)])
invoices.unlink()
return super().unlink()
| Allow delete only fiscal documents with draft state | [REF] Allow delete only fiscal documents with draft state
| Python | agpl-3.0 | OCA/l10n-brazil,akretion/l10n-brazil,akretion/l10n-brazil,akretion/l10n-brazil,OCA/l10n-brazil,OCA/l10n-brazil |
- from odoo import api, models
+ from odoo import _, api, models
+ from odoo.exceptions import UserError
+
+ from odoo.addons.l10n_br_fiscal.constants.fiscal import (
+ SITUACAO_EDOC_EM_DIGITACAO,
+ )
class FiscalDocument(models.Model):
_inherit = 'l10n_br_fiscal.document'
@api.multi
def unlink(self):
+ draft_documents = self.filtered(
+ lambda d: d.state == SITUACAO_EDOC_EM_DIGITACAO)
+
+ if draft_documents:
+ UserError(_("You cannot delete a fiscal document "
+ "which is not draft state."))
+
invoices = self.env['account.invoice'].search(
[('fiscal_document_id', 'in', self.ids)])
invoices.unlink()
return super().unlink()
| Allow delete only fiscal documents with draft state | ## Code Before:
from odoo import api, models
class FiscalDocument(models.Model):
_inherit = 'l10n_br_fiscal.document'
@api.multi
def unlink(self):
invoices = self.env['account.invoice'].search(
[('fiscal_document_id', 'in', self.ids)])
invoices.unlink()
return super().unlink()
## Instruction:
Allow delete only fiscal documents with draft state
## Code After:
from odoo import _, api, models
from odoo.exceptions import UserError
from odoo.addons.l10n_br_fiscal.constants.fiscal import (
SITUACAO_EDOC_EM_DIGITACAO,
)
class FiscalDocument(models.Model):
_inherit = 'l10n_br_fiscal.document'
@api.multi
def unlink(self):
draft_documents = self.filtered(
lambda d: d.state == SITUACAO_EDOC_EM_DIGITACAO)
if draft_documents:
UserError(_("You cannot delete a fiscal document "
"which is not draft state."))
invoices = self.env['account.invoice'].search(
[('fiscal_document_id', 'in', self.ids)])
invoices.unlink()
return super().unlink()
| // ... existing code ...
from odoo import _, api, models
from odoo.exceptions import UserError
from odoo.addons.l10n_br_fiscal.constants.fiscal import (
SITUACAO_EDOC_EM_DIGITACAO,
)
// ... modified code ...
@api.multi
def unlink(self):
draft_documents = self.filtered(
lambda d: d.state == SITUACAO_EDOC_EM_DIGITACAO)
if draft_documents:
UserError(_("You cannot delete a fiscal document "
"which is not draft state."))
invoices = self.env['account.invoice'].search(
[('fiscal_document_id', 'in', self.ids)])
// ... rest of the code ... |
1ed7d695eff134557990d8b1a5dffa51b6d1d2f6 | distarray/run_tests.py | distarray/run_tests.py |
from __future__ import print_function
import os
import sys
import shlex
import subprocess
import distarray
def _run_shell_command(specific_cmd):
"""Run a command with subprocess and pass the results through to stdout.
First, change directory to the project directory.
"""
path = os.path.split(os.path.split(distarray.__file__)[0])[0]
os.chdir(path)
proc = subprocess.Popen(shlex.split(specific_cmd),
stdout=subprocess.PIPE,
stderr=subprocess.STDOUT)
while True:
char = proc.stdout.read(1).decode()
if not char:
break
else:
print(char, end="")
sys.stdout.flush()
def test():
"""Run all DistArray tests."""
cmd = "make test"
_run_shell_command(cmd)
if __name__ == "__main__":
test()
|
from __future__ import print_function
import os
import sys
import shlex
import subprocess
import distarray
def _run_shell_command(specific_cmd):
"""Run a command with subprocess and pass the results through to stdout.
First, change directory to the project directory.
"""
path = os.path.split(os.path.split(distarray.__file__)[0])[0]
os.chdir(path)
proc = subprocess.Popen(shlex.split(specific_cmd),
stdout=subprocess.PIPE,
stderr=subprocess.STDOUT)
while True:
char = proc.stdout.read(1).decode()
if not char:
return proc.wait()
else:
print(char, end="")
sys.stdout.flush()
def test():
"""Run all DistArray tests."""
cmd = "make test"
return _run_shell_command(cmd)
if __name__ == "__main__":
sys.exit(test())
| Return returncode from shell command. | Return returncode from shell command. | Python | bsd-3-clause | enthought/distarray,RaoUmer/distarray,RaoUmer/distarray,enthought/distarray |
from __future__ import print_function
import os
import sys
import shlex
import subprocess
import distarray
def _run_shell_command(specific_cmd):
"""Run a command with subprocess and pass the results through to stdout.
First, change directory to the project directory.
"""
path = os.path.split(os.path.split(distarray.__file__)[0])[0]
os.chdir(path)
proc = subprocess.Popen(shlex.split(specific_cmd),
stdout=subprocess.PIPE,
stderr=subprocess.STDOUT)
while True:
char = proc.stdout.read(1).decode()
if not char:
- break
+ return proc.wait()
else:
print(char, end="")
sys.stdout.flush()
def test():
"""Run all DistArray tests."""
cmd = "make test"
- _run_shell_command(cmd)
+ return _run_shell_command(cmd)
if __name__ == "__main__":
- test()
+ sys.exit(test())
| Return returncode from shell command. | ## Code Before:
from __future__ import print_function
import os
import sys
import shlex
import subprocess
import distarray
def _run_shell_command(specific_cmd):
"""Run a command with subprocess and pass the results through to stdout.
First, change directory to the project directory.
"""
path = os.path.split(os.path.split(distarray.__file__)[0])[0]
os.chdir(path)
proc = subprocess.Popen(shlex.split(specific_cmd),
stdout=subprocess.PIPE,
stderr=subprocess.STDOUT)
while True:
char = proc.stdout.read(1).decode()
if not char:
break
else:
print(char, end="")
sys.stdout.flush()
def test():
"""Run all DistArray tests."""
cmd = "make test"
_run_shell_command(cmd)
if __name__ == "__main__":
test()
## Instruction:
Return returncode from shell command.
## Code After:
from __future__ import print_function
import os
import sys
import shlex
import subprocess
import distarray
def _run_shell_command(specific_cmd):
"""Run a command with subprocess and pass the results through to stdout.
First, change directory to the project directory.
"""
path = os.path.split(os.path.split(distarray.__file__)[0])[0]
os.chdir(path)
proc = subprocess.Popen(shlex.split(specific_cmd),
stdout=subprocess.PIPE,
stderr=subprocess.STDOUT)
while True:
char = proc.stdout.read(1).decode()
if not char:
return proc.wait()
else:
print(char, end="")
sys.stdout.flush()
def test():
"""Run all DistArray tests."""
cmd = "make test"
return _run_shell_command(cmd)
if __name__ == "__main__":
sys.exit(test())
| # ... existing code ...
char = proc.stdout.read(1).decode()
if not char:
return proc.wait()
else:
print(char, end="")
# ... modified code ...
"""Run all DistArray tests."""
cmd = "make test"
return _run_shell_command(cmd)
if __name__ == "__main__":
sys.exit(test())
# ... rest of the code ... |
73aa38a5d481a26278dd29364f16839cad0f22cf | manager/projects/ui/views/files.py | manager/projects/ui/views/files.py | from django.contrib.auth.decorators import login_required
from django.http import HttpRequest, HttpResponse
from django.shortcuts import render
from projects.api.views.files import ProjectsFilesViewSet
@login_required
def list(request: HttpRequest, *args, **kwargs) -> HttpResponse:
"""
Get a list of project files.
The trailing part of the URL becomes the `prefix` query
parameter, consistent with API ending e.g.
/<account>/<project>/files/sub?search=foo
is equivalent to:
/api/projects/<project>/files?prefix=sub&search=foo
"""
prefix = kwargs.get("prefix")
if prefix and not prefix.endswith("/"):
prefix += "/"
request.GET = request.GET.copy()
request.GET["prefix"] = prefix
request.GET["aggregate"] = True
viewset = ProjectsFilesViewSet.init("list", request, args, kwargs)
project = viewset.get_project()
files = viewset.get_queryset(project)
# List of tuples for directory breadcrumbs
dirs = [("root", "")]
path = ""
for name in prefix.split("/"):
if name:
path += name + "/"
dirs.append((name, path))
return render(
request,
"projects/files/list.html",
dict(prefix=prefix, dirs=dirs, files=files, project=project,),
)
| from django.contrib.auth.decorators import login_required
from django.http import HttpRequest, HttpResponse
from django.shortcuts import render
from projects.api.views.files import ProjectsFilesViewSet
@login_required
def list(request: HttpRequest, *args, **kwargs) -> HttpResponse:
"""
Get a list of project files.
The trailing part of the URL becomes the `prefix` query
parameter, consistent with API ending e.g.
/<account>/<project>/files/sub?search=foo
is equivalent to:
/api/projects/<project>/files?prefix=sub&search=foo
"""
request.GET = request.GET.copy()
request.GET["prefix"] = kwargs.get("prefix")
viewset = ProjectsFilesViewSet.init("list", request, args, kwargs)
project = viewset.get_project()
files = viewset.get_queryset(project)
context = viewset.get_response_context(queryset=files)
return render(
request, "projects/files/list.html", dict(project=project, **context),
)
| Update view for change in viewset | refactor(Files): Update view for change in viewset
| Python | apache-2.0 | stencila/hub,stencila/hub,stencila/hub,stencila/hub,stencila/hub | from django.contrib.auth.decorators import login_required
from django.http import HttpRequest, HttpResponse
from django.shortcuts import render
from projects.api.views.files import ProjectsFilesViewSet
@login_required
def list(request: HttpRequest, *args, **kwargs) -> HttpResponse:
"""
Get a list of project files.
The trailing part of the URL becomes the `prefix` query
parameter, consistent with API ending e.g.
/<account>/<project>/files/sub?search=foo
is equivalent to:
/api/projects/<project>/files?prefix=sub&search=foo
"""
- prefix = kwargs.get("prefix")
- if prefix and not prefix.endswith("/"):
- prefix += "/"
-
request.GET = request.GET.copy()
- request.GET["prefix"] = prefix
+ request.GET["prefix"] = kwargs.get("prefix")
- request.GET["aggregate"] = True
viewset = ProjectsFilesViewSet.init("list", request, args, kwargs)
project = viewset.get_project()
files = viewset.get_queryset(project)
+ context = viewset.get_response_context(queryset=files)
-
- # List of tuples for directory breadcrumbs
- dirs = [("root", "")]
- path = ""
- for name in prefix.split("/"):
- if name:
- path += name + "/"
- dirs.append((name, path))
return render(
+ request, "projects/files/list.html", dict(project=project, **context),
- request,
- "projects/files/list.html",
- dict(prefix=prefix, dirs=dirs, files=files, project=project,),
)
| Update view for change in viewset | ## Code Before:
from django.contrib.auth.decorators import login_required
from django.http import HttpRequest, HttpResponse
from django.shortcuts import render
from projects.api.views.files import ProjectsFilesViewSet
@login_required
def list(request: HttpRequest, *args, **kwargs) -> HttpResponse:
"""
Get a list of project files.
The trailing part of the URL becomes the `prefix` query
parameter, consistent with API ending e.g.
/<account>/<project>/files/sub?search=foo
is equivalent to:
/api/projects/<project>/files?prefix=sub&search=foo
"""
prefix = kwargs.get("prefix")
if prefix and not prefix.endswith("/"):
prefix += "/"
request.GET = request.GET.copy()
request.GET["prefix"] = prefix
request.GET["aggregate"] = True
viewset = ProjectsFilesViewSet.init("list", request, args, kwargs)
project = viewset.get_project()
files = viewset.get_queryset(project)
# List of tuples for directory breadcrumbs
dirs = [("root", "")]
path = ""
for name in prefix.split("/"):
if name:
path += name + "/"
dirs.append((name, path))
return render(
request,
"projects/files/list.html",
dict(prefix=prefix, dirs=dirs, files=files, project=project,),
)
## Instruction:
Update view for change in viewset
## Code After:
from django.contrib.auth.decorators import login_required
from django.http import HttpRequest, HttpResponse
from django.shortcuts import render
from projects.api.views.files import ProjectsFilesViewSet
@login_required
def list(request: HttpRequest, *args, **kwargs) -> HttpResponse:
"""
Get a list of project files.
The trailing part of the URL becomes the `prefix` query
parameter, consistent with API ending e.g.
/<account>/<project>/files/sub?search=foo
is equivalent to:
/api/projects/<project>/files?prefix=sub&search=foo
"""
request.GET = request.GET.copy()
request.GET["prefix"] = kwargs.get("prefix")
viewset = ProjectsFilesViewSet.init("list", request, args, kwargs)
project = viewset.get_project()
files = viewset.get_queryset(project)
context = viewset.get_response_context(queryset=files)
return render(
request, "projects/files/list.html", dict(project=project, **context),
)
| // ... existing code ...
/api/projects/<project>/files?prefix=sub&search=foo
"""
request.GET = request.GET.copy()
request.GET["prefix"] = kwargs.get("prefix")
viewset = ProjectsFilesViewSet.init("list", request, args, kwargs)
// ... modified code ...
project = viewset.get_project()
files = viewset.get_queryset(project)
context = viewset.get_response_context(queryset=files)
return render(
request, "projects/files/list.html", dict(project=project, **context),
)
// ... rest of the code ... |
154b64b2ee56fa4391251268ba4a85d178bedd60 | djangoautoconf/urls.py | djangoautoconf/urls.py | from django.conf.urls import patterns, include, url
from django.conf import settings
from django.conf.urls.static import static
# Uncomment the next two lines to enable the admin:
from django.contrib import admin
# from mezzanine.core.views import direct_to_template
admin.autodiscover()
# Must be defined before auto discover and urlpatterns var. So when there is root url
# injection, we first insert root url to this, then the last line will insert it to real urlpatterns
default_app_url_patterns = []
from djangoautoconf import auto_conf_urls
auto_conf_urls.autodiscover()
urlpatterns = patterns('',
# Examples:
# url(r'^default_django_15_and_below/', include('default_django_15_and_below.foo.urls')),
# Uncomment the admin/doc line below to enable admin documentation:
url(r'^admin/doc/', include('django.contrib.admindocs.urls')),
# Uncomment the next line to enable the admin:
url(r'^admin/', include(admin.site.urls)),
# url(r'^', include('demo.urls')),
# url(r'^obj_sys/', include('obj_sys.urls')),
# url("^$", direct_to_template, {"template": "index.html"}, name="home"),
)
urlpatterns = [
# ... the rest of your URLconf goes here ...
] + static(settings.MEDIA_URL, document_root=settings.MEDIA_ROOT)
urlpatterns += default_app_url_patterns
| from django.conf.urls import patterns, include, url
from django.conf import settings
from django.conf.urls.static import static
# Uncomment the next two lines to enable the admin:
from django.contrib import admin
# from mezzanine.core.views import direct_to_template
admin.autodiscover()
# Must be defined before auto discover and urlpatterns var. So when there is root url
# injection, we first insert root url to this, then the last line will insert it to real urlpatterns
default_app_url_patterns = []
from djangoautoconf import auto_conf_urls
auto_conf_urls.autodiscover()
urlpatterns = patterns('',
# Examples:
# url(r'^default_django_15_and_below/', include('default_django_15_and_below.foo.urls')),
# Uncomment the admin/doc line below to enable admin documentation:
url(r'^admin/doc/', include('django.contrib.admindocs.urls')),
# Uncomment the next line to enable the admin:
url(r'^admin/', include(admin.site.urls)),
# url(r'^', include('demo.urls')),
# url(r'^obj_sys/', include('obj_sys.urls')),
# url("^$", direct_to_template, {"template": "index.html"}, name="home"),
)
urlpatterns += static(settings.MEDIA_URL, document_root=settings.MEDIA_ROOT)
urlpatterns += default_app_url_patterns
| Fix the issue of override url by mistake. | Fix the issue of override url by mistake.
| Python | bsd-3-clause | weijia/djangoautoconf,weijia/djangoautoconf | from django.conf.urls import patterns, include, url
from django.conf import settings
from django.conf.urls.static import static
# Uncomment the next two lines to enable the admin:
from django.contrib import admin
# from mezzanine.core.views import direct_to_template
admin.autodiscover()
# Must be defined before auto discover and urlpatterns var. So when there is root url
# injection, we first insert root url to this, then the last line will insert it to real urlpatterns
default_app_url_patterns = []
from djangoautoconf import auto_conf_urls
auto_conf_urls.autodiscover()
urlpatterns = patterns('',
# Examples:
# url(r'^default_django_15_and_below/', include('default_django_15_and_below.foo.urls')),
# Uncomment the admin/doc line below to enable admin documentation:
url(r'^admin/doc/', include('django.contrib.admindocs.urls')),
# Uncomment the next line to enable the admin:
url(r'^admin/', include(admin.site.urls)),
# url(r'^', include('demo.urls')),
# url(r'^obj_sys/', include('obj_sys.urls')),
# url("^$", direct_to_template, {"template": "index.html"}, name="home"),
)
- urlpatterns = [
- # ... the rest of your URLconf goes here ...
- ] + static(settings.MEDIA_URL, document_root=settings.MEDIA_ROOT)
+ urlpatterns += static(settings.MEDIA_URL, document_root=settings.MEDIA_ROOT)
urlpatterns += default_app_url_patterns
| Fix the issue of override url by mistake. | ## Code Before:
from django.conf.urls import patterns, include, url
from django.conf import settings
from django.conf.urls.static import static
# Uncomment the next two lines to enable the admin:
from django.contrib import admin
# from mezzanine.core.views import direct_to_template
admin.autodiscover()
# Must be defined before auto discover and urlpatterns var. So when there is root url
# injection, we first insert root url to this, then the last line will insert it to real urlpatterns
default_app_url_patterns = []
from djangoautoconf import auto_conf_urls
auto_conf_urls.autodiscover()
urlpatterns = patterns('',
# Examples:
# url(r'^default_django_15_and_below/', include('default_django_15_and_below.foo.urls')),
# Uncomment the admin/doc line below to enable admin documentation:
url(r'^admin/doc/', include('django.contrib.admindocs.urls')),
# Uncomment the next line to enable the admin:
url(r'^admin/', include(admin.site.urls)),
# url(r'^', include('demo.urls')),
# url(r'^obj_sys/', include('obj_sys.urls')),
# url("^$", direct_to_template, {"template": "index.html"}, name="home"),
)
urlpatterns = [
# ... the rest of your URLconf goes here ...
] + static(settings.MEDIA_URL, document_root=settings.MEDIA_ROOT)
urlpatterns += default_app_url_patterns
## Instruction:
Fix the issue of override url by mistake.
## Code After:
from django.conf.urls import patterns, include, url
from django.conf import settings
from django.conf.urls.static import static
# Uncomment the next two lines to enable the admin:
from django.contrib import admin
# from mezzanine.core.views import direct_to_template
admin.autodiscover()
# Must be defined before auto discover and urlpatterns var. So when there is root url
# injection, we first insert root url to this, then the last line will insert it to real urlpatterns
default_app_url_patterns = []
from djangoautoconf import auto_conf_urls
auto_conf_urls.autodiscover()
urlpatterns = patterns('',
# Examples:
# url(r'^default_django_15_and_below/', include('default_django_15_and_below.foo.urls')),
# Uncomment the admin/doc line below to enable admin documentation:
url(r'^admin/doc/', include('django.contrib.admindocs.urls')),
# Uncomment the next line to enable the admin:
url(r'^admin/', include(admin.site.urls)),
# url(r'^', include('demo.urls')),
# url(r'^obj_sys/', include('obj_sys.urls')),
# url("^$", direct_to_template, {"template": "index.html"}, name="home"),
)
urlpatterns += static(settings.MEDIA_URL, document_root=settings.MEDIA_ROOT)
urlpatterns += default_app_url_patterns
| ...
# url("^$", direct_to_template, {"template": "index.html"}, name="home"),
)
urlpatterns += static(settings.MEDIA_URL, document_root=settings.MEDIA_ROOT)
urlpatterns += default_app_url_patterns
... |
0213bbb8f8075b2dc36a33380a66932c9d541f63 | src/sphobjinv/__init__.py | src/sphobjinv/__init__.py |
from sphobjinv.data import DataFields, DataObjBytes, DataObjStr
from sphobjinv.enum import HeaderFields, SourceTypes
from sphobjinv.error import (
SphobjinvError,
VersionError,
)
from sphobjinv.fileops import readbytes, readjson, urlwalk, writebytes, writejson
from sphobjinv.inventory import Inventory
from sphobjinv.re import p_data, pb_comments, pb_data, pb_project, pb_version
from sphobjinv.schema import json_schema
from sphobjinv.version import __version__
from sphobjinv.zlib import compress, decompress
|
from sphobjinv.data import DataFields, DataObjBytes, DataObjStr
from sphobjinv.enum import HeaderFields, SourceTypes
from sphobjinv.error import SphobjinvError, VersionError
from sphobjinv.fileops import readbytes, readjson, urlwalk, writebytes, writejson
from sphobjinv.inventory import Inventory
from sphobjinv.re import p_data, pb_comments, pb_data, pb_project, pb_version
from sphobjinv.schema import json_schema
from sphobjinv.version import __version__
from sphobjinv.zlib import compress, decompress
| Clean up the error imports | Clean up the error imports
The new errors that had been added for _intersphinx.py had left
the sphobjinv.error import line split. No need, when it all fits on
one line.
| Python | mit | bskinn/sphobjinv |
from sphobjinv.data import DataFields, DataObjBytes, DataObjStr
from sphobjinv.enum import HeaderFields, SourceTypes
+ from sphobjinv.error import SphobjinvError, VersionError
- from sphobjinv.error import (
- SphobjinvError,
- VersionError,
- )
from sphobjinv.fileops import readbytes, readjson, urlwalk, writebytes, writejson
from sphobjinv.inventory import Inventory
from sphobjinv.re import p_data, pb_comments, pb_data, pb_project, pb_version
from sphobjinv.schema import json_schema
from sphobjinv.version import __version__
from sphobjinv.zlib import compress, decompress
| Clean up the error imports | ## Code Before:
from sphobjinv.data import DataFields, DataObjBytes, DataObjStr
from sphobjinv.enum import HeaderFields, SourceTypes
from sphobjinv.error import (
SphobjinvError,
VersionError,
)
from sphobjinv.fileops import readbytes, readjson, urlwalk, writebytes, writejson
from sphobjinv.inventory import Inventory
from sphobjinv.re import p_data, pb_comments, pb_data, pb_project, pb_version
from sphobjinv.schema import json_schema
from sphobjinv.version import __version__
from sphobjinv.zlib import compress, decompress
## Instruction:
Clean up the error imports
## Code After:
from sphobjinv.data import DataFields, DataObjBytes, DataObjStr
from sphobjinv.enum import HeaderFields, SourceTypes
from sphobjinv.error import SphobjinvError, VersionError
from sphobjinv.fileops import readbytes, readjson, urlwalk, writebytes, writejson
from sphobjinv.inventory import Inventory
from sphobjinv.re import p_data, pb_comments, pb_data, pb_project, pb_version
from sphobjinv.schema import json_schema
from sphobjinv.version import __version__
from sphobjinv.zlib import compress, decompress
| ...
from sphobjinv.data import DataFields, DataObjBytes, DataObjStr
from sphobjinv.enum import HeaderFields, SourceTypes
from sphobjinv.error import SphobjinvError, VersionError
from sphobjinv.fileops import readbytes, readjson, urlwalk, writebytes, writejson
from sphobjinv.inventory import Inventory
... |
785fcdca3c9bfb908444d3b9339457c616761f2c | tests/flights_to_test.py | tests/flights_to_test.py | import unittest
import datetime
import json
import sys
sys.path.append('..')
import sabre_dev_studio
import sabre_dev_studio.sabre_exceptions as sabre_exceptions
'''
requires config.json in the same directory for api authentication
{
"sabre_client_id": -----,
"sabre_client_secret": -----
}
'''
class TestBasicInstaflights(unittest.TestCase):
def read_config(self):
raw_data = open('config.json').read()
data = json.loads(raw_data)
client_secret = data['sabre_client_secret']
client_id = data['sabre_client_id']
return (client_id, client_secret)
def setUp(self):
# Read from config
self.client_id, self.client_secret = self.read_config()
self.sds = sabre_dev_studio.SabreDevStudio()
self.sds.set_credentials(self.client_id, self.client_secret)
self.sds.authenticate()
def test_basic_request(self):
city = 'YTO'
instaf = self.sds.flights_to(city)
self.assertIsNotNone(instaf)
def test_no_authorization(self):
sds = sabre_dev_studio.SabreDevStudio()
with self.assertRaises(sabre_exceptions.NotAuthorizedError):
resp = sds.flights_to('YTO')
if __name__ == '__main__':
unittest.main()
| import unittest
import datetime
import json
import sys
sys.path.append('..')
import sabre_dev_studio
import sabre_dev_studio.sabre_exceptions as sabre_exceptions
'''
requires config.json in the same directory for api authentication
{
"sabre_client_id": -----,
"sabre_client_secret": -----
}
'''
class TestBasicFlightsTo(unittest.TestCase):
def read_config(self):
raw_data = open('config.json').read()
data = json.loads(raw_data)
client_secret = data['sabre_client_secret']
client_id = data['sabre_client_id']
return (client_id, client_secret)
def setUp(self):
# Read from config
self.client_id, self.client_secret = self.read_config()
self.sds = sabre_dev_studio.SabreDevStudio()
self.sds.set_credentials(self.client_id, self.client_secret)
self.sds.authenticate()
def test_basic_request(self):
city = 'YTO'
flights_to_city = self.sds.flights_to(city)
print(flights_to_city)
self.assertIsNotNone(flights_to_city)
def test_no_authorization(self):
sds = sabre_dev_studio.SabreDevStudio()
with self.assertRaises(sabre_exceptions.NotAuthorizedError):
resp = sds.flights_to('YTO')
if __name__ == '__main__':
unittest.main()
| Change instaflights name in flights_to tests | Change instaflights name in flights_to tests
| Python | mit | Jamil/sabre_dev_studio | import unittest
import datetime
import json
import sys
sys.path.append('..')
import sabre_dev_studio
import sabre_dev_studio.sabre_exceptions as sabre_exceptions
'''
requires config.json in the same directory for api authentication
{
"sabre_client_id": -----,
"sabre_client_secret": -----
}
'''
- class TestBasicInstaflights(unittest.TestCase):
+ class TestBasicFlightsTo(unittest.TestCase):
def read_config(self):
raw_data = open('config.json').read()
data = json.loads(raw_data)
client_secret = data['sabre_client_secret']
client_id = data['sabre_client_id']
return (client_id, client_secret)
def setUp(self):
# Read from config
self.client_id, self.client_secret = self.read_config()
self.sds = sabre_dev_studio.SabreDevStudio()
self.sds.set_credentials(self.client_id, self.client_secret)
self.sds.authenticate()
def test_basic_request(self):
city = 'YTO'
- instaf = self.sds.flights_to(city)
+ flights_to_city = self.sds.flights_to(city)
+ print(flights_to_city)
- self.assertIsNotNone(instaf)
+ self.assertIsNotNone(flights_to_city)
def test_no_authorization(self):
sds = sabre_dev_studio.SabreDevStudio()
with self.assertRaises(sabre_exceptions.NotAuthorizedError):
resp = sds.flights_to('YTO')
if __name__ == '__main__':
unittest.main()
| Change instaflights name in flights_to tests | ## Code Before:
import unittest
import datetime
import json
import sys
sys.path.append('..')
import sabre_dev_studio
import sabre_dev_studio.sabre_exceptions as sabre_exceptions
'''
requires config.json in the same directory for api authentication
{
"sabre_client_id": -----,
"sabre_client_secret": -----
}
'''
class TestBasicInstaflights(unittest.TestCase):
def read_config(self):
raw_data = open('config.json').read()
data = json.loads(raw_data)
client_secret = data['sabre_client_secret']
client_id = data['sabre_client_id']
return (client_id, client_secret)
def setUp(self):
# Read from config
self.client_id, self.client_secret = self.read_config()
self.sds = sabre_dev_studio.SabreDevStudio()
self.sds.set_credentials(self.client_id, self.client_secret)
self.sds.authenticate()
def test_basic_request(self):
city = 'YTO'
instaf = self.sds.flights_to(city)
self.assertIsNotNone(instaf)
def test_no_authorization(self):
sds = sabre_dev_studio.SabreDevStudio()
with self.assertRaises(sabre_exceptions.NotAuthorizedError):
resp = sds.flights_to('YTO')
if __name__ == '__main__':
unittest.main()
## Instruction:
Change instaflights name in flights_to tests
## Code After:
import unittest
import datetime
import json
import sys
sys.path.append('..')
import sabre_dev_studio
import sabre_dev_studio.sabre_exceptions as sabre_exceptions
'''
requires config.json in the same directory for api authentication
{
"sabre_client_id": -----,
"sabre_client_secret": -----
}
'''
class TestBasicFlightsTo(unittest.TestCase):
def read_config(self):
raw_data = open('config.json').read()
data = json.loads(raw_data)
client_secret = data['sabre_client_secret']
client_id = data['sabre_client_id']
return (client_id, client_secret)
def setUp(self):
# Read from config
self.client_id, self.client_secret = self.read_config()
self.sds = sabre_dev_studio.SabreDevStudio()
self.sds.set_credentials(self.client_id, self.client_secret)
self.sds.authenticate()
def test_basic_request(self):
city = 'YTO'
flights_to_city = self.sds.flights_to(city)
print(flights_to_city)
self.assertIsNotNone(flights_to_city)
def test_no_authorization(self):
sds = sabre_dev_studio.SabreDevStudio()
with self.assertRaises(sabre_exceptions.NotAuthorizedError):
resp = sds.flights_to('YTO')
if __name__ == '__main__':
unittest.main()
| ...
'''
class TestBasicFlightsTo(unittest.TestCase):
def read_config(self):
raw_data = open('config.json').read()
...
def test_basic_request(self):
city = 'YTO'
flights_to_city = self.sds.flights_to(city)
print(flights_to_city)
self.assertIsNotNone(flights_to_city)
def test_no_authorization(self):
... |
a24b2b303c1cd5e9f43353d55cc6b9d07b37b7f4 | ephemeral-cluster.py | ephemeral-cluster.py | import subprocess
import sys
import uuid
usage = """\
Run a command using a temporary docker-compose cluster, removing all containers \
and images after command completion (regardless of success or failure.)
Generally, this would be used with the ``run`` command to provide a clean room \
testing environment.
"""
if not sys.argv[1:]:
sys.stderr.write(usage)
sys.exit(1)
project = uuid.uuid1().hex
sys.stderr.write('Starting ephemeral cluster: {0}\n'.format(project))
try:
sys.exit(subprocess.check_call(['docker-compose', '-p', project] + sys.argv[1:]))
finally:
subprocess.check_call(['docker-compose', '-p', project, 'stop'])
subprocess.check_call(['docker-compose', '-p', project, 'rm', '-f', '-v'])
| import subprocess
import sys
import uuid
usage = """\
Run a command using a temporary docker-compose cluster, removing all containers \
and associated volumes after command completion (regardless of success or \
failure.)
Generally, this would be used with the ``run`` command to provide a clean room \
testing environment.
"""
if not sys.argv[1:]:
sys.stderr.write(usage)
sys.exit(1)
project = uuid.uuid1().hex
sys.stderr.write('Setting up ephemeral cluster ({0})...\n'.format(project))
try:
subprocess.check_call(['docker-compose', '-p', project] + sys.argv[1:])
except subprocess.CalledProcessError as error:
raise SystemExit(error.returncode)
finally:
sys.stderr.write('\nCleaning up ephemeral cluster ({0})...\n'.format(project))
subprocess.check_call(['docker-compose', '-p', project, 'stop'])
subprocess.check_call(['docker-compose', '-p', project, 'rm', '-f', '-v'])
| Fix forwarding ephemeral cluster exit code. | Fix forwarding ephemeral cluster exit code.
Summary: Also improves logging a little bit.
Test Plan:
$ python ephemeral-cluster.py run --rm --entrypoint=bash pgshovel -c "exit 10"
$ test $? -eq 10
Reviewers: jeff, tail
Reviewed By: tail
Differential Revision: http://phabricator.local.disqus.net/D19564
| Python | apache-2.0 | fuziontech/pgshovel,disqus/pgshovel,fuziontech/pgshovel,fuziontech/pgshovel,disqus/pgshovel | import subprocess
import sys
import uuid
usage = """\
Run a command using a temporary docker-compose cluster, removing all containers \
- and images after command completion (regardless of success or failure.)
+ and associated volumes after command completion (regardless of success or \
+ failure.)
Generally, this would be used with the ``run`` command to provide a clean room \
testing environment.
"""
if not sys.argv[1:]:
sys.stderr.write(usage)
sys.exit(1)
project = uuid.uuid1().hex
- sys.stderr.write('Starting ephemeral cluster: {0}\n'.format(project))
+ sys.stderr.write('Setting up ephemeral cluster ({0})...\n'.format(project))
try:
- sys.exit(subprocess.check_call(['docker-compose', '-p', project] + sys.argv[1:]))
+ subprocess.check_call(['docker-compose', '-p', project] + sys.argv[1:])
+ except subprocess.CalledProcessError as error:
+ raise SystemExit(error.returncode)
finally:
+ sys.stderr.write('\nCleaning up ephemeral cluster ({0})...\n'.format(project))
subprocess.check_call(['docker-compose', '-p', project, 'stop'])
subprocess.check_call(['docker-compose', '-p', project, 'rm', '-f', '-v'])
| Fix forwarding ephemeral cluster exit code. | ## Code Before:
import subprocess
import sys
import uuid
usage = """\
Run a command using a temporary docker-compose cluster, removing all containers \
and images after command completion (regardless of success or failure.)
Generally, this would be used with the ``run`` command to provide a clean room \
testing environment.
"""
if not sys.argv[1:]:
sys.stderr.write(usage)
sys.exit(1)
project = uuid.uuid1().hex
sys.stderr.write('Starting ephemeral cluster: {0}\n'.format(project))
try:
sys.exit(subprocess.check_call(['docker-compose', '-p', project] + sys.argv[1:]))
finally:
subprocess.check_call(['docker-compose', '-p', project, 'stop'])
subprocess.check_call(['docker-compose', '-p', project, 'rm', '-f', '-v'])
## Instruction:
Fix forwarding ephemeral cluster exit code.
## Code After:
import subprocess
import sys
import uuid
usage = """\
Run a command using a temporary docker-compose cluster, removing all containers \
and associated volumes after command completion (regardless of success or \
failure.)
Generally, this would be used with the ``run`` command to provide a clean room \
testing environment.
"""
if not sys.argv[1:]:
sys.stderr.write(usage)
sys.exit(1)
project = uuid.uuid1().hex
sys.stderr.write('Setting up ephemeral cluster ({0})...\n'.format(project))
try:
subprocess.check_call(['docker-compose', '-p', project] + sys.argv[1:])
except subprocess.CalledProcessError as error:
raise SystemExit(error.returncode)
finally:
sys.stderr.write('\nCleaning up ephemeral cluster ({0})...\n'.format(project))
subprocess.check_call(['docker-compose', '-p', project, 'stop'])
subprocess.check_call(['docker-compose', '-p', project, 'rm', '-f', '-v'])
| # ... existing code ...
usage = """\
Run a command using a temporary docker-compose cluster, removing all containers \
and associated volumes after command completion (regardless of success or \
failure.)
Generally, this would be used with the ``run`` command to provide a clean room \
# ... modified code ...
project = uuid.uuid1().hex
sys.stderr.write('Setting up ephemeral cluster ({0})...\n'.format(project))
try:
subprocess.check_call(['docker-compose', '-p', project] + sys.argv[1:])
except subprocess.CalledProcessError as error:
raise SystemExit(error.returncode)
finally:
sys.stderr.write('\nCleaning up ephemeral cluster ({0})...\n'.format(project))
subprocess.check_call(['docker-compose', '-p', project, 'stop'])
subprocess.check_call(['docker-compose', '-p', project, 'rm', '-f', '-v'])
# ... rest of the code ... |
a3c52c84da93c3e3007fa291213b97fd7d5b0e8f | tests.py | tests.py |
import TwitterSA
import unittest
class TwitterSATestCase(unittest.TestCase):
def setUp(self):
TwitterSA.app.config['TESTING'] = True
self.app = TwitterSA.app.test_client()
def tearDown(self):
pass
def test_invalid_search_query(self):
"""Test for invalid search queries"""
rv = self.app.get('/search?q=')
assert 'Invalid search query' in rv.data
rv = self.app.get('/search?nonsense=nonsense')
assert 'Invalid search query' in rv.data
def test_invalid_user_id(self):
"""Test for invalid user ids"""
rv = self.app.get('/user?uid=')
assert 'Invalid user id' in rv.data
rv = self.app.get('/user?nonsense=nonsense')
assert 'Invalid user id' in rv.data
if __name__ == '__main__':
unittest.main()
|
import TwitterSA
import unittest
class TwitterSATestCase(unittest.TestCase):
def setUp(self):
TwitterSA.app.config['TESTING'] = True
self.app = TwitterSA.app.test_client()
def tearDown(self):
pass
def test_twitter_api(self):
"""Test to make sure the API is getting tweets"""
tweets = TwitterSA.api.search(q='hello')
assert tweets and len(tweets)
def test_invalid_search_query(self):
"""Test for invalid search queries"""
rv = self.app.get('/search?q=')
assert 'Invalid search query' in rv.data
rv = self.app.get('/search?nonsense=nonsense')
assert 'Invalid search query' in rv.data
def test_invalid_user_id(self):
"""Test for invalid user ids"""
rv = self.app.get('/user?uid=')
assert 'Invalid user id' in rv.data
rv = self.app.get('/user?nonsense=nonsense')
assert 'Invalid user id' in rv.data
if __name__ == '__main__':
unittest.main()
| Add twitter API functionality test | Add twitter API functionality test
| Python | mit | jayelm/twittersa,jayelm/twittersa |
import TwitterSA
import unittest
class TwitterSATestCase(unittest.TestCase):
def setUp(self):
TwitterSA.app.config['TESTING'] = True
self.app = TwitterSA.app.test_client()
def tearDown(self):
pass
+
+ def test_twitter_api(self):
+ """Test to make sure the API is getting tweets"""
+ tweets = TwitterSA.api.search(q='hello')
+ assert tweets and len(tweets)
def test_invalid_search_query(self):
"""Test for invalid search queries"""
rv = self.app.get('/search?q=')
assert 'Invalid search query' in rv.data
rv = self.app.get('/search?nonsense=nonsense')
assert 'Invalid search query' in rv.data
def test_invalid_user_id(self):
"""Test for invalid user ids"""
rv = self.app.get('/user?uid=')
assert 'Invalid user id' in rv.data
rv = self.app.get('/user?nonsense=nonsense')
assert 'Invalid user id' in rv.data
if __name__ == '__main__':
unittest.main()
| Add twitter API functionality test | ## Code Before:
import TwitterSA
import unittest
class TwitterSATestCase(unittest.TestCase):
def setUp(self):
TwitterSA.app.config['TESTING'] = True
self.app = TwitterSA.app.test_client()
def tearDown(self):
pass
def test_invalid_search_query(self):
"""Test for invalid search queries"""
rv = self.app.get('/search?q=')
assert 'Invalid search query' in rv.data
rv = self.app.get('/search?nonsense=nonsense')
assert 'Invalid search query' in rv.data
def test_invalid_user_id(self):
"""Test for invalid user ids"""
rv = self.app.get('/user?uid=')
assert 'Invalid user id' in rv.data
rv = self.app.get('/user?nonsense=nonsense')
assert 'Invalid user id' in rv.data
if __name__ == '__main__':
unittest.main()
## Instruction:
Add twitter API functionality test
## Code After:
import TwitterSA
import unittest
class TwitterSATestCase(unittest.TestCase):
def setUp(self):
TwitterSA.app.config['TESTING'] = True
self.app = TwitterSA.app.test_client()
def tearDown(self):
pass
def test_twitter_api(self):
"""Test to make sure the API is getting tweets"""
tweets = TwitterSA.api.search(q='hello')
assert tweets and len(tweets)
def test_invalid_search_query(self):
"""Test for invalid search queries"""
rv = self.app.get('/search?q=')
assert 'Invalid search query' in rv.data
rv = self.app.get('/search?nonsense=nonsense')
assert 'Invalid search query' in rv.data
def test_invalid_user_id(self):
"""Test for invalid user ids"""
rv = self.app.get('/user?uid=')
assert 'Invalid user id' in rv.data
rv = self.app.get('/user?nonsense=nonsense')
assert 'Invalid user id' in rv.data
if __name__ == '__main__':
unittest.main()
| // ... existing code ...
def tearDown(self):
pass
def test_twitter_api(self):
"""Test to make sure the API is getting tweets"""
tweets = TwitterSA.api.search(q='hello')
assert tweets and len(tweets)
def test_invalid_search_query(self):
// ... rest of the code ... |
e4ee7034291fbeda48efa0d1c617be8a20eb49bd | algorithms/python/496_next_greater_element.py | algorithms/python/496_next_greater_element.py | class Solution(object):
def nextGreaterElement(self, findNums, nums):
"""
:type findNums: List[int]
:type nums: List[int]
:rtype: List[int]
"""
results = []
for findNum in findNums:
index = nums.index(findNum)
result = index + 1
for candidate in nums[index + 1:]:
if candidate > findNum:
results.append(candidate)
break
else:
result += 1
if result >= len(nums):
results.append(-1)
return results
| class Solution(object):
def nextGreaterElement(self, findNums, nums):
"""
:type findNums: List[int]
:type nums: List[int]
:rtype: List[int]
"""
results = []
for findNum in findNums:
index = nums.index(findNum)
result = index + 1
for candidate in nums[index + 1:]:
if candidate > findNum:
results.append(candidate)
break
else:
result += 1
if result >= len(nums):
results.append(-1)
return results
# Solution 2
class Solution(object):
def nextGreaterElement(self, findNums, nums):
"""
:type findNums: List[int]
:type nums: List[int]
:rtype: List[int]
"""
result_hash = {}
stack = []
for num in nums:
while stack and num > stack[-1]:
result_hash[stack.pop()] = num
stack.append(num)
return [result_hash.get(x, -1) for x in findNums]
| Add another solution for 496 next greater element | Add another solution for 496 next greater element
| Python | mit | ruichao-factual/leetcode | class Solution(object):
def nextGreaterElement(self, findNums, nums):
"""
:type findNums: List[int]
:type nums: List[int]
:rtype: List[int]
"""
results = []
for findNum in findNums:
index = nums.index(findNum)
result = index + 1
for candidate in nums[index + 1:]:
if candidate > findNum:
results.append(candidate)
break
else:
result += 1
if result >= len(nums):
results.append(-1)
return results
+
+ # Solution 2
+
+ class Solution(object):
+ def nextGreaterElement(self, findNums, nums):
+ """
+ :type findNums: List[int]
+ :type nums: List[int]
+ :rtype: List[int]
+ """
+ result_hash = {}
+ stack = []
+
+ for num in nums:
+ while stack and num > stack[-1]:
+ result_hash[stack.pop()] = num
+ stack.append(num)
+
+ return [result_hash.get(x, -1) for x in findNums]
+ | Add another solution for 496 next greater element | ## Code Before:
class Solution(object):
def nextGreaterElement(self, findNums, nums):
"""
:type findNums: List[int]
:type nums: List[int]
:rtype: List[int]
"""
results = []
for findNum in findNums:
index = nums.index(findNum)
result = index + 1
for candidate in nums[index + 1:]:
if candidate > findNum:
results.append(candidate)
break
else:
result += 1
if result >= len(nums):
results.append(-1)
return results
## Instruction:
Add another solution for 496 next greater element
## Code After:
class Solution(object):
def nextGreaterElement(self, findNums, nums):
"""
:type findNums: List[int]
:type nums: List[int]
:rtype: List[int]
"""
results = []
for findNum in findNums:
index = nums.index(findNum)
result = index + 1
for candidate in nums[index + 1:]:
if candidate > findNum:
results.append(candidate)
break
else:
result += 1
if result >= len(nums):
results.append(-1)
return results
# Solution 2
class Solution(object):
def nextGreaterElement(self, findNums, nums):
"""
:type findNums: List[int]
:type nums: List[int]
:rtype: List[int]
"""
result_hash = {}
stack = []
for num in nums:
while stack and num > stack[-1]:
result_hash[stack.pop()] = num
stack.append(num)
return [result_hash.get(x, -1) for x in findNums]
| ...
return results
# Solution 2
class Solution(object):
def nextGreaterElement(self, findNums, nums):
"""
:type findNums: List[int]
:type nums: List[int]
:rtype: List[int]
"""
result_hash = {}
stack = []
for num in nums:
while stack and num > stack[-1]:
result_hash[stack.pop()] = num
stack.append(num)
return [result_hash.get(x, -1) for x in findNums]
... |
a633fd37a4d795e7b565254ef10aaa0f2ad77f31 | vcontrol/rest/machines/shutdown.py | vcontrol/rest/machines/shutdown.py | from ..helpers import get_allowed
import subprocess
import web
class ShutdownMachineR:
"""
This endpoint is for shutting down a running machine.
"""
allow_origin, rest_url = get_allowed.get_allowed()
def GET(self, machine):
web.header('Access-Control-Allow-Origin', self.allow_origin)
try:
out = subprocess.check_output("/usr/local/bin/docker-machine stop "+machine, shell=True)
except:
out = "unable to stop machine"
return str(out)
| from ..helpers import get_allowed
import subprocess
import web
class ShutdownMachineR:
"""
This endpoint is for shutting down a running machine.
"""
allow_origin, rest_url = get_allowed.get_allowed()
def GET(self, machine):
try:
web.header('Access-Control-Allow-Origin', self.allow_origin)
except Exception as e: # no pragma
print(e.message)
try:
out = subprocess.check_output("/usr/local/bin/docker-machine stop "+machine, shell=True)
except:
out = "unable to stop machine"
return str(out)
| Put the web.header function in a try/except block | Put the web.header function in a try/except block
| Python | apache-2.0 | cglewis/vcontrol,CyberReboot/vcontrol,CyberReboot/vcontrol,cglewis/vcontrol,CyberReboot/vcontrol,cglewis/vcontrol | from ..helpers import get_allowed
import subprocess
import web
+
class ShutdownMachineR:
"""
This endpoint is for shutting down a running machine.
"""
allow_origin, rest_url = get_allowed.get_allowed()
+
def GET(self, machine):
+ try:
- web.header('Access-Control-Allow-Origin', self.allow_origin)
+ web.header('Access-Control-Allow-Origin', self.allow_origin)
+ except Exception as e: # no pragma
+ print(e.message)
try:
out = subprocess.check_output("/usr/local/bin/docker-machine stop "+machine, shell=True)
except:
out = "unable to stop machine"
return str(out)
| Put the web.header function in a try/except block | ## Code Before:
from ..helpers import get_allowed
import subprocess
import web
class ShutdownMachineR:
"""
This endpoint is for shutting down a running machine.
"""
allow_origin, rest_url = get_allowed.get_allowed()
def GET(self, machine):
web.header('Access-Control-Allow-Origin', self.allow_origin)
try:
out = subprocess.check_output("/usr/local/bin/docker-machine stop "+machine, shell=True)
except:
out = "unable to stop machine"
return str(out)
## Instruction:
Put the web.header function in a try/except block
## Code After:
from ..helpers import get_allowed
import subprocess
import web
class ShutdownMachineR:
"""
This endpoint is for shutting down a running machine.
"""
allow_origin, rest_url = get_allowed.get_allowed()
def GET(self, machine):
try:
web.header('Access-Control-Allow-Origin', self.allow_origin)
except Exception as e: # no pragma
print(e.message)
try:
out = subprocess.check_output("/usr/local/bin/docker-machine stop "+machine, shell=True)
except:
out = "unable to stop machine"
return str(out)
| ...
import subprocess
import web
class ShutdownMachineR:
...
"""
allow_origin, rest_url = get_allowed.get_allowed()
def GET(self, machine):
try:
web.header('Access-Control-Allow-Origin', self.allow_origin)
except Exception as e: # no pragma
print(e.message)
try:
out = subprocess.check_output("/usr/local/bin/docker-machine stop "+machine, shell=True)
... |
34f8c0a4a0a9f78c124cd07b121ce5b2fbf00136 | onadata/libs/utils/csv_import.py | onadata/libs/utils/csv_import.py | import unicodecsv as ucsv
from cStringIO import StringIO
from ondata.apps.api.viewsets.xform_submission_api import dict_lists2strings
from onadata.libs.utils.logger_tools import dict2xform, safe_create_instance
def submit_csv(username, request, csv_data):
if isinstance(csv_data, (str, unicode)):
csv_data = StringIO(csv_data)
elif not isinstance(csv_data, file):
raise TypeError(u'Invalid param type for `csv_data`. '
'Expected file, String or Unicode but '
'got {} instead.'.format(type(csv_data).__name__))
csv_reader = ucsv.DictReader(csv_data)
for row in csv_reader:
xml_file = StringIO(dict2xform(dict_lists2strings(row),
row.get('_uuid')))
safe_create_instance(username, xml_file, [], None, None)
| import unicodecsv as ucsv
from cStringIO import StringIO
from ondata.apps.api.viewsets.xform_submission_api import dict_lists2strings
from onadata.libs.utils.logger_tools import dict2xform, safe_create_instance
from django.db import transaction
class CSVImportException(Exception):
pass
def submit_csv(username, request, csv_data):
if isinstance(csv_data, (str, unicode)):
csv_data = StringIO(csv_data)
elif not isinstance(csv_data, file):
raise TypeError(u'Invalid param type for `csv_data`. '
'Expected file, String or Unicode but '
'got {} instead.'.format(type(csv_data).__name__))
csv_reader = ucsv.DictReader(csv_data)
with transaction.atomic():
for row in csv_reader:
# fetch submission uuid before nuking row metadata
_uuid = row.get('_uuid')
# nuke metadata (keys starting with '_')
for key in row.keys():
if key.startswith('_'):
del row[key]
xml_file = StringIO(dict2xform(dict_lists2strings(row), _uuid))
error, instance = safe_create_instance(
username, xml_file, [], None, None)
if error is None:
raise CSVImportException(error)
| Implement atomicity for CSV imports | JZ: Implement atomicity for CSV imports
CSV imports should happen for all rows or nothing at all!
Use `django.transactions` for rollbacks on submission on errors
Also remove metadata from CSV rows before submitting
| Python | bsd-2-clause | awemulya/fieldsight-kobocat,mainakibui/kobocat,qlands/onadata,smn/onadata,sounay/flaminggo-test,piqoni/onadata,qlands/onadata,jomolinare/kobocat,sounay/flaminggo-test,mainakibui/kobocat,kobotoolbox/kobocat,kobotoolbox/kobocat,smn/onadata,piqoni/onadata,hnjamba/onaclone,awemulya/fieldsight-kobocat,smn/onadata,jomolinare/kobocat,jomolinare/kobocat,smn/onadata,spatialdev/onadata,qlands/onadata,GeoODK/onadata,kobotoolbox/kobocat,spatialdev/onadata,hnjamba/onaclone,GeoODK/onadata,spatialdev/onadata,awemulya/fieldsight-kobocat,sounay/flaminggo-test,awemulya/fieldsight-kobocat,sounay/flaminggo-test,kobotoolbox/kobocat,piqoni/onadata,GeoODK/onadata,hnjamba/onaclone,qlands/onadata,hnjamba/onaclone,GeoODK/onadata,piqoni/onadata,spatialdev/onadata,mainakibui/kobocat,mainakibui/kobocat,jomolinare/kobocat | import unicodecsv as ucsv
from cStringIO import StringIO
from ondata.apps.api.viewsets.xform_submission_api import dict_lists2strings
from onadata.libs.utils.logger_tools import dict2xform, safe_create_instance
+ from django.db import transaction
+
+
+ class CSVImportException(Exception):
+ pass
def submit_csv(username, request, csv_data):
if isinstance(csv_data, (str, unicode)):
csv_data = StringIO(csv_data)
elif not isinstance(csv_data, file):
raise TypeError(u'Invalid param type for `csv_data`. '
'Expected file, String or Unicode but '
'got {} instead.'.format(type(csv_data).__name__))
csv_reader = ucsv.DictReader(csv_data)
+ with transaction.atomic():
- for row in csv_reader:
+ for row in csv_reader:
+ # fetch submission uuid before nuking row metadata
+ _uuid = row.get('_uuid')
+ # nuke metadata (keys starting with '_')
+ for key in row.keys():
+ if key.startswith('_'):
+ del row[key]
- xml_file = StringIO(dict2xform(dict_lists2strings(row),
+ xml_file = StringIO(dict2xform(dict_lists2strings(row), _uuid))
- row.get('_uuid')))
- safe_create_instance(username, xml_file, [], None, None)
+ error, instance = safe_create_instance(
+ username, xml_file, [], None, None)
+ if error is None:
+ raise CSVImportException(error)
| Implement atomicity for CSV imports | ## Code Before:
import unicodecsv as ucsv
from cStringIO import StringIO
from ondata.apps.api.viewsets.xform_submission_api import dict_lists2strings
from onadata.libs.utils.logger_tools import dict2xform, safe_create_instance
def submit_csv(username, request, csv_data):
if isinstance(csv_data, (str, unicode)):
csv_data = StringIO(csv_data)
elif not isinstance(csv_data, file):
raise TypeError(u'Invalid param type for `csv_data`. '
'Expected file, String or Unicode but '
'got {} instead.'.format(type(csv_data).__name__))
csv_reader = ucsv.DictReader(csv_data)
for row in csv_reader:
xml_file = StringIO(dict2xform(dict_lists2strings(row),
row.get('_uuid')))
safe_create_instance(username, xml_file, [], None, None)
## Instruction:
Implement atomicity for CSV imports
## Code After:
import unicodecsv as ucsv
from cStringIO import StringIO
from ondata.apps.api.viewsets.xform_submission_api import dict_lists2strings
from onadata.libs.utils.logger_tools import dict2xform, safe_create_instance
from django.db import transaction
class CSVImportException(Exception):
pass
def submit_csv(username, request, csv_data):
if isinstance(csv_data, (str, unicode)):
csv_data = StringIO(csv_data)
elif not isinstance(csv_data, file):
raise TypeError(u'Invalid param type for `csv_data`. '
'Expected file, String or Unicode but '
'got {} instead.'.format(type(csv_data).__name__))
csv_reader = ucsv.DictReader(csv_data)
with transaction.atomic():
for row in csv_reader:
# fetch submission uuid before nuking row metadata
_uuid = row.get('_uuid')
# nuke metadata (keys starting with '_')
for key in row.keys():
if key.startswith('_'):
del row[key]
xml_file = StringIO(dict2xform(dict_lists2strings(row), _uuid))
error, instance = safe_create_instance(
username, xml_file, [], None, None)
if error is None:
raise CSVImportException(error)
| # ... existing code ...
from ondata.apps.api.viewsets.xform_submission_api import dict_lists2strings
from onadata.libs.utils.logger_tools import dict2xform, safe_create_instance
from django.db import transaction
class CSVImportException(Exception):
pass
# ... modified code ...
csv_reader = ucsv.DictReader(csv_data)
with transaction.atomic():
for row in csv_reader:
# fetch submission uuid before nuking row metadata
_uuid = row.get('_uuid')
# nuke metadata (keys starting with '_')
for key in row.keys():
if key.startswith('_'):
del row[key]
xml_file = StringIO(dict2xform(dict_lists2strings(row), _uuid))
error, instance = safe_create_instance(
username, xml_file, [], None, None)
if error is None:
raise CSVImportException(error)
# ... rest of the code ... |
a27b03a89af6442dc8e1be3d310a8fc046a98ed4 | foampy/tests.py | foampy/tests.py |
from .core import *
from .dictionaries import *
from .types import *
from .foil import *
| """Tests for foamPy."""
from .core import *
from .dictionaries import *
from .types import *
from .foil import *
def test_load_all_torque_drag():
"""Test the `load_all_torque_drag` function."""
t, torque, drag = load_all_torque_drag(casedir="test")
assert t.max() == 4.0
| Add test for loading all torque and drag data | Add test for loading all torque and drag data
| Python | mit | petebachant/foamPy,petebachant/foamPy,petebachant/foamPy | + """Tests for foamPy."""
from .core import *
from .dictionaries import *
from .types import *
from .foil import *
+
+ def test_load_all_torque_drag():
+ """Test the `load_all_torque_drag` function."""
+ t, torque, drag = load_all_torque_drag(casedir="test")
+ assert t.max() == 4.0
+ | Add test for loading all torque and drag data | ## Code Before:
from .core import *
from .dictionaries import *
from .types import *
from .foil import *
## Instruction:
Add test for loading all torque and drag data
## Code After:
"""Tests for foamPy."""
from .core import *
from .dictionaries import *
from .types import *
from .foil import *
def test_load_all_torque_drag():
"""Test the `load_all_torque_drag` function."""
t, torque, drag = load_all_torque_drag(casedir="test")
assert t.max() == 4.0
| # ... existing code ...
"""Tests for foamPy."""
from .core import *
# ... modified code ...
from .types import *
from .foil import *
def test_load_all_torque_drag():
"""Test the `load_all_torque_drag` function."""
t, torque, drag = load_all_torque_drag(casedir="test")
assert t.max() == 4.0
# ... rest of the code ... |
c755934a9bc9f15f1e7dcf6d337c3dd3acf4e824 | checks/check_solarize.py | checks/check_solarize.py | import imgaug as ia
import imgaug.augmenters as iaa
def main():
image = ia.quokka_square((128, 128))
images_aug = iaa.Solarize(1.0)(images=[image] * (5*5))
ia.imshow(ia.draw_grid(images_aug))
if __name__ == "__main__":
main()
| from __future__ import print_function, division, absolute_import
import imgaug as ia
import imgaug.augmenters as iaa
import timeit
def main():
for size in [64, 128, 256, 512, 1024]:
for threshold in [64, 128, 192]:
time_iaa = timeit.timeit(
"iaa.solarize(image, %d)" % (threshold,),
number=1000,
setup=(
"import imgaug as ia; "
"import imgaug.augmenters as iaa; "
"image = ia.quokka_square((%d, %d))" % (size, size))
)
time_pil = timeit.timeit(
"np.asarray("
"PIL.ImageOps.solarize(PIL.Image.fromarray(image), %d)"
")" % (threshold,),
number=1000,
setup=(
"import numpy as np; "
"import PIL.Image; "
"import PIL.ImageOps; "
"import imgaug as ia; "
"image = ia.quokka_square((%d, %d))" % (size, size))
)
print("[size=%04d, thresh=%03d] iaa=%.4f pil=%.4f" % (
size, threshold, time_iaa, time_pil))
image = ia.quokka_square((128, 128))
images_aug = iaa.Solarize(1.0)(images=[image] * (5*5))
ia.imshow(ia.draw_grid(images_aug))
if __name__ == "__main__":
main()
| Add performance comparison with PIL | Add performance comparison with PIL
| Python | mit | aleju/ImageAugmenter,aleju/imgaug,aleju/imgaug | + from __future__ import print_function, division, absolute_import
import imgaug as ia
import imgaug.augmenters as iaa
+ import timeit
def main():
+ for size in [64, 128, 256, 512, 1024]:
+ for threshold in [64, 128, 192]:
+ time_iaa = timeit.timeit(
+ "iaa.solarize(image, %d)" % (threshold,),
+ number=1000,
+ setup=(
+ "import imgaug as ia; "
+ "import imgaug.augmenters as iaa; "
+ "image = ia.quokka_square((%d, %d))" % (size, size))
+ )
+ time_pil = timeit.timeit(
+ "np.asarray("
+ "PIL.ImageOps.solarize(PIL.Image.fromarray(image), %d)"
+ ")" % (threshold,),
+ number=1000,
+ setup=(
+ "import numpy as np; "
+ "import PIL.Image; "
+ "import PIL.ImageOps; "
+ "import imgaug as ia; "
+ "image = ia.quokka_square((%d, %d))" % (size, size))
+ )
+ print("[size=%04d, thresh=%03d] iaa=%.4f pil=%.4f" % (
+ size, threshold, time_iaa, time_pil))
+
image = ia.quokka_square((128, 128))
images_aug = iaa.Solarize(1.0)(images=[image] * (5*5))
ia.imshow(ia.draw_grid(images_aug))
if __name__ == "__main__":
main()
| Add performance comparison with PIL | ## Code Before:
import imgaug as ia
import imgaug.augmenters as iaa
def main():
image = ia.quokka_square((128, 128))
images_aug = iaa.Solarize(1.0)(images=[image] * (5*5))
ia.imshow(ia.draw_grid(images_aug))
if __name__ == "__main__":
main()
## Instruction:
Add performance comparison with PIL
## Code After:
from __future__ import print_function, division, absolute_import
import imgaug as ia
import imgaug.augmenters as iaa
import timeit
def main():
for size in [64, 128, 256, 512, 1024]:
for threshold in [64, 128, 192]:
time_iaa = timeit.timeit(
"iaa.solarize(image, %d)" % (threshold,),
number=1000,
setup=(
"import imgaug as ia; "
"import imgaug.augmenters as iaa; "
"image = ia.quokka_square((%d, %d))" % (size, size))
)
time_pil = timeit.timeit(
"np.asarray("
"PIL.ImageOps.solarize(PIL.Image.fromarray(image), %d)"
")" % (threshold,),
number=1000,
setup=(
"import numpy as np; "
"import PIL.Image; "
"import PIL.ImageOps; "
"import imgaug as ia; "
"image = ia.quokka_square((%d, %d))" % (size, size))
)
print("[size=%04d, thresh=%03d] iaa=%.4f pil=%.4f" % (
size, threshold, time_iaa, time_pil))
image = ia.quokka_square((128, 128))
images_aug = iaa.Solarize(1.0)(images=[image] * (5*5))
ia.imshow(ia.draw_grid(images_aug))
if __name__ == "__main__":
main()
| # ... existing code ...
from __future__ import print_function, division, absolute_import
import imgaug as ia
import imgaug.augmenters as iaa
import timeit
def main():
for size in [64, 128, 256, 512, 1024]:
for threshold in [64, 128, 192]:
time_iaa = timeit.timeit(
"iaa.solarize(image, %d)" % (threshold,),
number=1000,
setup=(
"import imgaug as ia; "
"import imgaug.augmenters as iaa; "
"image = ia.quokka_square((%d, %d))" % (size, size))
)
time_pil = timeit.timeit(
"np.asarray("
"PIL.ImageOps.solarize(PIL.Image.fromarray(image), %d)"
")" % (threshold,),
number=1000,
setup=(
"import numpy as np; "
"import PIL.Image; "
"import PIL.ImageOps; "
"import imgaug as ia; "
"image = ia.quokka_square((%d, %d))" % (size, size))
)
print("[size=%04d, thresh=%03d] iaa=%.4f pil=%.4f" % (
size, threshold, time_iaa, time_pil))
image = ia.quokka_square((128, 128))
images_aug = iaa.Solarize(1.0)(images=[image] * (5*5))
# ... rest of the code ... |
b5e13cb92f539545873d59553c03e1523eac1dbb | recipes/kaleido-core/run_test.py | recipes/kaleido-core/run_test.py | from subprocess import Popen, PIPE
import json
import platform
# Remove "sys.exit" after feedstock creation when running
# on linux-anvil-cos7-x86_64 image
if platform.system() == "Linux":
import sys
sys.exit(0)
if platform.system() == "Windows":
ext = ".cmd"
else:
ext = ""
p = Popen(
['kaleido' + ext, "plotly", "--disable-gpu"],
stdout=PIPE, stdin=PIPE, stderr=PIPE,
text=True
)
stdout_data = p.communicate(
input=json.dumps({"data": {"data": []}, "format": "png"})
)[0]
assert "iVBORw" in stdout_data
| from subprocess import Popen, PIPE
import json
import platform
# Remove "sys.exit" after feedstock creation when running
# on linux-anvil-cos7-x86_64 image
if platform.system() == "Linux":
import sys
sys.exit(0)
if platform.system() == "Windows":
ext = ".cmd"
else:
ext = ""
p = Popen(
['kaleido' + ext, "plotly", "--disable-gpu", "--no-sandbox", "--disable-breakpad"],
stdout=PIPE, stdin=PIPE, stderr=PIPE,
text=True
)
stdout_data = p.communicate(
input=json.dumps({"data": {"data": []}, "format": "png"})
)[0]
assert "iVBORw" in stdout_data
| Fix hanging test on MacOS | Fix hanging test on MacOS
| Python | bsd-3-clause | kwilcox/staged-recipes,ocefpaf/staged-recipes,stuertz/staged-recipes,johanneskoester/staged-recipes,jochym/staged-recipes,SylvainCorlay/staged-recipes,goanpeca/staged-recipes,igortg/staged-recipes,mariusvniekerk/staged-recipes,conda-forge/staged-recipes,patricksnape/staged-recipes,goanpeca/staged-recipes,johanneskoester/staged-recipes,scopatz/staged-recipes,ocefpaf/staged-recipes,ReimarBauer/staged-recipes,scopatz/staged-recipes,hadim/staged-recipes,hadim/staged-recipes,jochym/staged-recipes,patricksnape/staged-recipes,mariusvniekerk/staged-recipes,conda-forge/staged-recipes,stuertz/staged-recipes,jakirkham/staged-recipes,ReimarBauer/staged-recipes,igortg/staged-recipes,jakirkham/staged-recipes,SylvainCorlay/staged-recipes,kwilcox/staged-recipes | from subprocess import Popen, PIPE
import json
import platform
# Remove "sys.exit" after feedstock creation when running
# on linux-anvil-cos7-x86_64 image
if platform.system() == "Linux":
import sys
sys.exit(0)
if platform.system() == "Windows":
ext = ".cmd"
else:
ext = ""
p = Popen(
- ['kaleido' + ext, "plotly", "--disable-gpu"],
+ ['kaleido' + ext, "plotly", "--disable-gpu", "--no-sandbox", "--disable-breakpad"],
stdout=PIPE, stdin=PIPE, stderr=PIPE,
text=True
)
stdout_data = p.communicate(
input=json.dumps({"data": {"data": []}, "format": "png"})
)[0]
assert "iVBORw" in stdout_data
| Fix hanging test on MacOS | ## Code Before:
from subprocess import Popen, PIPE
import json
import platform
# Remove "sys.exit" after feedstock creation when running
# on linux-anvil-cos7-x86_64 image
if platform.system() == "Linux":
import sys
sys.exit(0)
if platform.system() == "Windows":
ext = ".cmd"
else:
ext = ""
p = Popen(
['kaleido' + ext, "plotly", "--disable-gpu"],
stdout=PIPE, stdin=PIPE, stderr=PIPE,
text=True
)
stdout_data = p.communicate(
input=json.dumps({"data": {"data": []}, "format": "png"})
)[0]
assert "iVBORw" in stdout_data
## Instruction:
Fix hanging test on MacOS
## Code After:
from subprocess import Popen, PIPE
import json
import platform
# Remove "sys.exit" after feedstock creation when running
# on linux-anvil-cos7-x86_64 image
if platform.system() == "Linux":
import sys
sys.exit(0)
if platform.system() == "Windows":
ext = ".cmd"
else:
ext = ""
p = Popen(
['kaleido' + ext, "plotly", "--disable-gpu", "--no-sandbox", "--disable-breakpad"],
stdout=PIPE, stdin=PIPE, stderr=PIPE,
text=True
)
stdout_data = p.communicate(
input=json.dumps({"data": {"data": []}, "format": "png"})
)[0]
assert "iVBORw" in stdout_data
| ...
p = Popen(
['kaleido' + ext, "plotly", "--disable-gpu", "--no-sandbox", "--disable-breakpad"],
stdout=PIPE, stdin=PIPE, stderr=PIPE,
text=True
... |
d42b9da06d5cde89a6116d711fc6ae216256cabc | shell/view/home/IconLayout.py | shell/view/home/IconLayout.py | import random
class IconLayout:
def __init__(self, width, height):
self._icons = []
self._width = width
self._height = height
def add_icon(self, icon):
self._icons.append(icon)
self._layout_icon(icon)
def remove_icon(self, icon):
self._icons.remove(icon)
def _is_valid_position(self, icon, x, y):
icon_size = icon.props.size
border = 20
if not (border < x < self._width - icon_size - border and \
border < y < self._height - icon_size - border):
return False
return True
def _layout_icon(self, icon):
while True:
x = random.random() * self._width
y = random.random() * self._height
if self._is_valid_position(icon, x, y):
break
icon.props.x = x
icon.props.y = y
| import random
class IconLayout:
def __init__(self, width, height):
self._icons = []
self._width = width
self._height = height
def add_icon(self, icon):
self._icons.append(icon)
self._layout_icon(icon)
def remove_icon(self, icon):
self._icons.remove(icon)
def _is_valid_position(self, icon, x, y):
icon_size = icon.get_property('size')
border = 20
if not (border < x < self._width - icon_size - border and \
border < y < self._height - icon_size - border):
return False
return True
def _layout_icon(self, icon):
while True:
x = random.random() * self._width
y = random.random() * self._height
if self._is_valid_position(icon, x, y):
break
icon.set_property('x', x)
icon.set_property('y', y)
| Use get/set_property rather than direct accessors | Use get/set_property rather than direct accessors
| Python | lgpl-2.1 | Daksh/sugar-toolkit-gtk3,manuq/sugar-toolkit-gtk3,tchx84/sugar-toolkit-gtk3,quozl/sugar-toolkit-gtk3,tchx84/sugar-toolkit-gtk3,i5o/sugar-toolkit-gtk3,puneetgkaur/backup_sugar_sugartoolkit,quozl/sugar-toolkit-gtk3,godiard/sugar-toolkit-gtk3,tchx84/debian-pkg-sugar-toolkit,tchx84/debian-pkg-sugar-toolkit,ceibal-tatu/sugar-toolkit,gusDuarte/sugar-toolkit-gtk3,tchx84/debian-pkg-sugar-toolkit,tchx84/debian-pkg-sugar-toolkit-gtk3,tchx84/debian-pkg-sugar-toolkit-gtk3,ceibal-tatu/sugar-toolkit-gtk3,manuq/sugar-toolkit-gtk3,samdroid-apps/sugar-toolkit-gtk3,puneetgkaur/sugar-toolkit-gtk3,sugarlabs/sugar-toolkit-gtk3,samdroid-apps/sugar-toolkit-gtk3,godiard/sugar-toolkit-gtk3,sugarlabs/sugar-toolkit,i5o/sugar-toolkit-gtk3,Daksh/sugar-toolkit-gtk3,godiard/sugar-toolkit-gtk3,tchx84/debian-pkg-sugar-toolkit-gtk3,ceibal-tatu/sugar-toolkit-gtk3,ceibal-tatu/sugar-toolkit,gusDuarte/sugar-toolkit-gtk3,puneetgkaur/sugar-toolkit-gtk3,quozl/sugar-toolkit-gtk3,i5o/sugar-toolkit-gtk3,i5o/sugar-toolkit-gtk3,Daksh/sugar-toolkit-gtk3,sugarlabs/sugar-toolkit-gtk3,samdroid-apps/sugar-toolkit-gtk3,puneetgkaur/sugar-toolkit-gtk3,samdroid-apps/sugar-toolkit-gtk3,gusDuarte/sugar-toolkit-gtk3,gusDuarte/sugar-toolkit-gtk3,ceibal-tatu/sugar-toolkit-gtk3,sugarlabs/sugar-toolkit,sugarlabs/sugar-toolkit,sugarlabs/sugar-toolkit-gtk3,quozl/sugar-toolkit-gtk3,manuq/sugar-toolkit-gtk3,puneetgkaur/backup_sugar_sugartoolkit,puneetgkaur/backup_sugar_sugartoolkit,sugarlabs/sugar-toolkit,ceibal-tatu/sugar-toolkit,tchx84/sugar-toolkit-gtk3 | import random
class IconLayout:
def __init__(self, width, height):
self._icons = []
self._width = width
self._height = height
def add_icon(self, icon):
self._icons.append(icon)
self._layout_icon(icon)
def remove_icon(self, icon):
self._icons.remove(icon)
def _is_valid_position(self, icon, x, y):
- icon_size = icon.props.size
+ icon_size = icon.get_property('size')
border = 20
if not (border < x < self._width - icon_size - border and \
border < y < self._height - icon_size - border):
return False
return True
def _layout_icon(self, icon):
while True:
x = random.random() * self._width
y = random.random() * self._height
if self._is_valid_position(icon, x, y):
break
- icon.props.x = x
- icon.props.y = y
+ icon.set_property('x', x)
+ icon.set_property('y', y)
| Use get/set_property rather than direct accessors | ## Code Before:
import random
class IconLayout:
def __init__(self, width, height):
self._icons = []
self._width = width
self._height = height
def add_icon(self, icon):
self._icons.append(icon)
self._layout_icon(icon)
def remove_icon(self, icon):
self._icons.remove(icon)
def _is_valid_position(self, icon, x, y):
icon_size = icon.props.size
border = 20
if not (border < x < self._width - icon_size - border and \
border < y < self._height - icon_size - border):
return False
return True
def _layout_icon(self, icon):
while True:
x = random.random() * self._width
y = random.random() * self._height
if self._is_valid_position(icon, x, y):
break
icon.props.x = x
icon.props.y = y
## Instruction:
Use get/set_property rather than direct accessors
## Code After:
import random
class IconLayout:
def __init__(self, width, height):
self._icons = []
self._width = width
self._height = height
def add_icon(self, icon):
self._icons.append(icon)
self._layout_icon(icon)
def remove_icon(self, icon):
self._icons.remove(icon)
def _is_valid_position(self, icon, x, y):
icon_size = icon.get_property('size')
border = 20
if not (border < x < self._width - icon_size - border and \
border < y < self._height - icon_size - border):
return False
return True
def _layout_icon(self, icon):
while True:
x = random.random() * self._width
y = random.random() * self._height
if self._is_valid_position(icon, x, y):
break
icon.set_property('x', x)
icon.set_property('y', y)
| ...
def _is_valid_position(self, icon, x, y):
icon_size = icon.get_property('size')
border = 20
...
break
icon.set_property('x', x)
icon.set_property('y', y)
... |
aabf28c02a4dff593e5e4b156052adb9b81a70c7 | skflow/ops/tests/test_dropout_ops.py | skflow/ops/tests/test_dropout_ops.py |
import numpy as np
import tensorflow as tf
from skflow import ops
class DropoutTest(tf.test.TestCase):
def test_dropout_float(self):
with self.test_session():
x = tf.placeholder(tf.float32, [5, 5])
y = ops.dropout(x, 0.5)
probs = tf.get_collection(ops.DROPOUTS)
self.assertEqual(len(probs), 1)
if __name__ == '__main__':
tf.test.main()
|
import numpy as np
import tensorflow as tf
from skflow import ops
class DropoutTest(tf.test.TestCase):
def test_dropout_float(self):
with self.test_session() as session:
x = tf.placeholder(tf.float32, [5, 5])
y = ops.dropout(x, 0.5)
probs = tf.get_collection(ops.DROPOUTS)
session.run(tf.initialize_all_variables())
self.assertEqual(len(probs), 1)
self.assertEqual(session.run(probs[0]), 0.5)
def test_dropout_tensor(self):
with self.test_session():
x = tf.placeholder(tf.float32, [5, 5])
y = tf.get_variable("prob", [], initializer=tf.constant_initializer(0.5))
z = ops.dropout(x, y)
probs = tf.get_collection(ops.DROPOUTS)
self.assertEqual(probs, [y])
if __name__ == '__main__':
tf.test.main()
| Test for dropout probability be a tensor | Test for dropout probability be a tensor
| Python | apache-2.0 | handroissuazo/tensorflow,elingg/tensorflow,awni/tensorflow,aselle/tensorflow,theflofly/tensorflow,ishay2b/tensorflow,XueqingLin/tensorflow,DavidNorman/tensorflow,odejesush/tensorflow,taknevski/tensorflow-xsmm,AndreasMadsen/tensorflow,Kongsea/tensorflow,hfp/tensorflow-xsmm,jhaux/tensorflow,JingJunYin/tensorflow,tensorflow/tensorflow,mixturemodel-flow/tensorflow,aam-at/tensorflow,ibab/tensorflow,tensorflow/tensorflow-pywrap_saved_model,Carmezim/tensorflow,yongtang/tensorflow,ageron/tensorflow,mrry/tensorflow,HKUST-SING/tensorflow,av8ramit/tensorflow,whn09/tensorflow,tensorflow/tensorflow-pywrap_tf_optimizer,ppwwyyxx/tensorflow,anilmuthineni/tensorflow,laosiaudi/tensorflow,gnieboer/tensorflow,cxxgtxy/tensorflow,gautam1858/tensorflow,gnieboer/tensorflow,annarev/tensorflow,haeusser/tensorflow,JingJunYin/tensorflow,snnn/tensorflow,lukeiwanski/tensorflow-opencl,ibab/tensorflow,alshedivat/tensorflow,AnishShah/tensorflow,anand-c-goog/tensorflow,DCSaunders/tensorflow,seanli9jan/tensorflow,scenarios/tensorflow,xodus7/tensorflow,rabipanda/tensorflow,LUTAN/tensorflow,dancingdan/tensorflow,jart/tensorflow,alisidd/tensorflow,zycdragonball/tensorflow,ghchinoy/tensorflow,tensorflow/tensorflow-pywrap_tf_optimizer,handroissuazo/tensorflow,nolanliou/tensorflow,tensorflow/tensorflow-experimental_link_static_libraries_once,vrv/tensorflow,andrewcmyers/tensorflow,girving/tensorflow,nanditav/15712-TensorFlow,AndreasMadsen/tensorflow,LUTAN/tensorflow,ivano666/tensorflow,asadziach/tensorflow,tntnatbry/tensorflow,DavidNorman/tensorflow,dancingdan/tensorflow,paolodedios/tensorflow,sandeepdsouza93/TensorFlow-15712,dongjoon-hyun/tensorflow,thesuperzapper/tensorflow,dyoung418/tensorflow,alivecor/tensorflow,asimshankar/tensorflow,seanli9jan/tensorflow,ppwwyyxx/tensorflow,hsaputra/tensorflow,strint/tensorflow,gautam1858/tensorflow,benoitsteiner/tensorflow-opencl,wangyum/tensorflow,benoitsteiner/tensorflow-opencl,johndpope/tensorflow,jalexvig/tensorflow,freedomtan/tensorflow,TakayukiSakai/tensorflow,Xeralux/tensorflow,dendisuhubdy/tensorflow,kevin-coder/tensorflow-fork,bowang/tensorflow,renyi533/tensorflow,tensorflow/tensorflow-pywrap_tf_optimizer,alsrgv/tensorflow,benoitsteiner/tensorflow,dyoung418/tensorflow,chris-chris/tensorflow,mavenlin/tensorflow,wangyum/tensorflow,mengxn/tensorflow,horance-liu/tensorflow,aam-at/tensorflow,ZhangXinNan/tensorflow,kevin-coder/tensorflow-fork,lakshayg/tensorflow,calebfoss/tensorflow,mavenlin/tensorflow,a-doumoulakis/tensorflow,tongwang01/tensorflow,gojira/tensorflow,tongwang01/tensorflow,with-git/tensorflow,llhe/tensorflow,jhaux/tensorflow,nburn42/tensorflow,Carmezim/tensorflow,dhalleine/tensorflow,jhseu/tensorflow,aam-at/tensorflow,johndpope/tensorflow,annarev/tensorflow,adit-chandra/tensorflow,cg31/tensorflow,AnishShah/tensorflow,brchiu/tensorflow,manipopopo/tensorflow,dyoung418/tensorflow,abhitopia/tensorflow,kevin-coder/tensorflow-fork,Bismarrck/tensorflow,karllessard/tensorflow,MoamerEncsConcordiaCa/tensorflow,ZhangXinNan/tensorflow,meteorcloudy/tensorflow,frreiss/tensorflow-fred,ychfan/tensorflow,chenjun0210/tensorflow,alheinecke/tensorflow-xsmm,Mazecreator/tensorflow,nburn42/tensorflow,renyi533/tensorflow,chemelnucfin/tensorflow,AnishShah/tensorflow,eadgarchen/tensorflow,alshedivat/tensorflow,lukeiwanski/tensorflow-opencl,martinwicke/tensorflow,hsaputra/tensorflow,LUTAN/tensorflow,theflofly/tensorflow,alistairlow/tensorflow,jendap/tensorflow,mavenlin/tensorflow,alivecor/tensorflow,panmari/tensorflow,jhseu/tensorflow,Mistobaan/tensorflow,arborh/tensorflow,asimshankar/tensorflow,alsrgv/tensorflow,unsiloai/syntaxnet-ops-hack,manazhao/tf_recsys,karllessard/tensorflow,LUTAN/tensorflow,laosiaudi/tensorflow,jendap/tensorflow,aldian/tensorflow,xodus7/tensorflow,sandeepgupta2k4/tensorflow,jhseu/tensorflow,MycChiu/tensorflow,snnn/tensorflow,seanli9jan/tensorflow,strint/tensorflow,cancan101/tensorflow,aldian/tensorflow,theflofly/tensorflow,ppwwyyxx/tensorflow,brchiu/tensorflow,apark263/tensorflow,chemelnucfin/tensorflow,gibiansky/tensorflow,dyoung418/tensorflow,TakayukiSakai/tensorflow,freedomtan/tensorflow,cxxgtxy/tensorflow,ibab/tensorflow,manazhao/tf_recsys,Xeralux/tensorflow,cxxgtxy/tensorflow,tillahoffmann/tensorflow,ivano666/tensorflow,gunan/tensorflow,seanli9jan/tensorflow,code-sauce/tensorflow,ivano666/tensorflow,adamtiger/tensorflow,haeusser/tensorflow,dancingdan/tensorflow,Bulochkin/tensorflow_pack,eadgarchen/tensorflow,ibmsoe/tensorflow,calebfoss/tensorflow,caisq/tensorflow,ppwwyyxx/tensorflow,arborh/tensorflow,nikste/tensorflow,manazhao/tf_recsys,petewarden/tensorflow_makefile,jostep/tensorflow,hsaputra/tensorflow,suiyuan2009/tensorflow,davidzchen/tensorflow,tensorflow/tensorflow-pywrap_tf_optimizer,jwlawson/tensorflow,freedomtan/tensorflow,nanditav/15712-TensorFlow,ninotoshi/tensorflow,pierreg/tensorflow,whn09/tensorflow,freedomtan/tensorflow,dansbecker/skflow,alistairlow/tensorflow,brchiu/tensorflow,elingg/tensorflow,EvenStrangest/tensorflow,lukeiwanski/tensorflow-opencl,DavidNorman/tensorflow,Moriadry/tensorflow,paolodedios/tensorflow,a-doumoulakis/tensorflow,alshedivat/tensorflow,Kongsea/tensorflow,asadziach/tensorflow,chemelnucfin/tensorflow,annarev/tensorflow,dancingdan/tensorflow,MycChiu/tensorflow,davidzchen/tensorflow,kobejean/tensorflow,eadgarchen/tensorflow,gunan/tensorflow,Carmezim/tensorflow,gnieboer/tensorflow,mengxn/tensorflow,ppwwyyxx/tensorflow,xzturn/tensorflow,jwlawson/tensorflow,Mazecreator/tensorflow,alsrgv/tensorflow,lukeiwanski/tensorflow,drpngx/tensorflow,thesuperzapper/tensorflow,LUTAN/tensorflow,alisidd/tensorflow,rabipanda/tensorflow,adit-chandra/tensorflow,rabipanda/tensorflow,alheinecke/tensorflow-xsmm,alisidd/tensorflow,jbedorf/tensorflow,SnakeJenny/TensorFlow,ghchinoy/tensorflow,alsrgv/tensorflow,davidzchen/tensorflow,kevin-coder/tensorflow-fork,peterbraden/tensorflow,Xeralux/tensorflow,ravindrapanda/tensorflow,peterbraden/tensorflow,memo/tensorflow,wangyum/tensorflow,taknevski/tensorflow-xsmm,allenlavoie/tensorflow,krikru/tensorflow-opencl,MoamerEncsConcordiaCa/tensorflow,llhe/tensorflow,kevin-coder/tensorflow-fork,SnakeJenny/TensorFlow,taknevski/tensorflow-xsmm,kchodorow/tensorflow,hsaputra/tensorflow,jbedorf/tensorflow,aam-at/tensorflow,yaroslavvb/tensorflow,SnakeJenny/TensorFlow,ishay2b/tensorflow,sjperkins/tensorflow,jeffzheng1/tensorflow,lakshayg/tensorflow,tensorflow/tensorflow-pywrap_saved_model,tomasreimers/tensorflow-emscripten,tensorflow/tensorflow-experimental_link_static_libraries_once,ppwwyyxx/tensorflow,cancan101/tensorflow,chemelnucfin/tensorflow,snnn/tensorflow,tensorflow/tensorflow-experimental_link_static_libraries_once,benoitsteiner/tensorflow-opencl,ibab/tensorflow,codrut3/tensorflow,wchan/tensorflow,ychfan/tensorflow,jalexvig/tensorflow,kobejean/tensorflow,asadziach/tensorflow,tomasreimers/tensorflow-emscripten,pavelchristof/gomoku-ai,Bulochkin/tensorflow_pack,chenjun0210/tensorflow,dyoung418/tensorflow,code-sauce/tensorflow,thjashin/tensorflow,taknevski/tensorflow-xsmm,ageron/tensorflow,anand-c-goog/tensorflow,ninotoshi/tensorflow,yanchen036/tensorflow,mortada/tensorflow,tomasreimers/tensorflow-emscripten,kamcpp/tensorflow,thesuperzapper/tensorflow,tensorflow/tensorflow-pywrap_saved_model,neilhan/tensorflow,admcrae/tensorflow,drpngx/tensorflow,frreiss/tensorflow-fred,yaroslavvb/tensorflow,xzturn/tensorflow,benoitsteiner/tensorflow,adamtiger/tensorflow,gojira/tensorflow,caisq/tensorflow,lukeiwanski/tensorflow-opencl,DCSaunders/tensorflow,ppries/tensorflow,davidzchen/tensorflow,benoitsteiner/tensorflow,Xeralux/tensorflow,gojira/tensorflow,raymondxyang/tensorflow,jhseu/tensorflow,strint/tensorflow,alivecor/tensorflow,drpngx/tensorflow,JingJunYin/tensorflow,snnn/tensorflow,guschmue/tensorflow,chemelnucfin/tensorflow,tensorflow/tensorflow-pywrap_saved_model,arborh/tensorflow,gnieboer/tensorflow,haeusser/tensorflow,raymondxyang/tensorflow,anilmuthineni/tensorflow,benoitsteiner/tensorflow-xsmm,manipopopo/tensorflow,ppries/tensorflow,asimshankar/tensorflow,chenjun0210/tensorflow,pcm17/tensorflow,vrv/tensorflow,xodus7/tensorflow,nolanliou/tensorflow,code-sauce/tensorflow,eerwitt/tensorflow,Bismarrck/tensorflow,DavidNorman/tensorflow,ravindrapanda/tensorflow,Mistobaan/tensorflow,jbedorf/tensorflow,girving/tensorflow,sarvex/tensorflow,jalexvig/tensorflow,hfp/tensorflow-xsmm,petewarden/tensorflow,dancingdan/tensorflow,ivano666/tensorflow,codrut3/tensorflow,DCSaunders/tensorflow,mixturemodel-flow/tensorflow,jendap/tensorflow,DCSaunders/tensorflow,MycChiu/tensorflow,AnishShah/tensorflow,MycChiu/tensorflow,gibiansky/tensorflow,Intel-tensorflow/tensorflow,JVillella/tensorflow,markslwong/tensorflow,jwlawson/tensorflow,mrry/tensorflow,nburn42/tensorflow,admcrae/tensorflow,jbedorf/tensorflow,ravindrapanda/tensorflow,seanli9jan/tensorflow,eaplatanios/tensorflow,hsaputra/tensorflow,EvenStrangest/tensorflow,gnieboer/tensorflow,adit-chandra/tensorflow,wchan/tensorflow,pcm17/tensorflow,thesuperzapper/tensorflow,martinwicke/tensorflow,seaotterman/tensorflow,gnieboer/tensorflow,lukeiwanski/tensorflow,sandeepdsouza93/TensorFlow-15712,asimshankar/tensorflow,Intel-tensorflow/tensorflow,guschmue/tensorflow,ibab/tensorflow,jbedorf/tensorflow,chemelnucfin/tensorflow,gautam1858/tensorflow,raymondxyang/tensorflow,HKUST-SING/tensorflow,lukeiwanski/tensorflow-opencl,ppries/tensorflow,sandeepdsouza93/TensorFlow-15712,pavelchristof/gomoku-ai,ishay2b/tensorflow,whn09/tensorflow,ibmsoe/tensorflow,meteorcloudy/tensorflow,Carmezim/tensorflow,whn09/tensorflow,a-doumoulakis/tensorflow,theflofly/tensorflow,jhseu/tensorflow,ArtsiomCh/tensorflow,ppries/tensorflow,jart/tensorflow,llhe/tensorflow,dhalleine/tensorflow,adit-chandra/tensorflow,DavidNorman/tensorflow,HKUST-SING/tensorflow,rdipietro/tensorflow,ppries/tensorflow,nolanliou/tensorflow,benoitsteiner/tensorflow,ghchinoy/tensorflow,sarvex/tensorflow,seanli9jan/tensorflow,Carmezim/tensorflow,chris-chris/tensorflow,davidzchen/tensorflow,eaplatanios/tensorflow,pavelchristof/gomoku-ai,thjashin/tensorflow,mortada/tensorflow,neilhan/tensorflow,Xeralux/tensorflow,mdrumond/tensorflow,paolodedios/tensorflow,cancan101/tensorflow,jhaux/tensorflow,hehongliang/tensorflow,dendisuhubdy/tensorflow,Bismarrck/tensorflow,renyi533/tensorflow,renyi533/tensorflow,tillahoffmann/tensorflow,apark263/tensorflow,chenjun0210/tensorflow,anilmuthineni/tensorflow,Intel-Corporation/tensorflow,theflofly/tensorflow,manipopopo/tensorflow,manjunaths/tensorflow,alshedivat/tensorflow,nightjean/Deep-Learning,lakshayg/tensorflow,kobejean/tensorflow,cg31/tensorflow,handroissuazo/tensorflow,xzturn/tensorflow,tensorflow/tensorflow-experimental_link_static_libraries_once,Carmezim/tensorflow,aldian/tensorflow,panmari/tensorflow,frreiss/tensorflow-fred,odejesush/tensorflow,tillahoffmann/tensorflow,ppwwyyxx/tensorflow,Moriadry/tensorflow,xzturn/tensorflow,ravindrapanda/tensorflow,jendap/tensorflow,jart/tensorflow,paolodedios/tensorflow,manjunaths/tensorflow,yufengg/tensorflow,johndpope/tensorflow,jhaux/tensorflow,caisq/tensorflow,jalexvig/tensorflow,annarev/tensorflow,naturali/tensorflow,handroissuazo/tensorflow,mrry/tensorflow,jwlawson/tensorflow,strint/tensorflow,maciekcc/tensorflow,nolanliou/tensorflow,meteorcloudy/tensorflow,ageron/tensorflow,lakshayg/tensorflow,odejesush/tensorflow,adit-chandra/tensorflow,sarvex/tensorflow,tensorflow/tensorflow-experimental_link_static_libraries_once,yongtang/tensorflow,karllessard/tensorflow,DCSaunders/tensorflow,snnn/tensorflow,yufengg/tensorflow,Bismarrck/tensorflow,scenarios/tensorflow,chenjun0210/tensorflow,LUTAN/tensorflow,thjashin/tensorflow,tongwang01/tensorflow,awni/tensorflow,ville-k/tensorflow,Kongsea/tensorflow,code-sauce/tensorflow,ran5515/DeepDecision,nanditav/15712-TensorFlow,apark263/tensorflow,AndreasMadsen/tensorflow,seaotterman/tensorflow,tomasreimers/tensorflow-emscripten,RapidApplicationDevelopment/tensorflow,nburn42/tensorflow,ghchinoy/tensorflow,bowang/tensorflow,alshedivat/tensorflow,ArtsiomCh/tensorflow,alivecor/tensorflow,ran5515/DeepDecision,dendisuhubdy/tensorflow,TakayukiSakai/tensorflow,Moriadry/tensorflow,karllessard/tensorflow,XueqingLin/tensorflow,tensorflow/tensorflow-experimental_link_static_libraries_once,manipopopo/tensorflow,AndreasMadsen/tensorflow,unsiloai/syntaxnet-ops-hack,ninotoshi/tensorflow,MostafaGazar/tensorflow,jwlawson/tensorflow,manjunaths/tensorflow,ZhangXinNan/tensorflow,nanditav/15712-TensorFlow,tomasreimers/tensorflow-emscripten,tiagofrepereira2012/tensorflow,panmari/tensorflow,abhitopia/tensorflow,kobejean/tensorflow,DavidNorman/tensorflow,dongjoon-hyun/tensorflow,strint/tensorflow,pierreg/tensorflow,tensorflow/tensorflow,horance-liu/tensorflow,drpngx/tensorflow,dyoung418/tensorflow,tornadozou/tensorflow,nburn42/tensorflow,XueqingLin/tensorflow,drpngx/tensorflow,Mazecreator/tensorflow,Intel-Corporation/tensorflow,abhitopia/tensorflow,cxxgtxy/tensorflow,aselle/tensorflow,asimshankar/tensorflow,Bulochkin/tensorflow_pack,seaotterman/tensorflow,jhseu/tensorflow,seanli9jan/tensorflow,zasdfgbnm/tensorflow,MoamerEncsConcordiaCa/tensorflow,tntnatbry/tensorflow,girving/tensorflow,MostafaGazar/tensorflow,cancan101/tensorflow,calebfoss/tensorflow,tensorflow/tensorflow,ghchinoy/tensorflow,with-git/tensorflow,haeusser/tensorflow,maciekcc/tensorflow,krikru/tensorflow-opencl,ZhangXinNan/tensorflow,codrut3/tensorflow,ychfan/tensorflow,hehongliang/tensorflow,pierreg/tensorflow,bowang/tensorflow,girving/tensorflow,girving/tensorflow,Intel-tensorflow/tensorflow,theflofly/tensorflow,manazhao/tf_recsys,admcrae/tensorflow,chenjun0210/tensorflow,jendap/tensorflow,benoitsteiner/tensorflow-opencl,ville-k/tensorflow,chenjun0210/tensorflow,thjashin/tensorflow,lukeiwanski/tensorflow,arborh/tensorflow,admcrae/tensorflow,Mistobaan/tensorflow,cancan101/tensorflow,ageron/tensorflow,dongjoon-hyun/tensorflow,hehongliang/tensorflow,vrv/tensorflow,odejesush/tensorflow,MostafaGazar/tensorflow,alistairlow/tensorflow,eerwitt/tensorflow,eaplatanios/tensorflow,xodus7/tensorflow,kamcpp/tensorflow,alistairlow/tensorflow,dancingdan/tensorflow,mengxn/tensorflow,rdipietro/tensorflow,ibmsoe/tensorflow,codrut3/tensorflow,drpngx/tensorflow,xzturn/tensorflow,jeffzheng1/tensorflow,HaebinShin/tensorflow,bowang/tensorflow,yongtang/tensorflow,eerwitt/tensorflow,ville-k/tensorflow,lukeiwanski/tensorflow,ibab/tensorflow,Carmezim/tensorflow,kchodorow/tensorflow,laszlocsomor/tensorflow,juharris/tensorflow,Intel-Corporation/tensorflow,guschmue/tensorflow,nightjean/Deep-Learning,gojira/tensorflow,zasdfgbnm/tensorflow,karllessard/tensorflow,Xeralux/tensorflow,frreiss/tensorflow-fred,tntnatbry/tensorflow,dendisuhubdy/tensorflow,ghchinoy/tensorflow,alistairlow/tensorflow,XueqingLin/tensorflow,Intel-tensorflow/tensorflow,taknevski/tensorflow-xsmm,suiyuan2009/tensorflow,nburn42/tensorflow,ravindrapanda/tensorflow,nightjean/Deep-Learning,brchiu/tensorflow,hsaputra/tensorflow,handroissuazo/tensorflow,wchan/tensorflow,horance-liu/tensorflow,tntnatbry/tensorflow,benoitsteiner/tensorflow-opencl,jostep/tensorflow,yanchen036/tensorflow,scenarios/tensorflow,kevin-coder/tensorflow-fork,paolodedios/tensorflow,pcm17/tensorflow,annarev/tensorflow,wangyum/tensorflow,gojira/tensorflow,calebfoss/tensorflow,asimshankar/tensorflow,nikste/tensorflow,martinwicke/tensorflow,jwlawson/tensorflow,jbedorf/tensorflow,alivecor/tensorflow,llhe/tensorflow,av8ramit/tensorflow,benoitsteiner/tensorflow-opencl,XueqingLin/tensorflow,ravindrapanda/tensorflow,vrv/tensorflow,ivano666/tensorflow,mdrumond/tensorflow,pierreg/tensorflow,vrv/tensorflow,Bulochkin/tensorflow_pack,seaotterman/tensorflow,alistairlow/tensorflow,hfp/tensorflow-xsmm,martinwicke/tensorflow,raymondxyang/tensorflow,mavenlin/tensorflow,adit-chandra/tensorflow,Bulochkin/tensorflow_pack,chemelnucfin/tensorflow,calebfoss/tensorflow,alheinecke/tensorflow-xsmm,yaroslavvb/tensorflow,ghchinoy/tensorflow,neilhan/tensorflow,nikste/tensorflow,anand-c-goog/tensorflow,memo/tensorflow,krikru/tensorflow-opencl,alheinecke/tensorflow-xsmm,adamtiger/tensorflow,jalexvig/tensorflow,frreiss/tensorflow-fred,kamcpp/tensorflow,aldian/tensorflow,ville-k/tensorflow,pavelchristof/gomoku-ai,Intel-tensorflow/tensorflow,alivecor/tensorflow,allenlavoie/tensorflow,av8ramit/tensorflow,petewarden/tensorflow_makefile,memo/tensorflow,sandeepgupta2k4/tensorflow,alisidd/tensorflow,manazhao/tf_recsys,jart/tensorflow,paolodedios/tensorflow,XueqingLin/tensorflow,haeusser/tensorflow,jwlawson/tensorflow,krikru/tensorflow-opencl,jwlawson/tensorflow,av8ramit/tensorflow,chenjun0210/tensorflow,caisq/tensorflow,lukeiwanski/tensorflow-opencl,cg31/tensorflow,ninotoshi/tensorflow,ravindrapanda/tensorflow,nikste/tensorflow,mengxn/tensorflow,davidzchen/tensorflow,jeffzheng1/tensorflow,Bulochkin/tensorflow_pack,ppwwyyxx/tensorflow,eerwitt/tensorflow,zycdragonball/tensorflow,snnn/tensorflow,tensorflow/tensorflow-pywrap_tf_optimizer,dancingdan/tensorflow,raymondxyang/tensorflow,eerwitt/tensorflow,mixturemodel-flow/tensorflow,ageron/tensorflow,kevin-coder/tensorflow-fork,yufengg/tensorflow,zasdfgbnm/tensorflow,manipopopo/tensorflow,ppwwyyxx/tensorflow,krikru/tensorflow-opencl,allenlavoie/tensorflow,HaebinShin/tensorflow,anilmuthineni/tensorflow,ageron/tensorflow,jart/tensorflow,dhalleine/tensorflow,petewarden/tensorflow_makefile,laszlocsomor/tensorflow,tensorflow/skflow,JVillella/tensorflow,davidzchen/tensorflow,odejesush/tensorflow,lukeiwanski/tensorflow-opencl,kamcpp/tensorflow,jalexvig/tensorflow,anand-c-goog/tensorflow,eadgarchen/tensorflow,eaplatanios/tensorflow,manipopopo/tensorflow,nburn42/tensorflow,jwlawson/tensorflow,manjunaths/tensorflow,ychfan/tensorflow,AnishShah/tensorflow,juharris/tensorflow,jeffzheng1/tensorflow,bowang/tensorflow,Mistobaan/tensorflow,Xeralux/tensorflow,neilhan/tensorflow,laszlocsomor/tensorflow,nanditav/15712-TensorFlow,hfp/tensorflow-xsmm,tillahoffmann/tensorflow,HaebinShin/tensorflow,naturali/tensorflow,markslwong/tensorflow,renyi533/tensorflow,nightjean/Deep-Learning,sandeepdsouza93/TensorFlow-15712,odejesush/tensorflow,HaebinShin/tensorflow,ZhangXinNan/tensorflow,haeusser/tensorflow,suiyuan2009/tensorflow,allenlavoie/tensorflow,codrut3/tensorflow,nikste/tensorflow,gojira/tensorflow,JVillella/tensorflow,paolodedios/tensorflow,sandeepdsouza93/TensorFlow-15712,nburn42/tensorflow,SnakeJenny/TensorFlow,gautam1858/tensorflow,sarvex/tensorflow,TakayukiSakai/tensorflow,yongtang/tensorflow,apark263/tensorflow,petewarden/tensorflow_makefile,av8ramit/tensorflow,admcrae/tensorflow,tomasreimers/tensorflow-emscripten,martinwicke/tensorflow,gautam1858/tensorflow,JingJunYin/tensorflow,aam-at/tensorflow,strint/tensorflow,dendisuhubdy/tensorflow,tensorflow/tensorflow,tensorflow/tensorflow-experimental_link_static_libraries_once,benoitsteiner/tensorflow-xsmm,annarev/tensorflow,tillahoffmann/tensorflow,Moriadry/tensorflow,dongjoon-hyun/tensorflow,kchodorow/tensorflow,EvenStrangest/tensorflow,mrry/tensorflow,yufengg/tensorflow,pierreg/tensorflow,anilmuthineni/tensorflow,maciekcc/tensorflow,renyi533/tensorflow,vrv/tensorflow,maciekcc/tensorflow,Carmezim/tensorflow,anilmuthineni/tensorflow,tensorflow/tensorflow-pywrap_tf_optimizer,elingg/tensorflow,sjperkins/tensorflow,lukeiwanski/tensorflow-opencl,laosiaudi/tensorflow,cg31/tensorflow,ArtsiomCh/tensorflow,asadziach/tensorflow,alistairlow/tensorflow,ghchinoy/tensorflow,lukeiwanski/tensorflow,with-git/tensorflow,EvenStrangest/tensorflow,kobejean/tensorflow,sandeepdsouza93/TensorFlow-15712,pavelchristof/gomoku-ai,vrv/tensorflow,ravindrapanda/tensorflow,aam-at/tensorflow,jalexvig/tensorflow,tongwang01/tensorflow,Bismarrck/tensorflow,jhseu/tensorflow,Intel-tensorflow/tensorflow,seaotterman/tensorflow,meteorcloudy/tensorflow,ishay2b/tensorflow,jendap/tensorflow,tongwang01/tensorflow,apark263/tensorflow,EvenStrangest/tensorflow,horance-liu/tensorflow,xzturn/tensorflow,guschmue/tensorflow,bowang/tensorflow,adit-chandra/tensorflow,jhaux/tensorflow,ageron/tensorflow,andrewcmyers/tensorflow,yanchen036/tensorflow,gunan/tensorflow,handroissuazo/tensorflow,JingJunYin/tensorflow,ibmsoe/tensorflow,Intel-Corporation/tensorflow,ravindrapanda/tensorflow,mavenlin/tensorflow,rdipietro/tensorflow,jhaux/tensorflow,mdrumond/tensorflow,tongwang01/tensorflow,petewarden/tensorflow,zycdragonball/tensorflow,ivano666/tensorflow,AnishShah/tensorflow,alsrgv/tensorflow,ppries/tensorflow,Xeralux/tensorflow,mrry/tensorflow,tensorflow/tensorflow-pywrap_saved_model,gunan/tensorflow,benoitsteiner/tensorflow-opencl,kobejean/tensorflow,Moriadry/tensorflow,gojira/tensorflow,JVillella/tensorflow,alheinecke/tensorflow-xsmm,renyi533/tensorflow,xzturn/tensorflow,tiagofrepereira2012/tensorflow,tongwang01/tensorflow,karllessard/tensorflow,xodus7/tensorflow,nanditav/15712-TensorFlow,freedomtan/tensorflow,ZhangXinNan/tensorflow,rabipanda/tensorflow,apark263/tensorflow,gnieboer/tensorflow,RapidApplicationDevelopment/tensorflow,Mistobaan/tensorflow,meteorcloudy/tensorflow,Moriadry/tensorflow,zasdfgbnm/tensorflow,ageron/tensorflow,kobejean/tensorflow,chris-chris/tensorflow,tillahoffmann/tensorflow,yaroslavvb/tensorflow,ran5515/DeepDecision,alheinecke/tensorflow-xsmm,mrry/tensorflow,laszlocsomor/tensorflow,apark263/tensorflow,nikste/tensorflow,sarvex/tensorflow,jbedorf/tensorflow,XueqingLin/tensorflow,markslwong/tensorflow,alheinecke/tensorflow-xsmm,anand-c-goog/tensorflow,allenlavoie/tensorflow,johndpope/tensorflow,karllessard/tensorflow,cg31/tensorflow,bowang/tensorflow,DCSaunders/tensorflow,ran5515/DeepDecision,horance-liu/tensorflow,MoamerEncsConcordiaCa/tensorflow,lukeiwanski/tensorflow,seanli9jan/tensorflow,ibab/tensorflow,anand-c-goog/tensorflow,neilhan/tensorflow,hfp/tensorflow-xsmm,DavidNorman/tensorflow,mengxn/tensorflow,cancan101/tensorflow,kamcpp/tensorflow,pavelchristof/gomoku-ai,renyi533/tensorflow,lukeiwanski/tensorflow-opencl,andrewcmyers/tensorflow,with-git/tensorflow,JVillella/tensorflow,mengxn/tensorflow,admcrae/tensorflow,paolodedios/tensorflow,xodus7/tensorflow,peterbraden/tensorflow,freedomtan/tensorflow,paolodedios/tensorflow,freedomtan/tensorflow,jeffzheng1/tensorflow,snnn/tensorflow,cg31/tensorflow,jhseu/tensorflow,nolanliou/tensorflow,manipopopo/tensorflow,ibmsoe/tensorflow,DavidNorman/tensorflow,tensorflow/tensorflow-pywrap_saved_model,nightjean/Deep-Learning,brchiu/tensorflow,gunan/tensorflow,lakshayg/tensorflow,eerwitt/tensorflow,xodus7/tensorflow,adit-chandra/tensorflow,zasdfgbnm/tensorflow,allenlavoie/tensorflow,ychfan/tensorflow,dancingdan/tensorflow,frreiss/tensorflow-fred,jbedorf/tensorflow,tornadozou/tensorflow,gojira/tensorflow,jbedorf/tensorflow,naturali/tensorflow,caisq/tensorflow,ninotoshi/tensorflow,lakshayg/tensorflow,hfp/tensorflow-xsmm,nolanliou/tensorflow,laosiaudi/tensorflow,caisq/tensorflow,suiyuan2009/tensorflow,jendap/tensorflow,LUTAN/tensorflow,jbedorf/tensorflow,memo/tensorflow,xzturn/tensorflow,Intel-tensorflow/tensorflow,gibiansky/tensorflow,kchodorow/tensorflow,laszlocsomor/tensorflow,with-git/tensorflow,tornadozou/tensorflow,rabipanda/tensorflow,memo/tensorflow,alsrgv/tensorflow,tntnatbry/tensorflow,sjperkins/tensorflow,whn09/tensorflow,neilhan/tensorflow,gautam1858/tensorflow,Bismarrck/tensorflow,mengxn/tensorflow,pierreg/tensorflow,kchodorow/tensorflow,with-git/tensorflow,llhe/tensorflow,wchan/tensorflow,ZhangXinNan/tensorflow,jostep/tensorflow,calebfoss/tensorflow,davidzchen/tensorflow,yanchen036/tensorflow,guschmue/tensorflow,anand-c-goog/tensorflow,meteorcloudy/tensorflow,awni/tensorflow,ville-k/tensorflow,jeffzheng1/tensorflow,jostep/tensorflow,eadgarchen/tensorflow,mengxn/tensorflow,Bismarrck/tensorflow,yongtang/tensorflow,nightjean/Deep-Learning,zasdfgbnm/tensorflow,ville-k/tensorflow,yanchen036/tensorflow,tomasreimers/tensorflow-emscripten,unsiloai/syntaxnet-ops-hack,manjunaths/tensorflow,admcrae/tensorflow,memo/tensorflow,dhalleine/tensorflow,brchiu/tensorflow,tntnatbry/tensorflow,krikru/tensorflow-opencl,benoitsteiner/tensorflow-xsmm,abhitopia/tensorflow,asimshankar/tensorflow,nanditav/15712-TensorFlow,jostep/tensorflow,chemelnucfin/tensorflow,kevin-coder/tensorflow-fork,XueqingLin/tensorflow,naturali/tensorflow,theflofly/tensorflow,Mistobaan/tensorflow,ran5515/DeepDecision,alshedivat/tensorflow,adamtiger/tensorflow,ageron/tensorflow,tntnatbry/tensorflow,eadgarchen/tensorflow,Mazecreator/tensorflow,mavenlin/tensorflow,seanli9jan/tensorflow,abhitopia/tensorflow,tensorflow/tensorflow-pywrap_saved_model,Xeralux/tensorflow,whn09/tensorflow,manipopopo/tensorflow,HKUST-SING/tensorflow,cxxgtxy/tensorflow,MostafaGazar/tensorflow,ychfan/tensorflow,martinwicke/tensorflow,elingg/tensorflow,johndpope/tensorflow,cancan101/tensorflow,abhitopia/tensorflow,girving/tensorflow,AnishShah/tensorflow,mortada/tensorflow,Bismarrck/tensorflow,alisidd/tensorflow,manazhao/tf_recsys,seaotterman/tensorflow,renyi533/tensorflow,asadziach/tensorflow,a-doumoulakis/tensorflow,mortada/tensorflow,gautam1858/tensorflow,tiagofrepereira2012/tensorflow,alsrgv/tensorflow,seanli9jan/tensorflow,AnishShah/tensorflow,snnn/tensorflow,codrut3/tensorflow,a-doumoulakis/tensorflow,rdipietro/tensorflow,tomasreimers/tensorflow-emscripten,benoitsteiner/tensorflow-xsmm,jhaux/tensorflow,elingg/tensorflow,laosiaudi/tensorflow,aselle/tensorflow,gunan/tensorflow,krikru/tensorflow-opencl,gnieboer/tensorflow,xodus7/tensorflow,snnn/tensorflow,petewarden/tensorflow,sjperkins/tensorflow,petewarden/tensorflow,naturali/tensorflow,hfp/tensorflow-xsmm,abhitopia/tensorflow,DavidNorman/tensorflow,llhe/tensorflow,annarev/tensorflow,yongtang/tensorflow,laosiaudi/tensorflow,llhe/tensorflow,anand-c-goog/tensorflow,wchan/tensorflow,hsaputra/tensorflow,yongtang/tensorflow,hfp/tensorflow-xsmm,mdrumond/tensorflow,chenjun0210/tensorflow,TakayukiSakai/tensorflow,whn09/tensorflow,seaotterman/tensorflow,gibiansky/tensorflow,hsaputra/tensorflow,Mistobaan/tensorflow,allenlavoie/tensorflow,hfp/tensorflow-xsmm,adit-chandra/tensorflow,maciekcc/tensorflow,eaplatanios/tensorflow,neilhan/tensorflow,DavidNorman/tensorflow,peterbraden/tensorflow,mrry/tensorflow,gunan/tensorflow,xzturn/tensorflow,JVillella/tensorflow,AnishShah/tensorflow,jart/tensorflow,scenarios/tensorflow,taknevski/tensorflow-xsmm,RapidApplicationDevelopment/tensorflow,mixturemodel-flow/tensorflow,taknevski/tensorflow-xsmm,lukeiwanski/tensorflow,a-doumoulakis/tensorflow,jostep/tensorflow,panmari/tensorflow,nikste/tensorflow,tensorflow/tensorflow-pywrap_saved_model,ychfan/tensorflow,ZhangXinNan/tensorflow,cancan101/tensorflow,thesuperzapper/tensorflow,awni/tensorflow,awni/tensorflow,ran5515/DeepDecision,aldian/tensorflow,aam-at/tensorflow,alsrgv/tensorflow,maciekcc/tensorflow,lakshayg/tensorflow,Moriadry/tensorflow,kchodorow/tensorflow,ghchinoy/tensorflow,benoitsteiner/tensorflow-xsmm,frreiss/tensorflow-fred,dongjoon-hyun/tensorflow,karllessard/tensorflow,handroissuazo/tensorflow,mortada/tensorflow,apark263/tensorflow,ppries/tensorflow,hfp/tensorflow-xsmm,arborh/tensorflow,Mistobaan/tensorflow,nightjean/Deep-Learning,ninotoshi/tensorflow,frreiss/tensorflow-fred,ibmsoe/tensorflow,kevin-coder/tensorflow-fork,codrut3/tensorflow,raymondxyang/tensorflow,Carmezim/tensorflow,eaplatanios/tensorflow,tensorflow/tensorflow-experimental_link_static_libraries_once,gunan/tensorflow,codrut3/tensorflow,raymondxyang/tensorflow,wangyum/tensorflow,rabipanda/tensorflow,av8ramit/tensorflow,chemelnucfin/tensorflow,ZhangXinNan/tensorflow,guschmue/tensorflow,tiagofrepereira2012/tensorflow,annarev/tensorflow,whn09/tensorflow,MoamerEncsConcordiaCa/tensorflow,ppries/tensorflow,jwlawson/tensorflow,benoitsteiner/tensorflow,manjunaths/tensorflow,Intel-tensorflow/tensorflow,eerwitt/tensorflow,eerwitt/tensorflow,RapidApplicationDevelopment/tensorflow,ishay2b/tensorflow,chris-chris/tensorflow,andrewcmyers/tensorflow,ychfan/tensorflow,sandeepdsouza93/TensorFlow-15712,laosiaudi/tensorflow,zasdfgbnm/tensorflow,ivano666/tensorflow,jalexvig/tensorflow,sandeepdsouza93/TensorFlow-15712,girving/tensorflow,dancingdan/tensorflow,karllessard/tensorflow,arborh/tensorflow,benoitsteiner/tensorflow-opencl,HKUST-SING/tensorflow,sarvex/tensorflow,jbedorf/tensorflow,gautam1858/tensorflow,handroissuazo/tensorflow,jostep/tensorflow,code-sauce/tensorflow,johndpope/tensorflow,alshedivat/tensorflow,ville-k/tensorflow,jendap/tensorflow,sjperkins/tensorflow,dendisuhubdy/tensorflow,drpngx/tensorflow,unsiloai/syntaxnet-ops-hack,laszlocsomor/tensorflow,chris-chris/tensorflow,zasdfgbnm/tensorflow,Intel-Corporation/tensorflow,juharris/tensorflow,yufengg/tensorflow,cxxgtxy/tensorflow,pcm17/tensorflow,yufengg/tensorflow,Intel-tensorflow/tensorflow,Mistobaan/tensorflow,yufengg/tensorflow,jeffzheng1/tensorflow,nolanliou/tensorflow,tornadozou/tensorflow,tiagofrepereira2012/tensorflow,XueqingLin/tensorflow,zasdfgbnm/tensorflow,ville-k/tensorflow,frreiss/tensorflow-fred,yongtang/tensorflow,jalexvig/tensorflow,aselle/tensorflow,chris-chris/tensorflow,hfp/tensorflow-xsmm,MoamerEncsConcordiaCa/tensorflow,tensorflow/tensorflow-pywrap_saved_model,mengxn/tensorflow,apark263/tensorflow,awni/tensorflow,ninotoshi/tensorflow,AnishShah/tensorflow,suiyuan2009/tensorflow,benoitsteiner/tensorflow-xsmm,gojira/tensorflow,yaroslavvb/tensorflow,adit-chandra/tensorflow,xzturn/tensorflow,whn09/tensorflow,aldian/tensorflow,brchiu/tensorflow,mortada/tensorflow,manjunaths/tensorflow,guschmue/tensorflow,gibiansky/tensorflow,theflofly/tensorflow,Mazecreator/tensorflow,Mazecreator/tensorflow,seaotterman/tensorflow,eadgarchen/tensorflow,sjperkins/tensorflow,manazhao/tf_recsys,ppwwyyxx/tensorflow,meteorcloudy/tensorflow,ArtsiomCh/tensorflow,thesuperzapper/tensorflow,martinwicke/tensorflow,mdrumond/tensorflow,panmari/tensorflow,nburn42/tensorflow,chris-chris/tensorflow,ZhangXinNan/tensorflow,JingJunYin/tensorflow,elingg/tensorflow,a-doumoulakis/tensorflow,manipopopo/tensorflow,arborh/tensorflow,rdipietro/tensorflow,dongjoon-hyun/tensorflow,naturali/tensorflow,SnakeJenny/TensorFlow,ghchinoy/tensorflow,asimshankar/tensorflow,benoitsteiner/tensorflow,kamcpp/tensorflow,thjashin/tensorflow,kevin-coder/tensorflow-fork,AndreasMadsen/tensorflow,petewarden/tensorflow_makefile,unsiloai/syntaxnet-ops-hack,Bismarrck/tensorflow,aselle/tensorflow,jhaux/tensorflow,petewarden/tensorflow,freedomtan/tensorflow,anilmuthineni/tensorflow,pcm17/tensorflow,tensorflow/tensorflow-pywrap_tf_optimizer,horance-liu/tensorflow,MostafaGazar/tensorflow,JingJunYin/tensorflow,nikste/tensorflow,aam-at/tensorflow,MoamerEncsConcordiaCa/tensorflow,hehongliang/tensorflow,davidzchen/tensorflow,pcm17/tensorflow,jendap/tensorflow,awni/tensorflow,Kongsea/tensorflow,Bulochkin/tensorflow_pack,tensorflow/tensorflow,asadziach/tensorflow,pcm17/tensorflow,laosiaudi/tensorflow,johndpope/tensorflow,DavidNorman/tensorflow,Bismarrck/tensorflow,allenlavoie/tensorflow,dyoung418/tensorflow,zycdragonball/tensorflow,HKUST-SING/tensorflow,with-git/tensorflow,naturali/tensorflow,tiagofrepereira2012/tensorflow,juharris/tensorflow,benoitsteiner/tensorflow-xsmm,mixturemodel-flow/tensorflow,thesuperzapper/tensorflow,alshedivat/tensorflow,mortada/tensorflow,benoitsteiner/tensorflow-opencl,tensorflow/tensorflow,MycChiu/tensorflow,jhseu/tensorflow,markslwong/tensorflow,taknevski/tensorflow-xsmm,Bulochkin/tensorflow_pack,MostafaGazar/tensorflow,eaplatanios/tensorflow,seaotterman/tensorflow,petewarden/tensorflow,sandeepgupta2k4/tensorflow,memo/tensorflow,Kongsea/tensorflow,mixturemodel-flow/tensorflow,vrv/tensorflow,rabipanda/tensorflow,renyi533/tensorflow,alisidd/tensorflow,frreiss/tensorflow-fred,tiagofrepereira2012/tensorflow,frreiss/tensorflow-fred,guschmue/tensorflow,adamtiger/tensorflow,ArtsiomCh/tensorflow,markslwong/tensorflow,alsrgv/tensorflow,MoamerEncsConcordiaCa/tensorflow,ishay2b/tensorflow,memo/tensorflow,Xeralux/tensorflow,dendisuhubdy/tensorflow,dhalleine/tensorflow,scenarios/tensorflow,drpngx/tensorflow,yaroslavvb/tensorflow,ppwwyyxx/tensorflow,tensorflow/tensorflow-pywrap_saved_model,juharris/tensorflow,Xeralux/tensorflow,unsiloai/syntaxnet-ops-hack,guschmue/tensorflow,handroissuazo/tensorflow,TakayukiSakai/tensorflow,rabipanda/tensorflow,aselle/tensorflow,odejesush/tensorflow,vrv/tensorflow,jart/tensorflow,nolanliou/tensorflow,elingg/tensorflow,eaplatanios/tensorflow,tensorflow/tensorflow-pywrap_tf_optimizer,LUTAN/tensorflow,manjunaths/tensorflow,aselle/tensorflow,Kongsea/tensorflow,laszlocsomor/tensorflow,HaebinShin/tensorflow,code-sauce/tensorflow,codrut3/tensorflow,gautam1858/tensorflow,AndreasMadsen/tensorflow,ppries/tensorflow,cxxgtxy/tensorflow,horance-liu/tensorflow,alisidd/tensorflow,nburn42/tensorflow,xzturn/tensorflow,arborh/tensorflow,HKUST-SING/tensorflow,rdipietro/tensorflow,tensorflow/tensorflow-experimental_link_static_libraries_once,jalexvig/tensorflow,bowang/tensorflow,dancingdan/tensorflow,sandeepgupta2k4/tensorflow,benoitsteiner/tensorflow-xsmm,martinwicke/tensorflow,karllessard/tensorflow,HKUST-SING/tensorflow,EvenStrangest/tensorflow,aam-at/tensorflow,sandeepdsouza93/TensorFlow-15712,hehongliang/tensorflow,tensorflow/tensorflow,HKUST-SING/tensorflow,petewarden/tensorflow,Intel-Corporation/tensorflow,scenarios/tensorflow,Intel-Corporation/tensorflow,freedomtan/tensorflow,tensorflow/tensorflow-experimental_link_static_libraries_once,anilmuthineni/tensorflow,aam-at/tensorflow,dongjoon-hyun/tensorflow,caisq/tensorflow,cg31/tensorflow,girving/tensorflow,gojira/tensorflow,EvenStrangest/tensorflow,sjperkins/tensorflow,RapidApplicationDevelopment/tensorflow,benoitsteiner/tensorflow-xsmm,peterbraden/tensorflow,pavelchristof/gomoku-ai,paolodedios/tensorflow,pavelchristof/gomoku-ai,jalexvig/tensorflow,admcrae/tensorflow,kobejean/tensorflow,johndpope/tensorflow,mixturemodel-flow/tensorflow,dyoung418/tensorflow,a-doumoulakis/tensorflow,alsrgv/tensorflow,adit-chandra/tensorflow,chemelnucfin/tensorflow,gibiansky/tensorflow,tensorflow/tensorflow,freedomtan/tensorflow,rdipietro/tensorflow,meteorcloudy/tensorflow,laszlocsomor/tensorflow,caisq/tensorflow,ArtsiomCh/tensorflow,ghchinoy/tensorflow,allenlavoie/tensorflow,ppwwyyxx/tensorflow,dongjoon-hyun/tensorflow,andrewcmyers/tensorflow,nburn42/tensorflow,gunan/tensorflow,renyi533/tensorflow,arborh/tensorflow,Mazecreator/tensorflow,lukeiwanski/tensorflow,calebfoss/tensorflow,nolanliou/tensorflow,theflofly/tensorflow,neilhan/tensorflow,drpngx/tensorflow,suiyuan2009/tensorflow,MycChiu/tensorflow,LUTAN/tensorflow,cxxgtxy/tensorflow,gibiansky/tensorflow,odejesush/tensorflow,Kongsea/tensorflow,girving/tensorflow,juharris/tensorflow,jhaux/tensorflow,av8ramit/tensorflow,JingJunYin/tensorflow,alshedivat/tensorflow,haeusser/tensorflow,JingJunYin/tensorflow,jart/tensorflow,kchodorow/tensorflow,admcrae/tensorflow,gautam1858/tensorflow,MostafaGazar/tensorflow,petewarden/tensorflow_makefile,hehongliang/tensorflow,code-sauce/tensorflow,zasdfgbnm/tensorflow,juharris/tensorflow,gautam1858/tensorflow,adamtiger/tensorflow,chemelnucfin/tensorflow,lakshayg/tensorflow,ville-k/tensorflow,benoitsteiner/tensorflow,eaplatanios/tensorflow,sandeepgupta2k4/tensorflow,zycdragonball/tensorflow,mavenlin/tensorflow,meteorcloudy/tensorflow,krikru/tensorflow-opencl,alsrgv/tensorflow,adit-chandra/tensorflow,asadziach/tensorflow,dhalleine/tensorflow,kamcpp/tensorflow,johndpope/tensorflow,peterbraden/tensorflow,SnakeJenny/TensorFlow,peterbraden/tensorflow,eaplatanios/tensorflow,kevin-coder/tensorflow-fork,strint/tensorflow,av8ramit/tensorflow,panmari/tensorflow,chris-chris/tensorflow,meteorcloudy/tensorflow,SnakeJenny/TensorFlow,jbedorf/tensorflow,tillahoffmann/tensorflow,with-git/tensorflow,seanli9jan/tensorflow,sarvex/tensorflow,thjashin/tensorflow,ibmsoe/tensorflow,MostafaGazar/tensorflow,rabipanda/tensorflow,unsiloai/syntaxnet-ops-hack,ghchinoy/tensorflow,kobejean/tensorflow,asadziach/tensorflow,AnishShah/tensorflow,ageron/tensorflow,JingJunYin/tensorflow,ninotoshi/tensorflow,Mazecreator/tensorflow,yanchen036/tensorflow,calebfoss/tensorflow,alistairlow/tensorflow,calebfoss/tensorflow,yaroslavvb/tensorflow,panmari/tensorflow,wchan/tensorflow,sandeepgupta2k4/tensorflow,markslwong/tensorflow,manjunaths/tensorflow,davidzchen/tensorflow,Bulochkin/tensorflow_pack,yanchen036/tensorflow,theflofly/tensorflow,peterbraden/tensorflow,eaplatanios/tensorflow,gnieboer/tensorflow,alheinecke/tensorflow-xsmm,girving/tensorflow,gibiansky/tensorflow,Mazecreator/tensorflow,ArtsiomCh/tensorflow,tongwang01/tensorflow,eadgarchen/tensorflow,girving/tensorflow,sarvex/tensorflow,tiagofrepereira2012/tensorflow,jeffzheng1/tensorflow,ishay2b/tensorflow,yongtang/tensorflow,kobejean/tensorflow,frreiss/tensorflow-fred,chemelnucfin/tensorflow,ZhangXinNan/tensorflow,HaebinShin/tensorflow,drpngx/tensorflow,alisidd/tensorflow,DCSaunders/tensorflow,neilhan/tensorflow,xodus7/tensorflow,annarev/tensorflow,juharris/tensorflow,naturali/tensorflow,hsaputra/tensorflow,adamtiger/tensorflow,apark263/tensorflow,Bulochkin/tensorflow_pack,tensorflow/tensorflow-experimental_link_static_libraries_once,code-sauce/tensorflow,mdrumond/tensorflow,MostafaGazar/tensorflow,JVillella/tensorflow,ychfan/tensorflow,thesuperzapper/tensorflow,horance-liu/tensorflow,TakayukiSakai/tensorflow,pcm17/tensorflow,horance-liu/tensorflow,maciekcc/tensorflow,asimshankar/tensorflow,unsiloai/syntaxnet-ops-hack,kchodorow/tensorflow,av8ramit/tensorflow,laszlocsomor/tensorflow,markslwong/tensorflow,SnakeJenny/TensorFlow,asadziach/tensorflow,sandeepgupta2k4/tensorflow,maciekcc/tensorflow,tensorflow/tensorflow,jendap/tensorflow,kamcpp/tensorflow,strint/tensorflow,davidzchen/tensorflow,johndpope/tensorflow,Kongsea/tensorflow,mavenlin/tensorflow,tensorflow/tensorflow,dendisuhubdy/tensorflow,petewarden/tensorflow,tensorflow/tensorflow-pywrap_saved_model,mdrumond/tensorflow,ibmsoe/tensorflow,jhseu/tensorflow,tensorflow/skflow,jhseu/tensorflow,yaroslavvb/tensorflow,alivecor/tensorflow,gunan/tensorflow,TakayukiSakai/tensorflow,laosiaudi/tensorflow,kchodorow/tensorflow,dhalleine/tensorflow,scenarios/tensorflow,ran5515/DeepDecision,aselle/tensorflow,apark263/tensorflow,eadgarchen/tensorflow,AndreasMadsen/tensorflow,Bulochkin/tensorflow_pack,DCSaunders/tensorflow,awni/tensorflow,petewarden/tensorflow,petewarden/tensorflow,benoitsteiner/tensorflow-xsmm,alistairlow/tensorflow,lukeiwanski/tensorflow,brchiu/tensorflow,nanditav/15712-TensorFlow,asimshankar/tensorflow,caisq/tensorflow,Intel-tensorflow/tensorflow,aselle/tensorflow,benoitsteiner/tensorflow,ageron/tensorflow,mrry/tensorflow,panmari/tensorflow,laszlocsomor/tensorflow,nikste/tensorflow,scenarios/tensorflow,rabipanda/tensorflow,gibiansky/tensorflow,abhitopia/tensorflow,dendisuhubdy/tensorflow,aselle/tensorflow,jeffzheng1/tensorflow,freedomtan/tensorflow,pierreg/tensorflow,tornadozou/tensorflow,ivano666/tensorflow,suiyuan2009/tensorflow,zasdfgbnm/tensorflow,jwlawson/tensorflow,paolodedios/tensorflow,Mistobaan/tensorflow,chris-chris/tensorflow,llhe/tensorflow,wchan/tensorflow,manipopopo/tensorflow,rdipietro/tensorflow,petewarden/tensorflow,jart/tensorflow,theflofly/tensorflow,tillahoffmann/tensorflow,tensorflow/tensorflow,raymondxyang/tensorflow,guschmue/tensorflow,tensorflow/tensorflow-pywrap_tf_optimizer,allenlavoie/tensorflow,benoitsteiner/tensorflow,yongtang/tensorflow,DavidNorman/tensorflow,abhitopia/tensorflow,hsaputra/tensorflow,MoamerEncsConcordiaCa/tensorflow,yaroslavvb/tensorflow,theflofly/tensorflow,zycdragonball/tensorflow,xodus7/tensorflow,wangyum/tensorflow,markslwong/tensorflow,AndreasMadsen/tensorflow,yanchen036/tensorflow,memo/tensorflow,odejesush/tensorflow,MycChiu/tensorflow,brchiu/tensorflow,petewarden/tensorflow_makefile,aam-at/tensorflow,dansbecker/skflow,tntnatbry/tensorflow,llhe/tensorflow,sjperkins/tensorflow,alsrgv/tensorflow,ibmsoe/tensorflow,benoitsteiner/tensorflow,petewarden/tensorflow_makefile,rdipietro/tensorflow,manipopopo/tensorflow,gunan/tensorflow,asimshankar/tensorflow,kobejean/tensorflow,mortada/tensorflow,jart/tensorflow,aldian/tensorflow,jostep/tensorflow,thjashin/tensorflow,annarev/tensorflow,sandeepgupta2k4/tensorflow,sjperkins/tensorflow,Moriadry/tensorflow,freedomtan/tensorflow,snnn/tensorflow,Bismarrck/tensorflow,RapidApplicationDevelopment/tensorflow,andrewcmyers/tensorflow,wangyum/tensorflow,brchiu/tensorflow,DCSaunders/tensorflow,haeusser/tensorflow,eadgarchen/tensorflow,aselle/tensorflow,sandeepgupta2k4/tensorflow,renyi533/tensorflow,wchan/tensorflow,elingg/tensorflow,zycdragonball/tensorflow,EvenStrangest/tensorflow,nolanliou/tensorflow,martinwicke/tensorflow,caisq/tensorflow,DCSaunders/tensorflow,arborh/tensorflow,xzturn/tensorflow,thjashin/tensorflow,Bulochkin/tensorflow_pack,tntnatbry/tensorflow,av8ramit/tensorflow,tornadozou/tensorflow,Intel-tensorflow/tensorflow,arborh/tensorflow,wangyum/tensorflow,av8ramit/tensorflow,petewarden/tensorflow,snnn/tensorflow,cancan101/tensorflow,tensorflow/tensorflow-pywrap_tf_optimizer,elingg/tensorflow,taknevski/tensorflow-xsmm,arborh/tensorflow,gojira/tensorflow,benoitsteiner/tensorflow-xsmm,alheinecke/tensorflow-xsmm,llhe/tensorflow,dongjoon-hyun/tensorflow,thjashin/tensorflow,krikru/tensorflow-opencl,brchiu/tensorflow,aam-at/tensorflow,mixturemodel-flow/tensorflow,cg31/tensorflow,ageron/tensorflow,MycChiu/tensorflow,tensorflow/tensorflow-pywrap_tf_optimizer,nightjean/Deep-Learning,Mistobaan/tensorflow,ArtsiomCh/tensorflow,anilmuthineni/tensorflow,ville-k/tensorflow,gautam1858/tensorflow,Intel-Corporation/tensorflow,dendisuhubdy/tensorflow,alistairlow/tensorflow,dancingdan/tensorflow,sjperkins/tensorflow,MycChiu/tensorflow,yongtang/tensorflow,eerwitt/tensorflow,haeusser/tensorflow,mrry/tensorflow,horance-liu/tensorflow,tornadozou/tensorflow,tomasreimers/tensorflow-emscripten,kamcpp/tensorflow,karllessard/tensorflow,dongjoon-hyun/tensorflow,xodus7/tensorflow,codrut3/tensorflow,HaebinShin/tensorflow,hehongliang/tensorflow,strint/tensorflow,code-sauce/tensorflow,tornadozou/tensorflow,RapidApplicationDevelopment/tensorflow,annarev/tensorflow,RapidApplicationDevelopment/tensorflow,aldian/tensorflow,cg31/tensorflow,davidzchen/tensorflow,alshedivat/tensorflow,gunan/tensorflow,alivecor/tensorflow,pierreg/tensorflow,mdrumond/tensorflow,mortada/tensorflow,dhalleine/tensorflow,jendap/tensorflow,thesuperzapper/tensorflow,wangyum/tensorflow,AndreasMadsen/tensorflow,scenarios/tensorflow,alshedivat/tensorflow,RapidApplicationDevelopment/tensorflow,pcm17/tensorflow,ibab/tensorflow,nanditav/15712-TensorFlow,lukeiwanski/tensorflow,allenlavoie/tensorflow,andrewcmyers/tensorflow,jhseu/tensorflow,anand-c-goog/tensorflow,rabipanda/tensorflow,dongjoon-hyun/tensorflow,markslwong/tensorflow,andrewcmyers/tensorflow,mdrumond/tensorflow,HaebinShin/tensorflow,jhaux/tensorflow,alisidd/tensorflow |
import numpy as np
import tensorflow as tf
from skflow import ops
class DropoutTest(tf.test.TestCase):
def test_dropout_float(self):
- with self.test_session():
+ with self.test_session() as session:
x = tf.placeholder(tf.float32, [5, 5])
y = ops.dropout(x, 0.5)
probs = tf.get_collection(ops.DROPOUTS)
+ session.run(tf.initialize_all_variables())
self.assertEqual(len(probs), 1)
+ self.assertEqual(session.run(probs[0]), 0.5)
+
+ def test_dropout_tensor(self):
+ with self.test_session():
+ x = tf.placeholder(tf.float32, [5, 5])
+ y = tf.get_variable("prob", [], initializer=tf.constant_initializer(0.5))
+ z = ops.dropout(x, y)
+ probs = tf.get_collection(ops.DROPOUTS)
+ self.assertEqual(probs, [y])
if __name__ == '__main__':
tf.test.main()
| Test for dropout probability be a tensor | ## Code Before:
import numpy as np
import tensorflow as tf
from skflow import ops
class DropoutTest(tf.test.TestCase):
def test_dropout_float(self):
with self.test_session():
x = tf.placeholder(tf.float32, [5, 5])
y = ops.dropout(x, 0.5)
probs = tf.get_collection(ops.DROPOUTS)
self.assertEqual(len(probs), 1)
if __name__ == '__main__':
tf.test.main()
## Instruction:
Test for dropout probability be a tensor
## Code After:
import numpy as np
import tensorflow as tf
from skflow import ops
class DropoutTest(tf.test.TestCase):
def test_dropout_float(self):
with self.test_session() as session:
x = tf.placeholder(tf.float32, [5, 5])
y = ops.dropout(x, 0.5)
probs = tf.get_collection(ops.DROPOUTS)
session.run(tf.initialize_all_variables())
self.assertEqual(len(probs), 1)
self.assertEqual(session.run(probs[0]), 0.5)
def test_dropout_tensor(self):
with self.test_session():
x = tf.placeholder(tf.float32, [5, 5])
y = tf.get_variable("prob", [], initializer=tf.constant_initializer(0.5))
z = ops.dropout(x, y)
probs = tf.get_collection(ops.DROPOUTS)
self.assertEqual(probs, [y])
if __name__ == '__main__':
tf.test.main()
| // ... existing code ...
def test_dropout_float(self):
with self.test_session() as session:
x = tf.placeholder(tf.float32, [5, 5])
y = ops.dropout(x, 0.5)
probs = tf.get_collection(ops.DROPOUTS)
session.run(tf.initialize_all_variables())
self.assertEqual(len(probs), 1)
self.assertEqual(session.run(probs[0]), 0.5)
def test_dropout_tensor(self):
with self.test_session():
x = tf.placeholder(tf.float32, [5, 5])
y = tf.get_variable("prob", [], initializer=tf.constant_initializer(0.5))
z = ops.dropout(x, y)
probs = tf.get_collection(ops.DROPOUTS)
self.assertEqual(probs, [y])
// ... rest of the code ... |
b7b691d82accc012ee4308849a82ba8514e4a156 | migrations/versions/20140430220209_4093ccb6d914.py | migrations/versions/20140430220209_4093ccb6d914.py |
# revision identifiers, used by Alembic.
revision = '4093ccb6d914'
down_revision = None
from alembic import op
import sqlalchemy as sa
from datetime import datetime
def upgrade():
op.create_table('gallery',
sa.Column('id', sa.Integer(), nullable=False),
sa.Column('name', sa.Text(), nullable=False),
sa.Column('folder', sa.types.VARCHAR(length=255), nullable=False),
sa.Column('share_code', sa.Text(), nullable=False),
sa.Column('modified', sa.DateTime(timezone=True), default=datetime.utcnow),
sa.Column('created', sa.DateTime(timezone=True), nullable=True),
sa.PrimaryKeyConstraint('id'),
sa.UniqueConstraint('folder')
)
op.create_table('user',
sa.Column('id', sa.Integer(), nullable=False),
sa.Column('name', sa.types.VARCHAR(length=255), nullable=False),
sa.Column('password', sa.Text(), nullable=False),
sa.Column('role', sa.Text(), nullable=False, server_default="user"),
sa.PrimaryKeyConstraint('id'),
sa.UniqueConstraint('name')
)
def downgrade():
op.drop_table('user')
op.drop_table('gallery')
|
# revision identifiers, used by Alembic.
revision = '4093ccb6d914'
down_revision = None
from alembic import op
import sqlalchemy as sa
from datetime import datetime
def upgrade():
op.create_table('gallery',
sa.Column('id', sa.Integer(), nullable=False),
sa.Column('name', sa.Text(), nullable=False),
sa.Column('folder', sa.types.VARCHAR(length=255), nullable=False),
sa.Column('share_code', sa.Text(), nullable=False),
sa.Column('modified', sa.DateTime(timezone=True), default=datetime.utcnow),
sa.Column('created', sa.DateTime(timezone=True), nullable=True),
sa.PrimaryKeyConstraint('id'),
sa.UniqueConstraint('folder')
)
op.create_table('user',
sa.Column('id', sa.Integer(), nullable=False),
sa.Column('name', sa.types.VARCHAR(length=255), nullable=False),
sa.Column('password', sa.Text(), nullable=False),
sa.Column('role', sa.types.VARCHAR(length=255), nullable=False, server_default="user"),
sa.PrimaryKeyConstraint('id'),
sa.UniqueConstraint('name')
)
def downgrade():
op.drop_table('user')
op.drop_table('gallery')
| Allow MySQL to set a default role | Allow MySQL to set a default role
| Python | mit | taeram/ineffable,taeram/ineffable,taeram/ineffable |
# revision identifiers, used by Alembic.
revision = '4093ccb6d914'
down_revision = None
from alembic import op
import sqlalchemy as sa
from datetime import datetime
def upgrade():
op.create_table('gallery',
sa.Column('id', sa.Integer(), nullable=False),
sa.Column('name', sa.Text(), nullable=False),
sa.Column('folder', sa.types.VARCHAR(length=255), nullable=False),
sa.Column('share_code', sa.Text(), nullable=False),
sa.Column('modified', sa.DateTime(timezone=True), default=datetime.utcnow),
sa.Column('created', sa.DateTime(timezone=True), nullable=True),
sa.PrimaryKeyConstraint('id'),
sa.UniqueConstraint('folder')
)
op.create_table('user',
sa.Column('id', sa.Integer(), nullable=False),
sa.Column('name', sa.types.VARCHAR(length=255), nullable=False),
sa.Column('password', sa.Text(), nullable=False),
- sa.Column('role', sa.Text(), nullable=False, server_default="user"),
+ sa.Column('role', sa.types.VARCHAR(length=255), nullable=False, server_default="user"),
sa.PrimaryKeyConstraint('id'),
sa.UniqueConstraint('name')
)
def downgrade():
op.drop_table('user')
op.drop_table('gallery')
| Allow MySQL to set a default role | ## Code Before:
# revision identifiers, used by Alembic.
revision = '4093ccb6d914'
down_revision = None
from alembic import op
import sqlalchemy as sa
from datetime import datetime
def upgrade():
op.create_table('gallery',
sa.Column('id', sa.Integer(), nullable=False),
sa.Column('name', sa.Text(), nullable=False),
sa.Column('folder', sa.types.VARCHAR(length=255), nullable=False),
sa.Column('share_code', sa.Text(), nullable=False),
sa.Column('modified', sa.DateTime(timezone=True), default=datetime.utcnow),
sa.Column('created', sa.DateTime(timezone=True), nullable=True),
sa.PrimaryKeyConstraint('id'),
sa.UniqueConstraint('folder')
)
op.create_table('user',
sa.Column('id', sa.Integer(), nullable=False),
sa.Column('name', sa.types.VARCHAR(length=255), nullable=False),
sa.Column('password', sa.Text(), nullable=False),
sa.Column('role', sa.Text(), nullable=False, server_default="user"),
sa.PrimaryKeyConstraint('id'),
sa.UniqueConstraint('name')
)
def downgrade():
op.drop_table('user')
op.drop_table('gallery')
## Instruction:
Allow MySQL to set a default role
## Code After:
# revision identifiers, used by Alembic.
revision = '4093ccb6d914'
down_revision = None
from alembic import op
import sqlalchemy as sa
from datetime import datetime
def upgrade():
op.create_table('gallery',
sa.Column('id', sa.Integer(), nullable=False),
sa.Column('name', sa.Text(), nullable=False),
sa.Column('folder', sa.types.VARCHAR(length=255), nullable=False),
sa.Column('share_code', sa.Text(), nullable=False),
sa.Column('modified', sa.DateTime(timezone=True), default=datetime.utcnow),
sa.Column('created', sa.DateTime(timezone=True), nullable=True),
sa.PrimaryKeyConstraint('id'),
sa.UniqueConstraint('folder')
)
op.create_table('user',
sa.Column('id', sa.Integer(), nullable=False),
sa.Column('name', sa.types.VARCHAR(length=255), nullable=False),
sa.Column('password', sa.Text(), nullable=False),
sa.Column('role', sa.types.VARCHAR(length=255), nullable=False, server_default="user"),
sa.PrimaryKeyConstraint('id'),
sa.UniqueConstraint('name')
)
def downgrade():
op.drop_table('user')
op.drop_table('gallery')
| # ... existing code ...
sa.Column('name', sa.types.VARCHAR(length=255), nullable=False),
sa.Column('password', sa.Text(), nullable=False),
sa.Column('role', sa.types.VARCHAR(length=255), nullable=False, server_default="user"),
sa.PrimaryKeyConstraint('id'),
sa.UniqueConstraint('name')
# ... rest of the code ... |
306cf5987c90d54d72037c19dd02f07be37cbb6f | make_mozilla/base/tests/decorators.py | make_mozilla/base/tests/decorators.py | from functools import wraps
from nose.plugins.attrib import attr
from nose.plugins.skip import SkipTest
__all__ = ['wip']
def fail(message):
raise AssertionError(message)
def wip(f):
@wraps(f)
def run_test(*args, **kwargs):
try:
f(*args, **kwargs)
except Exception as e:
raise SkipTest("WIP test failed: " + str(e))
fail("test passed but marked as work in progress")
return attr('wip')(run_test)
| from functools import wraps
from nose.plugins.attrib import attr
from nose.plugins.skip import SkipTest
import os
__all__ = ['wip']
def fail(message):
raise AssertionError(message)
def wip(f):
@wraps(f)
def run_test(*args, **kwargs):
try:
f(*args, **kwargs)
except Exception as e:
raise SkipTest("WIP test failed: " + str(e))
fail("test passed but marked as work in progress")
return attr('wip')(run_test)
def integration(f):
@wraps(f)
def run_test(*args, **kwargs):
integration_run = (os.getenv('INTEGRATION', None) is not None)
if integration_run:
f(*args, **kwargs)
else:
raise SkipTest("Skipping integration test")
return attr('integration')(run_test)
| Add integration test decorator to prevent certain tests running unless we really want them to. | Add integration test decorator to prevent certain tests running unless we really want them to.
| Python | bsd-3-clause | mozilla/make.mozilla.org,mozilla/make.mozilla.org,mozilla/make.mozilla.org,mozilla/make.mozilla.org | from functools import wraps
from nose.plugins.attrib import attr
from nose.plugins.skip import SkipTest
+ import os
__all__ = ['wip']
def fail(message):
raise AssertionError(message)
def wip(f):
@wraps(f)
def run_test(*args, **kwargs):
try:
f(*args, **kwargs)
except Exception as e:
raise SkipTest("WIP test failed: " + str(e))
fail("test passed but marked as work in progress")
return attr('wip')(run_test)
+ def integration(f):
+ @wraps(f)
+ def run_test(*args, **kwargs):
+ integration_run = (os.getenv('INTEGRATION', None) is not None)
+ if integration_run:
+ f(*args, **kwargs)
+ else:
+ raise SkipTest("Skipping integration test")
+ return attr('integration')(run_test)
+ | Add integration test decorator to prevent certain tests running unless we really want them to. | ## Code Before:
from functools import wraps
from nose.plugins.attrib import attr
from nose.plugins.skip import SkipTest
__all__ = ['wip']
def fail(message):
raise AssertionError(message)
def wip(f):
@wraps(f)
def run_test(*args, **kwargs):
try:
f(*args, **kwargs)
except Exception as e:
raise SkipTest("WIP test failed: " + str(e))
fail("test passed but marked as work in progress")
return attr('wip')(run_test)
## Instruction:
Add integration test decorator to prevent certain tests running unless we really want them to.
## Code After:
from functools import wraps
from nose.plugins.attrib import attr
from nose.plugins.skip import SkipTest
import os
__all__ = ['wip']
def fail(message):
raise AssertionError(message)
def wip(f):
@wraps(f)
def run_test(*args, **kwargs):
try:
f(*args, **kwargs)
except Exception as e:
raise SkipTest("WIP test failed: " + str(e))
fail("test passed but marked as work in progress")
return attr('wip')(run_test)
def integration(f):
@wraps(f)
def run_test(*args, **kwargs):
integration_run = (os.getenv('INTEGRATION', None) is not None)
if integration_run:
f(*args, **kwargs)
else:
raise SkipTest("Skipping integration test")
return attr('integration')(run_test)
| // ... existing code ...
from nose.plugins.attrib import attr
from nose.plugins.skip import SkipTest
import os
__all__ = ['wip']
// ... modified code ...
return attr('wip')(run_test)
def integration(f):
@wraps(f)
def run_test(*args, **kwargs):
integration_run = (os.getenv('INTEGRATION', None) is not None)
if integration_run:
f(*args, **kwargs)
else:
raise SkipTest("Skipping integration test")
return attr('integration')(run_test)
// ... rest of the code ... |
28a4f4ab9d6b7c3ea14d48c002273acfe05d7246 | bumblebee/util.py | bumblebee/util.py | import shlex
import exceptions
import subprocess
def bytefmt(num):
for unit in [ "", "Ki", "Mi", "Gi" ]:
if num < 1024.0:
return "{:.2f}{}B".format(num, unit)
num /= 1024.0
return "{:05.2f%}{}GiB".format(num)
def durationfmt(duration):
minutes, seconds = divmod(duration, 60)
hours, minutes = divmod(minutes, 60)
res = "{:02d}:{:02d}".format(minutes, seconds)
if hours > 0: res = "{:02d}:{}".format(hours, res)
return res
def execute(cmd):
args = shlex.split(cmd)
p = subprocess.Popen(args, stdout=subprocess.PIPE, stderr=subprocess.STDOUT)
out = p.communicate()
if p.returncode != 0:
raise exceptions.RuntimeError("{} exited with {}".format(cmd, p.returncode))
| import shlex
import subprocess
try:
from exceptions import RuntimeError
except ImportError:
# Python3 doesn't require this anymore
pass
def bytefmt(num):
for unit in [ "", "Ki", "Mi", "Gi" ]:
if num < 1024.0:
return "{:.2f}{}B".format(num, unit)
num /= 1024.0
return "{:05.2f%}{}GiB".format(num)
def durationfmt(duration):
minutes, seconds = divmod(duration, 60)
hours, minutes = divmod(minutes, 60)
res = "{:02d}:{:02d}".format(minutes, seconds)
if hours > 0: res = "{:02d}:{}".format(hours, res)
return res
def execute(cmd):
args = shlex.split(cmd)
p = subprocess.Popen(args, stdout=subprocess.PIPE, stderr=subprocess.STDOUT)
out = p.communicate()
if p.returncode != 0:
raise RuntimeError("{} exited with {}".format(cmd, p.returncode))
| Fix import error for Python3 | [core] Fix import error for Python3
Import exceptions module only for Python2.
fixes #22
| Python | mit | tobi-wan-kenobi/bumblebee-status,tobi-wan-kenobi/bumblebee-status | import shlex
- import exceptions
import subprocess
+ try:
+ from exceptions import RuntimeError
+ except ImportError:
+ # Python3 doesn't require this anymore
+ pass
def bytefmt(num):
for unit in [ "", "Ki", "Mi", "Gi" ]:
if num < 1024.0:
return "{:.2f}{}B".format(num, unit)
num /= 1024.0
return "{:05.2f%}{}GiB".format(num)
def durationfmt(duration):
minutes, seconds = divmod(duration, 60)
hours, minutes = divmod(minutes, 60)
res = "{:02d}:{:02d}".format(minutes, seconds)
if hours > 0: res = "{:02d}:{}".format(hours, res)
return res
def execute(cmd):
args = shlex.split(cmd)
p = subprocess.Popen(args, stdout=subprocess.PIPE, stderr=subprocess.STDOUT)
out = p.communicate()
if p.returncode != 0:
- raise exceptions.RuntimeError("{} exited with {}".format(cmd, p.returncode))
+ raise RuntimeError("{} exited with {}".format(cmd, p.returncode))
| Fix import error for Python3 | ## Code Before:
import shlex
import exceptions
import subprocess
def bytefmt(num):
for unit in [ "", "Ki", "Mi", "Gi" ]:
if num < 1024.0:
return "{:.2f}{}B".format(num, unit)
num /= 1024.0
return "{:05.2f%}{}GiB".format(num)
def durationfmt(duration):
minutes, seconds = divmod(duration, 60)
hours, minutes = divmod(minutes, 60)
res = "{:02d}:{:02d}".format(minutes, seconds)
if hours > 0: res = "{:02d}:{}".format(hours, res)
return res
def execute(cmd):
args = shlex.split(cmd)
p = subprocess.Popen(args, stdout=subprocess.PIPE, stderr=subprocess.STDOUT)
out = p.communicate()
if p.returncode != 0:
raise exceptions.RuntimeError("{} exited with {}".format(cmd, p.returncode))
## Instruction:
Fix import error for Python3
## Code After:
import shlex
import subprocess
try:
from exceptions import RuntimeError
except ImportError:
# Python3 doesn't require this anymore
pass
def bytefmt(num):
for unit in [ "", "Ki", "Mi", "Gi" ]:
if num < 1024.0:
return "{:.2f}{}B".format(num, unit)
num /= 1024.0
return "{:05.2f%}{}GiB".format(num)
def durationfmt(duration):
minutes, seconds = divmod(duration, 60)
hours, minutes = divmod(minutes, 60)
res = "{:02d}:{:02d}".format(minutes, seconds)
if hours > 0: res = "{:02d}:{}".format(hours, res)
return res
def execute(cmd):
args = shlex.split(cmd)
p = subprocess.Popen(args, stdout=subprocess.PIPE, stderr=subprocess.STDOUT)
out = p.communicate()
if p.returncode != 0:
raise RuntimeError("{} exited with {}".format(cmd, p.returncode))
| ...
import shlex
import subprocess
try:
from exceptions import RuntimeError
except ImportError:
# Python3 doesn't require this anymore
pass
def bytefmt(num):
...
if p.returncode != 0:
raise RuntimeError("{} exited with {}".format(cmd, p.returncode))
... |
0683645a2fb2323a9534d985005d843aada66040 | anypytools/__init__.py | anypytools/__init__.py | """AnyPyTools library."""
import sys
import platform
import logging
from anypytools.abcutils import AnyPyProcess
from anypytools.macroutils import AnyMacro
from anypytools import macro_commands
logger = logging.getLogger('abt.anypytools')
logger.addHandler(logging.NullHandler())
__all__ = [
'datautils', 'h5py_wrapper', 'AnyPyProcess', 'AnyMacro', 'macro_commands',
'print_versions',
]
__version__ = '0.10.2'
def print_versions():
"""Print all the versions of software that AnyPyTools relies on."""
import numpy as np
import scipy as sp
print("-=" * 38)
print("AnyPyTools version: %s" % __version__)
print("NumPy version: %s" % np.__version__)
print("SciPy version: %s" % sp.__version__)
print("Python version: %s" % sys.version)
(sysname, nodename, release, version, machine, processor) = \
platform.uname()
print("Platform: %s-%s-%s (%s)" % (sysname, release, machine, version))
if sysname == "Linux":
print("Linux dist: %s" % " ".join(platform.linux_distribution()[:-1]))
if not processor:
processor = "not recognized"
print("Processor: %s" % processor)
print("Byte-ordering: %s" % sys.byteorder)
print("-=" * 38)
| """AnyPyTools library."""
import sys
import platform
import logging
from anypytools.abcutils import AnyPyProcess, execute_anybodycon
from anypytools.macroutils import AnyMacro
from anypytools import macro_commands
logger = logging.getLogger('abt.anypytools')
logger.addHandler(logging.NullHandler())
__all__ = [
'datautils', 'h5py_wrapper', 'AnyPyProcess', 'AnyMacro', 'macro_commands',
'print_versions', 'execute_anybodycon',
]
__version__ = '0.10.2'
def print_versions():
"""Print all the versions of software that AnyPyTools relies on."""
import numpy as np
import scipy as sp
print("-=" * 38)
print("AnyPyTools version: %s" % __version__)
print("NumPy version: %s" % np.__version__)
print("SciPy version: %s" % sp.__version__)
print("Python version: %s" % sys.version)
(sysname, nodename, release, version, machine, processor) = \
platform.uname()
print("Platform: %s-%s-%s (%s)" % (sysname, release, machine, version))
if sysname == "Linux":
print("Linux dist: %s" % " ".join(platform.linux_distribution()[:-1]))
if not processor:
processor = "not recognized"
print("Processor: %s" % processor)
print("Byte-ordering: %s" % sys.byteorder)
print("-=" * 38)
| Add execute_anybodycon to toplevel package | Add execute_anybodycon to toplevel package
| Python | mit | AnyBody-Research-Group/AnyPyTools | """AnyPyTools library."""
import sys
import platform
import logging
- from anypytools.abcutils import AnyPyProcess
+ from anypytools.abcutils import AnyPyProcess, execute_anybodycon
from anypytools.macroutils import AnyMacro
from anypytools import macro_commands
logger = logging.getLogger('abt.anypytools')
logger.addHandler(logging.NullHandler())
__all__ = [
'datautils', 'h5py_wrapper', 'AnyPyProcess', 'AnyMacro', 'macro_commands',
- 'print_versions',
+ 'print_versions', 'execute_anybodycon',
]
__version__ = '0.10.2'
def print_versions():
"""Print all the versions of software that AnyPyTools relies on."""
import numpy as np
import scipy as sp
print("-=" * 38)
print("AnyPyTools version: %s" % __version__)
print("NumPy version: %s" % np.__version__)
print("SciPy version: %s" % sp.__version__)
print("Python version: %s" % sys.version)
(sysname, nodename, release, version, machine, processor) = \
platform.uname()
print("Platform: %s-%s-%s (%s)" % (sysname, release, machine, version))
if sysname == "Linux":
print("Linux dist: %s" % " ".join(platform.linux_distribution()[:-1]))
if not processor:
processor = "not recognized"
print("Processor: %s" % processor)
print("Byte-ordering: %s" % sys.byteorder)
print("-=" * 38)
| Add execute_anybodycon to toplevel package | ## Code Before:
"""AnyPyTools library."""
import sys
import platform
import logging
from anypytools.abcutils import AnyPyProcess
from anypytools.macroutils import AnyMacro
from anypytools import macro_commands
logger = logging.getLogger('abt.anypytools')
logger.addHandler(logging.NullHandler())
__all__ = [
'datautils', 'h5py_wrapper', 'AnyPyProcess', 'AnyMacro', 'macro_commands',
'print_versions',
]
__version__ = '0.10.2'
def print_versions():
"""Print all the versions of software that AnyPyTools relies on."""
import numpy as np
import scipy as sp
print("-=" * 38)
print("AnyPyTools version: %s" % __version__)
print("NumPy version: %s" % np.__version__)
print("SciPy version: %s" % sp.__version__)
print("Python version: %s" % sys.version)
(sysname, nodename, release, version, machine, processor) = \
platform.uname()
print("Platform: %s-%s-%s (%s)" % (sysname, release, machine, version))
if sysname == "Linux":
print("Linux dist: %s" % " ".join(platform.linux_distribution()[:-1]))
if not processor:
processor = "not recognized"
print("Processor: %s" % processor)
print("Byte-ordering: %s" % sys.byteorder)
print("-=" * 38)
## Instruction:
Add execute_anybodycon to toplevel package
## Code After:
"""AnyPyTools library."""
import sys
import platform
import logging
from anypytools.abcutils import AnyPyProcess, execute_anybodycon
from anypytools.macroutils import AnyMacro
from anypytools import macro_commands
logger = logging.getLogger('abt.anypytools')
logger.addHandler(logging.NullHandler())
__all__ = [
'datautils', 'h5py_wrapper', 'AnyPyProcess', 'AnyMacro', 'macro_commands',
'print_versions', 'execute_anybodycon',
]
__version__ = '0.10.2'
def print_versions():
"""Print all the versions of software that AnyPyTools relies on."""
import numpy as np
import scipy as sp
print("-=" * 38)
print("AnyPyTools version: %s" % __version__)
print("NumPy version: %s" % np.__version__)
print("SciPy version: %s" % sp.__version__)
print("Python version: %s" % sys.version)
(sysname, nodename, release, version, machine, processor) = \
platform.uname()
print("Platform: %s-%s-%s (%s)" % (sysname, release, machine, version))
if sysname == "Linux":
print("Linux dist: %s" % " ".join(platform.linux_distribution()[:-1]))
if not processor:
processor = "not recognized"
print("Processor: %s" % processor)
print("Byte-ordering: %s" % sys.byteorder)
print("-=" * 38)
| // ... existing code ...
import logging
from anypytools.abcutils import AnyPyProcess, execute_anybodycon
from anypytools.macroutils import AnyMacro
from anypytools import macro_commands
// ... modified code ...
__all__ = [
'datautils', 'h5py_wrapper', 'AnyPyProcess', 'AnyMacro', 'macro_commands',
'print_versions', 'execute_anybodycon',
]
// ... rest of the code ... |
6f5be9af15898f089c3ee83ca1f05fbd4570fcfa | src/cms/apps/news/models.py | src/cms/apps/news/models.py | """Models used by the news publication application."""
from django.db import models
from cms.apps.pages.models import Page, PageBase, PageField, HtmlField
from cms.apps.news.content import NewsFeed, NewsArticle
class Article(PageBase):
"""A news article."""
news_feed = PageField(Page,
"newsfeed")
content = HtmlField(blank=True,
null=True)
summary = models.TextField(blank=True,
null=True,
help_text="A short summary of this article. This will be used on news pages and RSS feeds. If not specified, then a summarized version of the content will be used.")
is_featured = models.BooleanField("featured",
default=False,
help_text="Featured articles will remain at the top of any news feeds.")
def get_absolute_url(self):
"""Returns the absolute URL of the article."""
return self.parent.content.reverse("article_detail", self.publication_date.year, self.publication_date.month, self.url_title, "")
class Meta:
verbose_name = "news article"
Page.register_content(NewsFeed)
| """Models used by the news publication application."""
from django.db import models
from cms.apps.pages.models import Page, PageBase, PageField, HtmlField
from cms.apps.news.content import NewsFeed, NewsArticle
class Article(PageBase):
"""A news article."""
news_feed = PageField(Page,
"newsfeed")
content = HtmlField(blank=True,
null=True)
summary = models.TextField(blank=True,
null=True,
help_text="A short summary of this article. This will be used on news pages and RSS feeds. If not specified, then a summarized version of the content will be used.")
is_featured = models.BooleanField("featured",
default=False,
help_text="Featured articles will remain at the top of any news feeds.")
def get_absolute_url(self):
"""Returns the absolute URL of the article."""
return self.parent.content.reverse("article_detail", self.publication_date.year, self.publication_date.month, self.url_title, "")
class Meta:
verbose_name = "news article"
unique_together = (("news_feed", "url_title",),)
Page.register_content(NewsFeed)
| Set unique together on news article. | Set unique together on news article. | Python | bsd-3-clause | lewiscollard/cms,etianen/cms,etianen/cms,danielsamuels/cms,jamesfoley/cms,lewiscollard/cms,jamesfoley/cms,dan-gamble/cms,lewiscollard/cms,jamesfoley/cms,dan-gamble/cms,dan-gamble/cms,jamesfoley/cms,danielsamuels/cms,etianen/cms,danielsamuels/cms | """Models used by the news publication application."""
from django.db import models
from cms.apps.pages.models import Page, PageBase, PageField, HtmlField
from cms.apps.news.content import NewsFeed, NewsArticle
class Article(PageBase):
"""A news article."""
news_feed = PageField(Page,
"newsfeed")
content = HtmlField(blank=True,
null=True)
summary = models.TextField(blank=True,
null=True,
help_text="A short summary of this article. This will be used on news pages and RSS feeds. If not specified, then a summarized version of the content will be used.")
is_featured = models.BooleanField("featured",
default=False,
help_text="Featured articles will remain at the top of any news feeds.")
def get_absolute_url(self):
"""Returns the absolute URL of the article."""
return self.parent.content.reverse("article_detail", self.publication_date.year, self.publication_date.month, self.url_title, "")
class Meta:
verbose_name = "news article"
+ unique_together = (("news_feed", "url_title",),)
-
-
Page.register_content(NewsFeed)
| Set unique together on news article. | ## Code Before:
"""Models used by the news publication application."""
from django.db import models
from cms.apps.pages.models import Page, PageBase, PageField, HtmlField
from cms.apps.news.content import NewsFeed, NewsArticle
class Article(PageBase):
"""A news article."""
news_feed = PageField(Page,
"newsfeed")
content = HtmlField(blank=True,
null=True)
summary = models.TextField(blank=True,
null=True,
help_text="A short summary of this article. This will be used on news pages and RSS feeds. If not specified, then a summarized version of the content will be used.")
is_featured = models.BooleanField("featured",
default=False,
help_text="Featured articles will remain at the top of any news feeds.")
def get_absolute_url(self):
"""Returns the absolute URL of the article."""
return self.parent.content.reverse("article_detail", self.publication_date.year, self.publication_date.month, self.url_title, "")
class Meta:
verbose_name = "news article"
Page.register_content(NewsFeed)
## Instruction:
Set unique together on news article.
## Code After:
"""Models used by the news publication application."""
from django.db import models
from cms.apps.pages.models import Page, PageBase, PageField, HtmlField
from cms.apps.news.content import NewsFeed, NewsArticle
class Article(PageBase):
"""A news article."""
news_feed = PageField(Page,
"newsfeed")
content = HtmlField(blank=True,
null=True)
summary = models.TextField(blank=True,
null=True,
help_text="A short summary of this article. This will be used on news pages and RSS feeds. If not specified, then a summarized version of the content will be used.")
is_featured = models.BooleanField("featured",
default=False,
help_text="Featured articles will remain at the top of any news feeds.")
def get_absolute_url(self):
"""Returns the absolute URL of the article."""
return self.parent.content.reverse("article_detail", self.publication_date.year, self.publication_date.month, self.url_title, "")
class Meta:
verbose_name = "news article"
unique_together = (("news_feed", "url_title",),)
Page.register_content(NewsFeed)
| // ... existing code ...
class Meta:
verbose_name = "news article"
unique_together = (("news_feed", "url_title",),)
// ... rest of the code ... |
3990e3aa64cff288def07ee36e24026cc15282c0 | taiga/projects/issues/serializers.py | taiga/projects/issues/serializers.py |
from rest_framework import serializers
from taiga.base.serializers import PickleField, NeighborsSerializerMixin
from . import models
class IssueSerializer(serializers.ModelSerializer):
tags = PickleField(required=False)
comment = serializers.SerializerMethodField("get_comment")
is_closed = serializers.Field(source="is_closed")
class Meta:
model = models.Issue
def get_comment(self, obj):
return ""
class IssueNeighborsSerializer(NeighborsSerializerMixin, IssueSerializer):
def serialize_neighbor(self, neighbor):
return NeighborIssueSerializer(neighbor).data
class NeighborIssueSerializer(serializers.ModelSerializer):
class Meta:
model = models.Issue
fields = ("id", "ref", "subject")
depth = 0
|
from rest_framework import serializers
from taiga.base.serializers import PickleField, NeighborsSerializerMixin
from . import models
class IssueSerializer(serializers.ModelSerializer):
tags = PickleField(required=False)
is_closed = serializers.Field(source="is_closed")
class Meta:
model = models.Issue
class IssueNeighborsSerializer(NeighborsSerializerMixin, IssueSerializer):
def serialize_neighbor(self, neighbor):
return NeighborIssueSerializer(neighbor).data
class NeighborIssueSerializer(serializers.ModelSerializer):
class Meta:
model = models.Issue
fields = ("id", "ref", "subject")
depth = 0
| Remove unnecessary field from IssueSerializer | Remove unnecessary field from IssueSerializer
| Python | agpl-3.0 | forging2012/taiga-back,EvgeneOskin/taiga-back,xdevelsistemas/taiga-back-community,seanchen/taiga-back,bdang2012/taiga-back-casting,Rademade/taiga-back,crr0004/taiga-back,dayatz/taiga-back,rajiteh/taiga-back,dycodedev/taiga-back,crr0004/taiga-back,obimod/taiga-back,Zaneh-/bearded-tribble-back,seanchen/taiga-back,gauravjns/taiga-back,joshisa/taiga-back,19kestier/taiga-back,jeffdwyatt/taiga-back,taigaio/taiga-back,WALR/taiga-back,joshisa/taiga-back,astronaut1712/taiga-back,taigaio/taiga-back,coopsource/taiga-back,gam-phon/taiga-back,Rademade/taiga-back,obimod/taiga-back,obimod/taiga-back,CMLL/taiga-back,frt-arch/taiga-back,dycodedev/taiga-back,bdang2012/taiga-back-casting,Tigerwhit4/taiga-back,19kestier/taiga-back,EvgeneOskin/taiga-back,EvgeneOskin/taiga-back,astagi/taiga-back,bdang2012/taiga-back-casting,Zaneh-/bearded-tribble-back,dayatz/taiga-back,CoolCloud/taiga-back,astronaut1712/taiga-back,jeffdwyatt/taiga-back,crr0004/taiga-back,WALR/taiga-back,gam-phon/taiga-back,CMLL/taiga-back,seanchen/taiga-back,astagi/taiga-back,gauravjns/taiga-back,gam-phon/taiga-back,WALR/taiga-back,jeffdwyatt/taiga-back,Tigerwhit4/taiga-back,Zaneh-/bearded-tribble-back,seanchen/taiga-back,xdevelsistemas/taiga-back-community,coopsource/taiga-back,astagi/taiga-back,EvgeneOskin/taiga-back,obimod/taiga-back,gam-phon/taiga-back,coopsource/taiga-back,CoolCloud/taiga-back,rajiteh/taiga-back,dycodedev/taiga-back,bdang2012/taiga-back-casting,19kestier/taiga-back,astronaut1712/taiga-back,forging2012/taiga-back,CMLL/taiga-back,frt-arch/taiga-back,astagi/taiga-back,WALR/taiga-back,forging2012/taiga-back,rajiteh/taiga-back,frt-arch/taiga-back,Rademade/taiga-back,xdevelsistemas/taiga-back-community,taigaio/taiga-back,joshisa/taiga-back,gauravjns/taiga-back,Rademade/taiga-back,crr0004/taiga-back,forging2012/taiga-back,joshisa/taiga-back,CMLL/taiga-back,dycodedev/taiga-back,coopsource/taiga-back,CoolCloud/taiga-back,Rademade/taiga-back,astronaut1712/taiga-back,jeffdwyatt/taiga-back,CoolCloud/taiga-back,gauravjns/taiga-back,rajiteh/taiga-back,dayatz/taiga-back,Tigerwhit4/taiga-back,Tigerwhit4/taiga-back |
from rest_framework import serializers
from taiga.base.serializers import PickleField, NeighborsSerializerMixin
from . import models
class IssueSerializer(serializers.ModelSerializer):
tags = PickleField(required=False)
- comment = serializers.SerializerMethodField("get_comment")
is_closed = serializers.Field(source="is_closed")
class Meta:
model = models.Issue
-
- def get_comment(self, obj):
- return ""
class IssueNeighborsSerializer(NeighborsSerializerMixin, IssueSerializer):
def serialize_neighbor(self, neighbor):
return NeighborIssueSerializer(neighbor).data
class NeighborIssueSerializer(serializers.ModelSerializer):
class Meta:
model = models.Issue
fields = ("id", "ref", "subject")
depth = 0
| Remove unnecessary field from IssueSerializer | ## Code Before:
from rest_framework import serializers
from taiga.base.serializers import PickleField, NeighborsSerializerMixin
from . import models
class IssueSerializer(serializers.ModelSerializer):
tags = PickleField(required=False)
comment = serializers.SerializerMethodField("get_comment")
is_closed = serializers.Field(source="is_closed")
class Meta:
model = models.Issue
def get_comment(self, obj):
return ""
class IssueNeighborsSerializer(NeighborsSerializerMixin, IssueSerializer):
def serialize_neighbor(self, neighbor):
return NeighborIssueSerializer(neighbor).data
class NeighborIssueSerializer(serializers.ModelSerializer):
class Meta:
model = models.Issue
fields = ("id", "ref", "subject")
depth = 0
## Instruction:
Remove unnecessary field from IssueSerializer
## Code After:
from rest_framework import serializers
from taiga.base.serializers import PickleField, NeighborsSerializerMixin
from . import models
class IssueSerializer(serializers.ModelSerializer):
tags = PickleField(required=False)
is_closed = serializers.Field(source="is_closed")
class Meta:
model = models.Issue
class IssueNeighborsSerializer(NeighborsSerializerMixin, IssueSerializer):
def serialize_neighbor(self, neighbor):
return NeighborIssueSerializer(neighbor).data
class NeighborIssueSerializer(serializers.ModelSerializer):
class Meta:
model = models.Issue
fields = ("id", "ref", "subject")
depth = 0
| ...
class IssueSerializer(serializers.ModelSerializer):
tags = PickleField(required=False)
is_closed = serializers.Field(source="is_closed")
...
class Meta:
model = models.Issue
... |
76551f7a05506a872ec6535eb3263710650ea8ce | glue/core/data_factories/__init__.py | glue/core/data_factories/__init__.py | from .helpers import *
from .gridded import *
from .pandas import *
from .excel import *
from .image import *
from .dendrogram import *
from .tables import *
| from .helpers import *
from .gridded import *
from .pandas import *
from .excel import *
from .image import *
from .tables import *
from .dendrogram import *
| Order of import matters for disambiguation, but this should be fixed later to avoid this. | Order of import matters for disambiguation, but this should be fixed later to avoid this.
| Python | bsd-3-clause | saimn/glue,stscieisenhamer/glue,stscieisenhamer/glue,JudoWill/glue,saimn/glue,JudoWill/glue | from .helpers import *
from .gridded import *
from .pandas import *
from .excel import *
from .image import *
+ from .tables import *
from .dendrogram import *
- from .tables import *
| Order of import matters for disambiguation, but this should be fixed later to avoid this. | ## Code Before:
from .helpers import *
from .gridded import *
from .pandas import *
from .excel import *
from .image import *
from .dendrogram import *
from .tables import *
## Instruction:
Order of import matters for disambiguation, but this should be fixed later to avoid this.
## Code After:
from .helpers import *
from .gridded import *
from .pandas import *
from .excel import *
from .image import *
from .tables import *
from .dendrogram import *
| ...
from .excel import *
from .image import *
from .tables import *
from .dendrogram import *
... |
f57c9643a32cca012fdccac40899c6de38e35af9 | ass/ets/__init__.py | ass/ets/__init__.py |
from bundles import Environment, Assets, Bundle, Manifest
import filters as f
from options import Option, Options, Undefined, dict_getter
|
from bundles import Environment, Assets, Bundle, Manifest
import filters as f
from options import Option, Options, Undefined, dict_getter
from pipeable import worker
| Make @worker available for import on ass.ets. | Make @worker available for import on ass.ets.
| Python | bsd-2-clause | kaste/ass.ets,kaste/ass.ets |
from bundles import Environment, Assets, Bundle, Manifest
import filters as f
from options import Option, Options, Undefined, dict_getter
+ from pipeable import worker
| Make @worker available for import on ass.ets. | ## Code Before:
from bundles import Environment, Assets, Bundle, Manifest
import filters as f
from options import Option, Options, Undefined, dict_getter
## Instruction:
Make @worker available for import on ass.ets.
## Code After:
from bundles import Environment, Assets, Bundle, Manifest
import filters as f
from options import Option, Options, Undefined, dict_getter
from pipeable import worker
| // ... existing code ...
import filters as f
from options import Option, Options, Undefined, dict_getter
from pipeable import worker
// ... rest of the code ... |
6dd04ed490c49c85bf91db2cb0bf2bed82b5967b | fasttsne/__init__.py | fasttsne/__init__.py | import scipy.linalg as la
import numpy as np
from py_bh_tsne import _TSNE as TSNE
def fast_tsne(data, pca_d=50, d=2, perplexity=30., theta=0.5):
"""
Run Barnes-Hut T-SNE on _data_.
@param data The data.
@param pca_d The dimensionality of data is reduced via PCA
to this dimensionality.
@param d The embedding dimensionality. Must be fixed to
2.
@param perplexity The perplexity controls the effective number of
neighbors.
@param theta If set to 0, exact t-SNE is run, which takes
very long for dataset > 5000 samples.
"""
N, _ = data.shape
# inplace!!
data -= data.mean(axis=0)
# working with covariance + (svd on cov.) is
# much faster than svd on data directly.
cov = np.dot(data.T, data)/N
u, s, v = la.svd(cov, full_matrices=False)
u = u[:,0:pca_d]
X = np.dot(data, u)
tsne = TSNE()
Y = tsne.run(X, N, pca_d, d, perplexity, theta)
return Y
| import scipy.linalg as la
import numpy as np
from fasttsne import _TSNE as TSNE
def fast_tsne(data, pca_d=None, d=2, perplexity=30., theta=0.5):
"""
Run Barnes-Hut T-SNE on _data_.
@param data The data.
@param pca_d The dimensionality of data is reduced via PCA
to this dimensionality.
@param d The embedding dimensionality. Must be fixed to
2.
@param perplexity The perplexity controls the effective number of
neighbors.
@param theta If set to 0, exact t-SNE is run, which takes
very long for dataset > 5000 samples.
"""
N, _ = data.shape
# inplace!!
if pca_d is None:
X = data
else:
# do PCA
data -= data.mean(axis=0)
# working with covariance + (svd on cov.) is
# much faster than svd on data directly.
cov = np.dot(data.T, data)/N
u, s, v = la.svd(cov, full_matrices=False)
u = u[:,0:pca_d]
X = np.dot(data, u)
tsne = TSNE()
Y = tsne.run(X, N, X.shape[1], d, perplexity, theta)
return Y
| FIX (from Justin Bayer): avoid memory segfault when pca_d is choosen too big. | FIX (from Justin Bayer): avoid memory segfault when pca_d is choosen too big.
| Python | bsd-3-clause | pryvkin10x/tsne,douglasbagnall/py_bh_tsne,douglasbagnall/py_bh_tsne,pryvkin10x/tsne,pryvkin10x/tsne | import scipy.linalg as la
import numpy as np
- from py_bh_tsne import _TSNE as TSNE
+ from fasttsne import _TSNE as TSNE
- def fast_tsne(data, pca_d=50, d=2, perplexity=30., theta=0.5):
+ def fast_tsne(data, pca_d=None, d=2, perplexity=30., theta=0.5):
"""
Run Barnes-Hut T-SNE on _data_.
@param data The data.
@param pca_d The dimensionality of data is reduced via PCA
to this dimensionality.
@param d The embedding dimensionality. Must be fixed to
2.
@param perplexity The perplexity controls the effective number of
neighbors.
@param theta If set to 0, exact t-SNE is run, which takes
very long for dataset > 5000 samples.
"""
N, _ = data.shape
# inplace!!
+
+ if pca_d is None:
+ X = data
+ else:
+ # do PCA
- data -= data.mean(axis=0)
+ data -= data.mean(axis=0)
- # working with covariance + (svd on cov.) is
+ # working with covariance + (svd on cov.) is
- # much faster than svd on data directly.
+ # much faster than svd on data directly.
- cov = np.dot(data.T, data)/N
+ cov = np.dot(data.T, data)/N
- u, s, v = la.svd(cov, full_matrices=False)
+ u, s, v = la.svd(cov, full_matrices=False)
- u = u[:,0:pca_d]
+ u = u[:,0:pca_d]
- X = np.dot(data, u)
+ X = np.dot(data, u)
tsne = TSNE()
- Y = tsne.run(X, N, pca_d, d, perplexity, theta)
+ Y = tsne.run(X, N, X.shape[1], d, perplexity, theta)
return Y
| FIX (from Justin Bayer): avoid memory segfault when pca_d is choosen too big. | ## Code Before:
import scipy.linalg as la
import numpy as np
from py_bh_tsne import _TSNE as TSNE
def fast_tsne(data, pca_d=50, d=2, perplexity=30., theta=0.5):
"""
Run Barnes-Hut T-SNE on _data_.
@param data The data.
@param pca_d The dimensionality of data is reduced via PCA
to this dimensionality.
@param d The embedding dimensionality. Must be fixed to
2.
@param perplexity The perplexity controls the effective number of
neighbors.
@param theta If set to 0, exact t-SNE is run, which takes
very long for dataset > 5000 samples.
"""
N, _ = data.shape
# inplace!!
data -= data.mean(axis=0)
# working with covariance + (svd on cov.) is
# much faster than svd on data directly.
cov = np.dot(data.T, data)/N
u, s, v = la.svd(cov, full_matrices=False)
u = u[:,0:pca_d]
X = np.dot(data, u)
tsne = TSNE()
Y = tsne.run(X, N, pca_d, d, perplexity, theta)
return Y
## Instruction:
FIX (from Justin Bayer): avoid memory segfault when pca_d is choosen too big.
## Code After:
import scipy.linalg as la
import numpy as np
from fasttsne import _TSNE as TSNE
def fast_tsne(data, pca_d=None, d=2, perplexity=30., theta=0.5):
"""
Run Barnes-Hut T-SNE on _data_.
@param data The data.
@param pca_d The dimensionality of data is reduced via PCA
to this dimensionality.
@param d The embedding dimensionality. Must be fixed to
2.
@param perplexity The perplexity controls the effective number of
neighbors.
@param theta If set to 0, exact t-SNE is run, which takes
very long for dataset > 5000 samples.
"""
N, _ = data.shape
# inplace!!
if pca_d is None:
X = data
else:
# do PCA
data -= data.mean(axis=0)
# working with covariance + (svd on cov.) is
# much faster than svd on data directly.
cov = np.dot(data.T, data)/N
u, s, v = la.svd(cov, full_matrices=False)
u = u[:,0:pca_d]
X = np.dot(data, u)
tsne = TSNE()
Y = tsne.run(X, N, X.shape[1], d, perplexity, theta)
return Y
| # ... existing code ...
from fasttsne import _TSNE as TSNE
def fast_tsne(data, pca_d=None, d=2, perplexity=30., theta=0.5):
"""
Run Barnes-Hut T-SNE on _data_.
# ... modified code ...
# inplace!!
if pca_d is None:
X = data
else:
# do PCA
data -= data.mean(axis=0)
# working with covariance + (svd on cov.) is
# much faster than svd on data directly.
cov = np.dot(data.T, data)/N
u, s, v = la.svd(cov, full_matrices=False)
u = u[:,0:pca_d]
X = np.dot(data, u)
tsne = TSNE()
Y = tsne.run(X, N, X.shape[1], d, perplexity, theta)
return Y
# ... rest of the code ... |
94d47cfc6db684beda275f8658660a3bd92b319d | src/syft/grid/client/request_api/user_api.py | src/syft/grid/client/request_api/user_api.py | from typing import Any
from typing import Dict
# third party
from pandas import DataFrame
# syft relative
from ...messages.user_messages import CreateUserMessage
from ...messages.user_messages import DeleteUserMessage
from ...messages.user_messages import GetUserMessage
from ...messages.user_messages import GetUsersMessage
from ...messages.user_messages import UpdateUserMessage
from .request_api import GridRequestAPI
class UserRequestAPI(GridRequestAPI):
response_key = "user"
def __init__(self, send):
super().__init__(
create_msg=CreateUserMessage,
get_msg=GetUserMessage,
get_all_msg=GetUsersMessage,
update_msg=UpdateUserMessage,
delete_msg=DeleteUserMessage,
send=send,
response_key=UserRequestAPI.response_key,
)
def __getitem__(self, key):
return self.get(user_id=key)
def __delitem__(self, key):
self.delete(user_id=key)
| from typing import Any
from typing import Callable
# syft relative
from ...messages.user_messages import CreateUserMessage
from ...messages.user_messages import DeleteUserMessage
from ...messages.user_messages import GetUserMessage
from ...messages.user_messages import GetUsersMessage
from ...messages.user_messages import UpdateUserMessage
from .request_api import GridRequestAPI
class UserRequestAPI(GridRequestAPI):
response_key = "user"
def __init__(self, send: Callable):
super().__init__(
create_msg=CreateUserMessage,
get_msg=GetUserMessage,
get_all_msg=GetUsersMessage,
update_msg=UpdateUserMessage,
delete_msg=DeleteUserMessage,
send=send,
response_key=UserRequestAPI.response_key,
)
def __getitem__(self, key: int) -> Any:
return self.get(user_id=key)
def __delitem__(self, key: int) -> None:
self.delete(user_id=key)
| Update User API - ADD type hints - Remove unused imports | Update User API
- ADD type hints
- Remove unused imports
| Python | apache-2.0 | OpenMined/PySyft,OpenMined/PySyft,OpenMined/PySyft,OpenMined/PySyft | from typing import Any
- from typing import Dict
+ from typing import Callable
-
- # third party
- from pandas import DataFrame
# syft relative
from ...messages.user_messages import CreateUserMessage
from ...messages.user_messages import DeleteUserMessage
from ...messages.user_messages import GetUserMessage
from ...messages.user_messages import GetUsersMessage
from ...messages.user_messages import UpdateUserMessage
from .request_api import GridRequestAPI
class UserRequestAPI(GridRequestAPI):
response_key = "user"
- def __init__(self, send):
+ def __init__(self, send: Callable):
super().__init__(
create_msg=CreateUserMessage,
get_msg=GetUserMessage,
get_all_msg=GetUsersMessage,
update_msg=UpdateUserMessage,
delete_msg=DeleteUserMessage,
send=send,
response_key=UserRequestAPI.response_key,
)
- def __getitem__(self, key):
+ def __getitem__(self, key: int) -> Any:
return self.get(user_id=key)
- def __delitem__(self, key):
+ def __delitem__(self, key: int) -> None:
self.delete(user_id=key)
| Update User API - ADD type hints - Remove unused imports | ## Code Before:
from typing import Any
from typing import Dict
# third party
from pandas import DataFrame
# syft relative
from ...messages.user_messages import CreateUserMessage
from ...messages.user_messages import DeleteUserMessage
from ...messages.user_messages import GetUserMessage
from ...messages.user_messages import GetUsersMessage
from ...messages.user_messages import UpdateUserMessage
from .request_api import GridRequestAPI
class UserRequestAPI(GridRequestAPI):
response_key = "user"
def __init__(self, send):
super().__init__(
create_msg=CreateUserMessage,
get_msg=GetUserMessage,
get_all_msg=GetUsersMessage,
update_msg=UpdateUserMessage,
delete_msg=DeleteUserMessage,
send=send,
response_key=UserRequestAPI.response_key,
)
def __getitem__(self, key):
return self.get(user_id=key)
def __delitem__(self, key):
self.delete(user_id=key)
## Instruction:
Update User API - ADD type hints - Remove unused imports
## Code After:
from typing import Any
from typing import Callable
# syft relative
from ...messages.user_messages import CreateUserMessage
from ...messages.user_messages import DeleteUserMessage
from ...messages.user_messages import GetUserMessage
from ...messages.user_messages import GetUsersMessage
from ...messages.user_messages import UpdateUserMessage
from .request_api import GridRequestAPI
class UserRequestAPI(GridRequestAPI):
response_key = "user"
def __init__(self, send: Callable):
super().__init__(
create_msg=CreateUserMessage,
get_msg=GetUserMessage,
get_all_msg=GetUsersMessage,
update_msg=UpdateUserMessage,
delete_msg=DeleteUserMessage,
send=send,
response_key=UserRequestAPI.response_key,
)
def __getitem__(self, key: int) -> Any:
return self.get(user_id=key)
def __delitem__(self, key: int) -> None:
self.delete(user_id=key)
| // ... existing code ...
from typing import Any
from typing import Callable
# syft relative
// ... modified code ...
response_key = "user"
def __init__(self, send: Callable):
super().__init__(
create_msg=CreateUserMessage,
...
)
def __getitem__(self, key: int) -> Any:
return self.get(user_id=key)
def __delitem__(self, key: int) -> None:
self.delete(user_id=key)
// ... rest of the code ... |
ee9b6b1640745bb7b757f1ec8603b19d4f678fb8 | core/observables/file.py | core/observables/file.py | from __future__ import unicode_literals
from mongoengine import *
from core.observables import Observable
from core.observables import Hash
class File(Observable):
value = StringField(verbose_name="SHA256 hash")
mime_type = StringField(verbose_name="MIME type")
hashes = DictField(verbose_name="Hashes")
body = ReferenceField("AttachedFile")
filenames = ListField(StringField(), verbose_name="Filenames")
DISPLAY_FIELDS = Observable.DISPLAY_FIELDS + [("mime_type", "MIME Type")]
@staticmethod
def check_type(txt):
return True
def info(self):
i = Observable.info(self)
i['mime_type'] = self.mime_type
i['hashes'] = self.hashes
return i
| from __future__ import unicode_literals
from flask import url_for
from flask_mongoengine.wtf import model_form
from mongoengine import *
from core.observables import Observable
from core.database import StringListField
class File(Observable):
value = StringField(verbose_name="Value")
mime_type = StringField(verbose_name="MIME type")
hashes = DictField(verbose_name="Hashes")
body = ReferenceField("AttachedFile")
filenames = ListField(StringField(), verbose_name="Filenames")
DISPLAY_FIELDS = Observable.DISPLAY_FIELDS + [("mime_type", "MIME Type")]
exclude_fields = Observable.exclude_fields + ['hashes', 'body']
@classmethod
def get_form(klass):
form = model_form(klass, exclude=klass.exclude_fields)
form.filenames = StringListField("Filenames")
return form
@staticmethod
def check_type(txt):
return True
def info(self):
i = Observable.info(self)
i['mime_type'] = self.mime_type
i['hashes'] = self.hashes
return i
| Clean up File edit view | Clean up File edit view
| Python | apache-2.0 | yeti-platform/yeti,yeti-platform/yeti,yeti-platform/yeti,yeti-platform/yeti | from __future__ import unicode_literals
+ from flask import url_for
+ from flask_mongoengine.wtf import model_form
from mongoengine import *
+
from core.observables import Observable
- from core.observables import Hash
+ from core.database import StringListField
class File(Observable):
- value = StringField(verbose_name="SHA256 hash")
+ value = StringField(verbose_name="Value")
-
mime_type = StringField(verbose_name="MIME type")
hashes = DictField(verbose_name="Hashes")
body = ReferenceField("AttachedFile")
filenames = ListField(StringField(), verbose_name="Filenames")
DISPLAY_FIELDS = Observable.DISPLAY_FIELDS + [("mime_type", "MIME Type")]
+ exclude_fields = Observable.exclude_fields + ['hashes', 'body']
+
+ @classmethod
+ def get_form(klass):
+ form = model_form(klass, exclude=klass.exclude_fields)
+ form.filenames = StringListField("Filenames")
+ return form
@staticmethod
def check_type(txt):
return True
def info(self):
i = Observable.info(self)
i['mime_type'] = self.mime_type
i['hashes'] = self.hashes
return i
| Clean up File edit view | ## Code Before:
from __future__ import unicode_literals
from mongoengine import *
from core.observables import Observable
from core.observables import Hash
class File(Observable):
value = StringField(verbose_name="SHA256 hash")
mime_type = StringField(verbose_name="MIME type")
hashes = DictField(verbose_name="Hashes")
body = ReferenceField("AttachedFile")
filenames = ListField(StringField(), verbose_name="Filenames")
DISPLAY_FIELDS = Observable.DISPLAY_FIELDS + [("mime_type", "MIME Type")]
@staticmethod
def check_type(txt):
return True
def info(self):
i = Observable.info(self)
i['mime_type'] = self.mime_type
i['hashes'] = self.hashes
return i
## Instruction:
Clean up File edit view
## Code After:
from __future__ import unicode_literals
from flask import url_for
from flask_mongoengine.wtf import model_form
from mongoengine import *
from core.observables import Observable
from core.database import StringListField
class File(Observable):
value = StringField(verbose_name="Value")
mime_type = StringField(verbose_name="MIME type")
hashes = DictField(verbose_name="Hashes")
body = ReferenceField("AttachedFile")
filenames = ListField(StringField(), verbose_name="Filenames")
DISPLAY_FIELDS = Observable.DISPLAY_FIELDS + [("mime_type", "MIME Type")]
exclude_fields = Observable.exclude_fields + ['hashes', 'body']
@classmethod
def get_form(klass):
form = model_form(klass, exclude=klass.exclude_fields)
form.filenames = StringListField("Filenames")
return form
@staticmethod
def check_type(txt):
return True
def info(self):
i = Observable.info(self)
i['mime_type'] = self.mime_type
i['hashes'] = self.hashes
return i
| # ... existing code ...
from __future__ import unicode_literals
from flask import url_for
from flask_mongoengine.wtf import model_form
from mongoengine import *
from core.observables import Observable
from core.database import StringListField
# ... modified code ...
class File(Observable):
value = StringField(verbose_name="Value")
mime_type = StringField(verbose_name="MIME type")
hashes = DictField(verbose_name="Hashes")
...
DISPLAY_FIELDS = Observable.DISPLAY_FIELDS + [("mime_type", "MIME Type")]
exclude_fields = Observable.exclude_fields + ['hashes', 'body']
@classmethod
def get_form(klass):
form = model_form(klass, exclude=klass.exclude_fields)
form.filenames = StringListField("Filenames")
return form
@staticmethod
# ... rest of the code ... |
d3a203725d13a7abef091f0070f90826d3225dbc | settings_travis.py | settings_travis.py | import ssl
LDAP_SERVER = 'ldap.rserver.de'
LDAP_PORT = 3389
LDAP_SSL_PORT = 6636
LDAP_REQUIRE_CERT = ssl.CERT_NONE
| import ssl
LDAP_SERVER = 'ldap.rserver.de'
LDAP_PORT = 3389
LDAP_SSL_PORT = 6636
LDAP_REQUIRE_CERT = ssl.CERT_NONE
LDAP_TLS_VERSION = ssl.PROTOCOL_TLSv1
| Fix travis unit test for python 3.3 | Fix travis unit test for python 3.3
| Python | bsd-2-clause | rroemhild/flask-ldapconn | import ssl
LDAP_SERVER = 'ldap.rserver.de'
LDAP_PORT = 3389
LDAP_SSL_PORT = 6636
LDAP_REQUIRE_CERT = ssl.CERT_NONE
+ LDAP_TLS_VERSION = ssl.PROTOCOL_TLSv1
| Fix travis unit test for python 3.3 | ## Code Before:
import ssl
LDAP_SERVER = 'ldap.rserver.de'
LDAP_PORT = 3389
LDAP_SSL_PORT = 6636
LDAP_REQUIRE_CERT = ssl.CERT_NONE
## Instruction:
Fix travis unit test for python 3.3
## Code After:
import ssl
LDAP_SERVER = 'ldap.rserver.de'
LDAP_PORT = 3389
LDAP_SSL_PORT = 6636
LDAP_REQUIRE_CERT = ssl.CERT_NONE
LDAP_TLS_VERSION = ssl.PROTOCOL_TLSv1
| # ... existing code ...
LDAP_SSL_PORT = 6636
LDAP_REQUIRE_CERT = ssl.CERT_NONE
LDAP_TLS_VERSION = ssl.PROTOCOL_TLSv1
# ... rest of the code ... |
e880522e226b593be2985cdc85cb0ebd87e53a98 | astral/models/tests/factories.py | astral/models/tests/factories.py | import factory
import faker
import random
import uuid
from astral.models.stream import Stream
from astral.models.node import Node
from astral.models.ticket import Ticket
ELIXIR_CREATION = lambda class_to_create, **kwargs: class_to_create(**kwargs)
factory.Factory.set_creation_function(ELIXIR_CREATION)
class StreamFactory(factory.Factory):
id = factory.Sequence(lambda n: int(n) + 1)
name = factory.LazyAttribute(lambda a: ' '.join(faker.lorem.words()))
class NodeFactory(factory.Factory):
ip_address = factory.LazyAttribute(lambda a: faker.internet.ip_address())
uuid = factory.LazyAttribute(lambda a: random.randrange(1000, 1000000))
port = factory.LazyAttribute(lambda a: random.randrange(1000, 10000))
class ThisNodeFactory(factory.Factory):
FACTORY_FOR = Node
ip_address = '127.0.0.1'
uuid = factory.LazyAttribute(lambda a: uuid.getnode())
port = factory.LazyAttribute(lambda a: random.randrange(1000, 10000))
class TicketFactory(factory.Factory):
source = factory.LazyAttribute(lambda a: NodeFactory())
destination = factory.LazyAttribute(lambda a: NodeFactory())
stream = factory.LazyAttribute(lambda a: StreamFactory())
| import factory
import faker
import random
import uuid
from astral.models.stream import Stream
from astral.models.node import Node
from astral.models.ticket import Ticket
ELIXIR_CREATION = lambda class_to_create, **kwargs: class_to_create(**kwargs)
factory.Factory.set_creation_function(ELIXIR_CREATION)
class StreamFactory(factory.Factory):
id = factory.Sequence(lambda n: int(n) + 1)
name = factory.LazyAttribute(lambda a: ' '.join(faker.lorem.words()))
source = factory.LazyAttribute(lambda a: NodeFactory())
class NodeFactory(factory.Factory):
ip_address = factory.LazyAttribute(lambda a: faker.internet.ip_address())
uuid = factory.LazyAttribute(lambda a: random.randrange(1000, 1000000))
port = factory.LazyAttribute(lambda a: random.randrange(1000, 10000))
class ThisNodeFactory(factory.Factory):
FACTORY_FOR = Node
ip_address = '127.0.0.1'
uuid = factory.LazyAttribute(lambda a: uuid.getnode())
port = factory.LazyAttribute(lambda a: random.randrange(1000, 10000))
class TicketFactory(factory.Factory):
source = factory.LazyAttribute(lambda a: NodeFactory())
destination = factory.LazyAttribute(lambda a: NodeFactory())
stream = factory.LazyAttribute(lambda a: StreamFactory())
| Make sure streams always have a source. | Make sure streams always have a source.
| Python | mit | peplin/astral | import factory
import faker
import random
import uuid
from astral.models.stream import Stream
from astral.models.node import Node
from astral.models.ticket import Ticket
ELIXIR_CREATION = lambda class_to_create, **kwargs: class_to_create(**kwargs)
factory.Factory.set_creation_function(ELIXIR_CREATION)
class StreamFactory(factory.Factory):
id = factory.Sequence(lambda n: int(n) + 1)
name = factory.LazyAttribute(lambda a: ' '.join(faker.lorem.words()))
+ source = factory.LazyAttribute(lambda a: NodeFactory())
class NodeFactory(factory.Factory):
ip_address = factory.LazyAttribute(lambda a: faker.internet.ip_address())
uuid = factory.LazyAttribute(lambda a: random.randrange(1000, 1000000))
port = factory.LazyAttribute(lambda a: random.randrange(1000, 10000))
class ThisNodeFactory(factory.Factory):
FACTORY_FOR = Node
ip_address = '127.0.0.1'
uuid = factory.LazyAttribute(lambda a: uuid.getnode())
port = factory.LazyAttribute(lambda a: random.randrange(1000, 10000))
class TicketFactory(factory.Factory):
source = factory.LazyAttribute(lambda a: NodeFactory())
destination = factory.LazyAttribute(lambda a: NodeFactory())
stream = factory.LazyAttribute(lambda a: StreamFactory())
| Make sure streams always have a source. | ## Code Before:
import factory
import faker
import random
import uuid
from astral.models.stream import Stream
from astral.models.node import Node
from astral.models.ticket import Ticket
ELIXIR_CREATION = lambda class_to_create, **kwargs: class_to_create(**kwargs)
factory.Factory.set_creation_function(ELIXIR_CREATION)
class StreamFactory(factory.Factory):
id = factory.Sequence(lambda n: int(n) + 1)
name = factory.LazyAttribute(lambda a: ' '.join(faker.lorem.words()))
class NodeFactory(factory.Factory):
ip_address = factory.LazyAttribute(lambda a: faker.internet.ip_address())
uuid = factory.LazyAttribute(lambda a: random.randrange(1000, 1000000))
port = factory.LazyAttribute(lambda a: random.randrange(1000, 10000))
class ThisNodeFactory(factory.Factory):
FACTORY_FOR = Node
ip_address = '127.0.0.1'
uuid = factory.LazyAttribute(lambda a: uuid.getnode())
port = factory.LazyAttribute(lambda a: random.randrange(1000, 10000))
class TicketFactory(factory.Factory):
source = factory.LazyAttribute(lambda a: NodeFactory())
destination = factory.LazyAttribute(lambda a: NodeFactory())
stream = factory.LazyAttribute(lambda a: StreamFactory())
## Instruction:
Make sure streams always have a source.
## Code After:
import factory
import faker
import random
import uuid
from astral.models.stream import Stream
from astral.models.node import Node
from astral.models.ticket import Ticket
ELIXIR_CREATION = lambda class_to_create, **kwargs: class_to_create(**kwargs)
factory.Factory.set_creation_function(ELIXIR_CREATION)
class StreamFactory(factory.Factory):
id = factory.Sequence(lambda n: int(n) + 1)
name = factory.LazyAttribute(lambda a: ' '.join(faker.lorem.words()))
source = factory.LazyAttribute(lambda a: NodeFactory())
class NodeFactory(factory.Factory):
ip_address = factory.LazyAttribute(lambda a: faker.internet.ip_address())
uuid = factory.LazyAttribute(lambda a: random.randrange(1000, 1000000))
port = factory.LazyAttribute(lambda a: random.randrange(1000, 10000))
class ThisNodeFactory(factory.Factory):
FACTORY_FOR = Node
ip_address = '127.0.0.1'
uuid = factory.LazyAttribute(lambda a: uuid.getnode())
port = factory.LazyAttribute(lambda a: random.randrange(1000, 10000))
class TicketFactory(factory.Factory):
source = factory.LazyAttribute(lambda a: NodeFactory())
destination = factory.LazyAttribute(lambda a: NodeFactory())
stream = factory.LazyAttribute(lambda a: StreamFactory())
| ...
id = factory.Sequence(lambda n: int(n) + 1)
name = factory.LazyAttribute(lambda a: ' '.join(faker.lorem.words()))
source = factory.LazyAttribute(lambda a: NodeFactory())
... |
a589aa63f250a347ab24b7309e65ef25c7281437 | src/sentry/utils/imports.py | src/sentry/utils/imports.py | from __future__ import absolute_import
import pkgutil
import six
class ModuleProxyCache(dict):
def __missing__(self, key):
if '.' not in key:
return __import__(key)
module_name, class_name = key.rsplit('.', 1)
module = __import__(module_name, {}, {}, [class_name])
handler = getattr(module, class_name)
# We cache a NoneType for missing imports to avoid repeated lookups
self[key] = handler
return handler
_cache = ModuleProxyCache()
def import_string(path):
"""
Path must be module.path.ClassName
>>> cls = import_string('sentry.models.Group')
"""
result = _cache[path]
return result
def import_submodules(context, root_module, path):
"""
Import all submodules and register them in the ``context`` namespace.
>>> import_submodules(locals(), __name__, __path__)
"""
for loader, module_name, is_pkg in pkgutil.walk_packages(path, root_module + '.'):
module = loader.find_module(module_name).load_module(module_name)
for k, v in six.iteritems(vars(module)):
if not k.startswith('_'):
context[k] = v
context[module_name] = module
| from __future__ import absolute_import
import pkgutil
import six
class ModuleProxyCache(dict):
def __missing__(self, key):
if '.' not in key:
return __import__(key)
module_name, class_name = key.rsplit('.', 1)
module = __import__(module_name, {}, {}, [class_name])
handler = getattr(module, class_name)
# We cache a NoneType for missing imports to avoid repeated lookups
self[key] = handler
return handler
_cache = ModuleProxyCache()
def import_string(path):
"""
Path must be module.path.ClassName
>>> cls = import_string('sentry.models.Group')
"""
result = _cache[path]
return result
def import_submodules(context, root_module, path):
"""
Import all submodules and register them in the ``context`` namespace.
>>> import_submodules(locals(), __name__, __path__)
"""
for loader, module_name, is_pkg in pkgutil.walk_packages(path, root_module + '.'):
# this causes a Runtime error with model conflicts
# module = loader.find_module(module_name).load_module(module_name)
module = __import__(module_name, globals(), locals(), ['__name__'])
for k, v in six.iteritems(vars(module)):
if not k.startswith('_'):
context[k] = v
context[module_name] = module
| Correct import behavior to prevent Runtime error | Correct import behavior to prevent Runtime error
| Python | bsd-3-clause | gencer/sentry,jean/sentry,fotinakis/sentry,gencer/sentry,looker/sentry,BuildingLink/sentry,gencer/sentry,jean/sentry,JackDanger/sentry,fotinakis/sentry,BuildingLink/sentry,beeftornado/sentry,zenefits/sentry,beeftornado/sentry,looker/sentry,ifduyue/sentry,ifduyue/sentry,JamesMura/sentry,ifduyue/sentry,JamesMura/sentry,BuildingLink/sentry,zenefits/sentry,fotinakis/sentry,ifduyue/sentry,zenefits/sentry,zenefits/sentry,beeftornado/sentry,BuildingLink/sentry,JamesMura/sentry,jean/sentry,gencer/sentry,ifduyue/sentry,jean/sentry,alexm92/sentry,mvaled/sentry,gencer/sentry,alexm92/sentry,looker/sentry,JackDanger/sentry,mvaled/sentry,mvaled/sentry,BuildingLink/sentry,JamesMura/sentry,JackDanger/sentry,fotinakis/sentry,looker/sentry,zenefits/sentry,mvaled/sentry,mvaled/sentry,jean/sentry,JamesMura/sentry,alexm92/sentry,looker/sentry,mvaled/sentry | from __future__ import absolute_import
import pkgutil
import six
class ModuleProxyCache(dict):
def __missing__(self, key):
if '.' not in key:
return __import__(key)
module_name, class_name = key.rsplit('.', 1)
module = __import__(module_name, {}, {}, [class_name])
handler = getattr(module, class_name)
# We cache a NoneType for missing imports to avoid repeated lookups
self[key] = handler
return handler
_cache = ModuleProxyCache()
def import_string(path):
"""
Path must be module.path.ClassName
>>> cls = import_string('sentry.models.Group')
"""
result = _cache[path]
return result
def import_submodules(context, root_module, path):
"""
Import all submodules and register them in the ``context`` namespace.
>>> import_submodules(locals(), __name__, __path__)
"""
for loader, module_name, is_pkg in pkgutil.walk_packages(path, root_module + '.'):
+ # this causes a Runtime error with model conflicts
- module = loader.find_module(module_name).load_module(module_name)
+ # module = loader.find_module(module_name).load_module(module_name)
+ module = __import__(module_name, globals(), locals(), ['__name__'])
for k, v in six.iteritems(vars(module)):
if not k.startswith('_'):
context[k] = v
context[module_name] = module
| Correct import behavior to prevent Runtime error | ## Code Before:
from __future__ import absolute_import
import pkgutil
import six
class ModuleProxyCache(dict):
def __missing__(self, key):
if '.' not in key:
return __import__(key)
module_name, class_name = key.rsplit('.', 1)
module = __import__(module_name, {}, {}, [class_name])
handler = getattr(module, class_name)
# We cache a NoneType for missing imports to avoid repeated lookups
self[key] = handler
return handler
_cache = ModuleProxyCache()
def import_string(path):
"""
Path must be module.path.ClassName
>>> cls = import_string('sentry.models.Group')
"""
result = _cache[path]
return result
def import_submodules(context, root_module, path):
"""
Import all submodules and register them in the ``context`` namespace.
>>> import_submodules(locals(), __name__, __path__)
"""
for loader, module_name, is_pkg in pkgutil.walk_packages(path, root_module + '.'):
module = loader.find_module(module_name).load_module(module_name)
for k, v in six.iteritems(vars(module)):
if not k.startswith('_'):
context[k] = v
context[module_name] = module
## Instruction:
Correct import behavior to prevent Runtime error
## Code After:
from __future__ import absolute_import
import pkgutil
import six
class ModuleProxyCache(dict):
def __missing__(self, key):
if '.' not in key:
return __import__(key)
module_name, class_name = key.rsplit('.', 1)
module = __import__(module_name, {}, {}, [class_name])
handler = getattr(module, class_name)
# We cache a NoneType for missing imports to avoid repeated lookups
self[key] = handler
return handler
_cache = ModuleProxyCache()
def import_string(path):
"""
Path must be module.path.ClassName
>>> cls = import_string('sentry.models.Group')
"""
result = _cache[path]
return result
def import_submodules(context, root_module, path):
"""
Import all submodules and register them in the ``context`` namespace.
>>> import_submodules(locals(), __name__, __path__)
"""
for loader, module_name, is_pkg in pkgutil.walk_packages(path, root_module + '.'):
# this causes a Runtime error with model conflicts
# module = loader.find_module(module_name).load_module(module_name)
module = __import__(module_name, globals(), locals(), ['__name__'])
for k, v in six.iteritems(vars(module)):
if not k.startswith('_'):
context[k] = v
context[module_name] = module
| # ... existing code ...
"""
for loader, module_name, is_pkg in pkgutil.walk_packages(path, root_module + '.'):
# this causes a Runtime error with model conflicts
# module = loader.find_module(module_name).load_module(module_name)
module = __import__(module_name, globals(), locals(), ['__name__'])
for k, v in six.iteritems(vars(module)):
if not k.startswith('_'):
# ... rest of the code ... |
5a6b19f956dfde65a1d8316fd4bebe4697846e45 | connman_dispatcher/detect.py | connman_dispatcher/detect.py | import glib
import dbus
from dbus.mainloop.glib import DBusGMainLoop
from pyee import EventEmitter
import logbook
logger = logbook.Logger('connman-dispatcher')
__all__ = ['detector']
def property_changed(_, message):
if message.get_member() == "PropertyChanged":
_, state = message.get_args_list()
if state == 'online' and not detector.is_online:
logger.info('network state change: online' )
detector.emit('up')
detector.is_online = True
elif state == 'idle':
logger.info('network state change: offline' )
detector.emit('down')
detector.is_online = False
detector = EventEmitter()
detector.is_online = is_online()
DBusGMainLoop(set_as_default=True)
bus = dbus.SystemBus()
bus.add_match_string_non_blocking("interface='net.connman.Manager'")
bus.add_message_filter(property_changed)
manager = dbus.Interface(bus.get_object('net.connman', "/"), 'net.connman.Manager')
def is_online():
properties = manager.GetProperties()
if properties['State'] == 'online':
return True
return False
def run():
mainloop = glib.MainLoop()
mainloop.run()
detector.run = run
detector.is_online = is_online
| import glib
import dbus
from dbus.mainloop.glib import DBusGMainLoop
from pyee import EventEmitter
import logbook
logger = logbook.Logger('connman-dispatcher')
__all__ = ['detector']
def property_changed(_, message):
if message.get_member() == "PropertyChanged":
_, state = message.get_args_list()
if state == 'online' and detector.state == 'offline':
logger.info('network state change: online' )
detector.emit('up')
detector.state = 'online'
elif state == 'idle':
logger.info('network state change: offline' )
detector.emit('down')
detector.state = 'online'
detector = EventEmitter()
DBusGMainLoop(set_as_default=True)
bus = dbus.SystemBus()
bus.add_match_string_non_blocking("interface='net.connman.Manager'")
bus.add_message_filter(property_changed)
manager = dbus.Interface(bus.get_object('net.connman', "/"), 'net.connman.Manager')
def is_online():
properties = manager.GetProperties()
if properties['State'] == 'online':
return True
return False
def run():
mainloop = glib.MainLoop()
mainloop.run()
detector.run = run
detector.is_online = is_online
detector.state = 'online' if is_online() else 'offline'
| Use .state instead of .is_online to keep internal state | Use .state instead of .is_online to keep internal state
| Python | isc | a-sk/connman-dispatcher | import glib
import dbus
from dbus.mainloop.glib import DBusGMainLoop
from pyee import EventEmitter
import logbook
logger = logbook.Logger('connman-dispatcher')
__all__ = ['detector']
def property_changed(_, message):
if message.get_member() == "PropertyChanged":
_, state = message.get_args_list()
- if state == 'online' and not detector.is_online:
+ if state == 'online' and detector.state == 'offline':
logger.info('network state change: online' )
detector.emit('up')
- detector.is_online = True
+ detector.state = 'online'
elif state == 'idle':
logger.info('network state change: offline' )
detector.emit('down')
- detector.is_online = False
+ detector.state = 'online'
detector = EventEmitter()
- detector.is_online = is_online()
DBusGMainLoop(set_as_default=True)
bus = dbus.SystemBus()
bus.add_match_string_non_blocking("interface='net.connman.Manager'")
bus.add_message_filter(property_changed)
manager = dbus.Interface(bus.get_object('net.connman', "/"), 'net.connman.Manager')
def is_online():
properties = manager.GetProperties()
if properties['State'] == 'online':
return True
return False
def run():
mainloop = glib.MainLoop()
mainloop.run()
detector.run = run
detector.is_online = is_online
+ detector.state = 'online' if is_online() else 'offline'
| Use .state instead of .is_online to keep internal state | ## Code Before:
import glib
import dbus
from dbus.mainloop.glib import DBusGMainLoop
from pyee import EventEmitter
import logbook
logger = logbook.Logger('connman-dispatcher')
__all__ = ['detector']
def property_changed(_, message):
if message.get_member() == "PropertyChanged":
_, state = message.get_args_list()
if state == 'online' and not detector.is_online:
logger.info('network state change: online' )
detector.emit('up')
detector.is_online = True
elif state == 'idle':
logger.info('network state change: offline' )
detector.emit('down')
detector.is_online = False
detector = EventEmitter()
detector.is_online = is_online()
DBusGMainLoop(set_as_default=True)
bus = dbus.SystemBus()
bus.add_match_string_non_blocking("interface='net.connman.Manager'")
bus.add_message_filter(property_changed)
manager = dbus.Interface(bus.get_object('net.connman', "/"), 'net.connman.Manager')
def is_online():
properties = manager.GetProperties()
if properties['State'] == 'online':
return True
return False
def run():
mainloop = glib.MainLoop()
mainloop.run()
detector.run = run
detector.is_online = is_online
## Instruction:
Use .state instead of .is_online to keep internal state
## Code After:
import glib
import dbus
from dbus.mainloop.glib import DBusGMainLoop
from pyee import EventEmitter
import logbook
logger = logbook.Logger('connman-dispatcher')
__all__ = ['detector']
def property_changed(_, message):
if message.get_member() == "PropertyChanged":
_, state = message.get_args_list()
if state == 'online' and detector.state == 'offline':
logger.info('network state change: online' )
detector.emit('up')
detector.state = 'online'
elif state == 'idle':
logger.info('network state change: offline' )
detector.emit('down')
detector.state = 'online'
detector = EventEmitter()
DBusGMainLoop(set_as_default=True)
bus = dbus.SystemBus()
bus.add_match_string_non_blocking("interface='net.connman.Manager'")
bus.add_message_filter(property_changed)
manager = dbus.Interface(bus.get_object('net.connman', "/"), 'net.connman.Manager')
def is_online():
properties = manager.GetProperties()
if properties['State'] == 'online':
return True
return False
def run():
mainloop = glib.MainLoop()
mainloop.run()
detector.run = run
detector.is_online = is_online
detector.state = 'online' if is_online() else 'offline'
| ...
if message.get_member() == "PropertyChanged":
_, state = message.get_args_list()
if state == 'online' and detector.state == 'offline':
logger.info('network state change: online' )
detector.emit('up')
detector.state = 'online'
elif state == 'idle':
logger.info('network state change: offline' )
detector.emit('down')
detector.state = 'online'
detector = EventEmitter()
DBusGMainLoop(set_as_default=True)
...
detector.run = run
detector.is_online = is_online
detector.state = 'online' if is_online() else 'offline'
... |
3747158af790a38ccfce217426ee5261877e9f0e | project/api/management/commands/seed_database.py | project/api/management/commands/seed_database.py | from django.core.management.base import BaseCommand
from api.factories import (
InternationalFactory,
)
class Command(BaseCommand):
help = "Command to seed database."
def handle(self, *args, **options):
InternationalFactory()
| from django.core.management.base import BaseCommand
class Command(BaseCommand):
help = "Command to seed database."
from api.factories import (
InternationalFactory,
)
def handle(self, *args, **options):
self.InternationalFactory()
| Fix seeding in management command | Fix seeding in management command
| Python | bsd-2-clause | barberscore/barberscore-api,barberscore/barberscore-api,barberscore/barberscore-api,barberscore/barberscore-api,dbinetti/barberscore-django,dbinetti/barberscore,dbinetti/barberscore-django,dbinetti/barberscore | from django.core.management.base import BaseCommand
-
-
- from api.factories import (
- InternationalFactory,
- )
class Command(BaseCommand):
help = "Command to seed database."
+ from api.factories import (
+ InternationalFactory,
+ )
+
def handle(self, *args, **options):
- InternationalFactory()
+ self.InternationalFactory()
| Fix seeding in management command | ## Code Before:
from django.core.management.base import BaseCommand
from api.factories import (
InternationalFactory,
)
class Command(BaseCommand):
help = "Command to seed database."
def handle(self, *args, **options):
InternationalFactory()
## Instruction:
Fix seeding in management command
## Code After:
from django.core.management.base import BaseCommand
class Command(BaseCommand):
help = "Command to seed database."
from api.factories import (
InternationalFactory,
)
def handle(self, *args, **options):
self.InternationalFactory()
| # ... existing code ...
from django.core.management.base import BaseCommand
# ... modified code ...
help = "Command to seed database."
from api.factories import (
InternationalFactory,
)
def handle(self, *args, **options):
self.InternationalFactory()
# ... rest of the code ... |
d883a0fd09a42ff84ebb2ccf331692167370444b | ESLog/esloghandler.py | ESLog/esloghandler.py |
from datetime import datetime
import logging
import os
import json
import urllib.request
import urllib.parse
class ESLogHandler(logging.Handler):
def __init__(self, url, index=None, doc_type="log", level=logging.NOTSET):
logging.Handler.__init__(self, level=level)
self.url = urllib.parse.urlparse(url)
print(self.url)
# end __init__
def emit(self, record):
# Break the record down to a dictionary
message = dict()
message["timestamp"] = datetime.now().isoformat()
message["level"] = record.levelname
message["name"] = record.name
message["lineno"] = record.lineno
message["message"] = record.msg
json_message = json.dumps(message)
json_message_bytes = json_message.encode("utf8")
urllib.request.urlopen(self.url, data=json_message_bytes)
# end emit
# end ESLogHandler
|
from datetime import datetime
import logging
import os
import json
import urllib.request
import urllib.parse
class ESLogHandler(logging.Handler):
def __init__(self, url, index=None, doc_type="log", level=logging.NOTSET):
logging.Handler.__init__(self, level=level)
# Parse the url
self.url = urllib.parse.urlparse(url)
# If no scheme is given, set it to http
if not self.url.scheme:
self.url.scheme = "http"
# If a scheme is given but it is not http, raise an exception
elif self.url.scheme != "http":
raise ValueError("Only HTTP is supported.")
# If no port is given default to 9200
if not self.url.port:
self.url.port = "9200"
# If no path is given or it is only a / use thi index and doc_type to construct one.
if not self.url.path or self.url.path == "/":
# an index is mandatory for Elasticsearch, doc_type too but it defaults to log
if not index:
raise ValueError("Elasticsearch index cannot be ommitted.")
else:
self.url.path = os.path.join("/", index, doc_type)
# end __init__
def emit(self, record):
# Break the record down to a dictionary
message = dict()
message["timestamp"] = datetime.now().isoformat()
message["level"] = record.levelname
message["name"] = record.name
message["lineno"] = record.lineno
message["message"] = record.msg
json_message = json.dumps(message)
json_message_bytes = json_message.encode("utf8")
urllib.request.urlopen(self.url, data=json_message_bytes)
# end emit
# end ESLogHandler
| Revert "trying to simplefy __init__" | Revert "trying to simplefy __init__"
This reverts commit f2e3887bcd53b1c35af2d9dfe2363ea9e2a407f5.
| Python | mit | Rio/ESLog |
from datetime import datetime
import logging
import os
import json
import urllib.request
import urllib.parse
class ESLogHandler(logging.Handler):
def __init__(self, url, index=None, doc_type="log", level=logging.NOTSET):
logging.Handler.__init__(self, level=level)
-
+
+ # Parse the url
self.url = urllib.parse.urlparse(url)
-
- print(self.url)
+
+ # If no scheme is given, set it to http
+ if not self.url.scheme:
+ self.url.scheme = "http"
+
+ # If a scheme is given but it is not http, raise an exception
+ elif self.url.scheme != "http":
+ raise ValueError("Only HTTP is supported.")
+
+ # If no port is given default to 9200
+ if not self.url.port:
+ self.url.port = "9200"
+
+ # If no path is given or it is only a / use thi index and doc_type to construct one.
+ if not self.url.path or self.url.path == "/":
+ # an index is mandatory for Elasticsearch, doc_type too but it defaults to log
+ if not index:
+ raise ValueError("Elasticsearch index cannot be ommitted.")
+
+ else:
+ self.url.path = os.path.join("/", index, doc_type)
# end __init__
def emit(self, record):
# Break the record down to a dictionary
message = dict()
message["timestamp"] = datetime.now().isoformat()
message["level"] = record.levelname
message["name"] = record.name
message["lineno"] = record.lineno
message["message"] = record.msg
json_message = json.dumps(message)
json_message_bytes = json_message.encode("utf8")
urllib.request.urlopen(self.url, data=json_message_bytes)
# end emit
# end ESLogHandler
| Revert "trying to simplefy __init__" | ## Code Before:
from datetime import datetime
import logging
import os
import json
import urllib.request
import urllib.parse
class ESLogHandler(logging.Handler):
def __init__(self, url, index=None, doc_type="log", level=logging.NOTSET):
logging.Handler.__init__(self, level=level)
self.url = urllib.parse.urlparse(url)
print(self.url)
# end __init__
def emit(self, record):
# Break the record down to a dictionary
message = dict()
message["timestamp"] = datetime.now().isoformat()
message["level"] = record.levelname
message["name"] = record.name
message["lineno"] = record.lineno
message["message"] = record.msg
json_message = json.dumps(message)
json_message_bytes = json_message.encode("utf8")
urllib.request.urlopen(self.url, data=json_message_bytes)
# end emit
# end ESLogHandler
## Instruction:
Revert "trying to simplefy __init__"
## Code After:
from datetime import datetime
import logging
import os
import json
import urllib.request
import urllib.parse
class ESLogHandler(logging.Handler):
def __init__(self, url, index=None, doc_type="log", level=logging.NOTSET):
logging.Handler.__init__(self, level=level)
# Parse the url
self.url = urllib.parse.urlparse(url)
# If no scheme is given, set it to http
if not self.url.scheme:
self.url.scheme = "http"
# If a scheme is given but it is not http, raise an exception
elif self.url.scheme != "http":
raise ValueError("Only HTTP is supported.")
# If no port is given default to 9200
if not self.url.port:
self.url.port = "9200"
# If no path is given or it is only a / use thi index and doc_type to construct one.
if not self.url.path or self.url.path == "/":
# an index is mandatory for Elasticsearch, doc_type too but it defaults to log
if not index:
raise ValueError("Elasticsearch index cannot be ommitted.")
else:
self.url.path = os.path.join("/", index, doc_type)
# end __init__
def emit(self, record):
# Break the record down to a dictionary
message = dict()
message["timestamp"] = datetime.now().isoformat()
message["level"] = record.levelname
message["name"] = record.name
message["lineno"] = record.lineno
message["message"] = record.msg
json_message = json.dumps(message)
json_message_bytes = json_message.encode("utf8")
urllib.request.urlopen(self.url, data=json_message_bytes)
# end emit
# end ESLogHandler
| // ... existing code ...
def __init__(self, url, index=None, doc_type="log", level=logging.NOTSET):
logging.Handler.__init__(self, level=level)
# Parse the url
self.url = urllib.parse.urlparse(url)
# If no scheme is given, set it to http
if not self.url.scheme:
self.url.scheme = "http"
# If a scheme is given but it is not http, raise an exception
elif self.url.scheme != "http":
raise ValueError("Only HTTP is supported.")
# If no port is given default to 9200
if not self.url.port:
self.url.port = "9200"
# If no path is given or it is only a / use thi index and doc_type to construct one.
if not self.url.path or self.url.path == "/":
# an index is mandatory for Elasticsearch, doc_type too but it defaults to log
if not index:
raise ValueError("Elasticsearch index cannot be ommitted.")
else:
self.url.path = os.path.join("/", index, doc_type)
# end __init__
// ... rest of the code ... |
191d73fb6d30b691da8d9c55bfd36f055aea19d5 | backend/pokehelper.py | backend/pokehelper.py | import json
import os
class Pokehelper(object):
def __init__(self):
basepath = os.path.dirname(__file__)
filepath = os.path.abspath(os.path.join(basepath, 'data/pokemon.json' ))
with open(filepath) as pokejson:
self.pokelist = json.load(pokejson)
###
### LIST STARTS AT 0, EVERY PokeNumber needs a -1
###
def get_pokefamily(self, pokemon_number):
return self.pokelist[pokemon_number-1]['family']
def get_evolution_name(self, pokemon_number):
# NOT a safe method to use, just for testing purposes
return self.pokelist[pokemon_number]['name']
def get_pokename(self, poke_number):
return self.pokelist[poke_number-1]['name']
def get_base_attack(self, poke_number):
return self.pokelist[poke_number-1]['stats']['attack']
def get_base_defense(self, poke_number):
return self.pokelist[poke_number-1]['stats']['defense']
| import json
import os
emptymon = {'moves1': [], 'family': 1, 'name': 'not-in-database', 'moves2': [], 'type2': 'nil', 'id': -1, 'candy': -1, 'type1': 'nil', 'stats': {'stamina': -1, 'attack': -1, 'defense': -1}}
class Pokehelper(object):
def __init__(self):
basepath = os.path.dirname(__file__)
filepath = os.path.abspath(os.path.join(basepath, 'data/pokemon.json'))
with open(filepath) as pokejson:
self.pokelist = json.load(pokejson)
###
# LIST STARTS AT 0, EVERY PokeNumber needs a -1
###
def get_pokefamily(self, poke_number):
if (poke_number > 151):
return emptymon['family']
return self.pokelist[poke_number-1]['family']
def get_evolution_name(self, poke_number):
# NOT a safe method to use, just for testing purposes
if (poke_number > 151):
return emptymon['name']
return self.pokelist[poke_number]['name']
def get_pokename(self, poke_number):
if (poke_number > 151):
return emptymon['name']
return self.pokelist[poke_number-1]['name']
def get_base_attack(self, poke_number):
if (poke_number > 151):
return emptymon['stats']['attack']
return self.pokelist[poke_number-1]['stats']['attack']
def get_base_defense(self, poke_number):
if (poke_number > 151):
return emptymon['stats']['defense']
return self.pokelist[poke_number-1]['stats']['defense']
| Add fallback if pokemon_id > 151 | Add fallback if pokemon_id > 151
| Python | mit | Phaetec/pogo-cruncher,Phaetec/pogo-cruncher,Phaetec/pogo-cruncher | import json
import os
+
+ emptymon = {'moves1': [], 'family': 1, 'name': 'not-in-database', 'moves2': [], 'type2': 'nil', 'id': -1, 'candy': -1, 'type1': 'nil', 'stats': {'stamina': -1, 'attack': -1, 'defense': -1}}
+
class Pokehelper(object):
def __init__(self):
basepath = os.path.dirname(__file__)
- filepath = os.path.abspath(os.path.join(basepath, 'data/pokemon.json' ))
+ filepath = os.path.abspath(os.path.join(basepath, 'data/pokemon.json'))
with open(filepath) as pokejson:
self.pokelist = json.load(pokejson)
- ###
+ ###
- ### LIST STARTS AT 0, EVERY PokeNumber needs a -1
+ # LIST STARTS AT 0, EVERY PokeNumber needs a -1
- ###
+ ###
- def get_pokefamily(self, pokemon_number):
+ def get_pokefamily(self, poke_number):
+ if (poke_number > 151):
+ return emptymon['family']
- return self.pokelist[pokemon_number-1]['family']
+ return self.pokelist[poke_number-1]['family']
- def get_evolution_name(self, pokemon_number):
+ def get_evolution_name(self, poke_number):
# NOT a safe method to use, just for testing purposes
+ if (poke_number > 151):
+ return emptymon['name']
- return self.pokelist[pokemon_number]['name']
+ return self.pokelist[poke_number]['name']
def get_pokename(self, poke_number):
+ if (poke_number > 151):
+ return emptymon['name']
return self.pokelist[poke_number-1]['name']
def get_base_attack(self, poke_number):
+ if (poke_number > 151):
+ return emptymon['stats']['attack']
return self.pokelist[poke_number-1]['stats']['attack']
def get_base_defense(self, poke_number):
+ if (poke_number > 151):
+ return emptymon['stats']['defense']
return self.pokelist[poke_number-1]['stats']['defense']
| Add fallback if pokemon_id > 151 | ## Code Before:
import json
import os
class Pokehelper(object):
def __init__(self):
basepath = os.path.dirname(__file__)
filepath = os.path.abspath(os.path.join(basepath, 'data/pokemon.json' ))
with open(filepath) as pokejson:
self.pokelist = json.load(pokejson)
###
### LIST STARTS AT 0, EVERY PokeNumber needs a -1
###
def get_pokefamily(self, pokemon_number):
return self.pokelist[pokemon_number-1]['family']
def get_evolution_name(self, pokemon_number):
# NOT a safe method to use, just for testing purposes
return self.pokelist[pokemon_number]['name']
def get_pokename(self, poke_number):
return self.pokelist[poke_number-1]['name']
def get_base_attack(self, poke_number):
return self.pokelist[poke_number-1]['stats']['attack']
def get_base_defense(self, poke_number):
return self.pokelist[poke_number-1]['stats']['defense']
## Instruction:
Add fallback if pokemon_id > 151
## Code After:
import json
import os
emptymon = {'moves1': [], 'family': 1, 'name': 'not-in-database', 'moves2': [], 'type2': 'nil', 'id': -1, 'candy': -1, 'type1': 'nil', 'stats': {'stamina': -1, 'attack': -1, 'defense': -1}}
class Pokehelper(object):
def __init__(self):
basepath = os.path.dirname(__file__)
filepath = os.path.abspath(os.path.join(basepath, 'data/pokemon.json'))
with open(filepath) as pokejson:
self.pokelist = json.load(pokejson)
###
# LIST STARTS AT 0, EVERY PokeNumber needs a -1
###
def get_pokefamily(self, poke_number):
if (poke_number > 151):
return emptymon['family']
return self.pokelist[poke_number-1]['family']
def get_evolution_name(self, poke_number):
# NOT a safe method to use, just for testing purposes
if (poke_number > 151):
return emptymon['name']
return self.pokelist[poke_number]['name']
def get_pokename(self, poke_number):
if (poke_number > 151):
return emptymon['name']
return self.pokelist[poke_number-1]['name']
def get_base_attack(self, poke_number):
if (poke_number > 151):
return emptymon['stats']['attack']
return self.pokelist[poke_number-1]['stats']['attack']
def get_base_defense(self, poke_number):
if (poke_number > 151):
return emptymon['stats']['defense']
return self.pokelist[poke_number-1]['stats']['defense']
| # ... existing code ...
import json
import os
emptymon = {'moves1': [], 'family': 1, 'name': 'not-in-database', 'moves2': [], 'type2': 'nil', 'id': -1, 'candy': -1, 'type1': 'nil', 'stats': {'stamina': -1, 'attack': -1, 'defense': -1}}
class Pokehelper(object):
# ... modified code ...
basepath = os.path.dirname(__file__)
filepath = os.path.abspath(os.path.join(basepath, 'data/pokemon.json'))
with open(filepath) as pokejson:
self.pokelist = json.load(pokejson)
###
# LIST STARTS AT 0, EVERY PokeNumber needs a -1
###
def get_pokefamily(self, poke_number):
if (poke_number > 151):
return emptymon['family']
return self.pokelist[poke_number-1]['family']
def get_evolution_name(self, poke_number):
# NOT a safe method to use, just for testing purposes
if (poke_number > 151):
return emptymon['name']
return self.pokelist[poke_number]['name']
def get_pokename(self, poke_number):
if (poke_number > 151):
return emptymon['name']
return self.pokelist[poke_number-1]['name']
def get_base_attack(self, poke_number):
if (poke_number > 151):
return emptymon['stats']['attack']
return self.pokelist[poke_number-1]['stats']['attack']
def get_base_defense(self, poke_number):
if (poke_number > 151):
return emptymon['stats']['defense']
return self.pokelist[poke_number-1]['stats']['defense']
# ... rest of the code ... |
bd7c5c5544a6d09062da05a4780524e8981f1737 | captainhook/checkers/block_branches.py | captainhook/checkers/block_branches.py | import argparse
from .utils import bash
CHECK_NAME = 'block_branch'
def run(files, temp_folder, arg=None):
"Check we're not committing to a blocked branch"
parser = get_parser()
argos = parser.parse_args(arg.split())
current_branch = bash('git symbolic-ref HEAD').value().decode('utf-8')
current_branch = current_branch.replace('refs/heads/', '').strip()
if current_branch in argos.branches:
return ("Branch '{0}' is blocked from being "
"committed to.".format(current_branch))
def get_parser():
parser = argparse.ArgumentParser()
parser.add_argument('branches', metavar='B', nargs='+',
help='a branch to block commits to')
return parser
| import argparse
from .utils import bash
CHECK_NAME = 'block_branch'
def run(files, temp_folder, arg=None):
"Check we're not committing to a blocked branch"
parser = get_parser()
argos = parser.parse_args(arg.split())
current_branch = bash('git symbolic-ref HEAD').value()
current_branch = current_branch.replace('refs/heads/', '').strip()
if current_branch in argos.branches:
return ("Branch '{0}' is blocked from being "
"committed to.".format(current_branch))
def get_parser():
parser = argparse.ArgumentParser()
parser.add_argument('branches', metavar='B', nargs='+',
help='a branch to block commits to')
return parser
| Remove decode from block branches check | Remove decode from block branches check
It’s now done by `bash()`.
| Python | bsd-3-clause | alexcouper/captainhook | import argparse
from .utils import bash
CHECK_NAME = 'block_branch'
def run(files, temp_folder, arg=None):
"Check we're not committing to a blocked branch"
parser = get_parser()
argos = parser.parse_args(arg.split())
- current_branch = bash('git symbolic-ref HEAD').value().decode('utf-8')
+ current_branch = bash('git symbolic-ref HEAD').value()
current_branch = current_branch.replace('refs/heads/', '').strip()
if current_branch in argos.branches:
return ("Branch '{0}' is blocked from being "
"committed to.".format(current_branch))
def get_parser():
parser = argparse.ArgumentParser()
parser.add_argument('branches', metavar='B', nargs='+',
help='a branch to block commits to')
return parser
| Remove decode from block branches check | ## Code Before:
import argparse
from .utils import bash
CHECK_NAME = 'block_branch'
def run(files, temp_folder, arg=None):
"Check we're not committing to a blocked branch"
parser = get_parser()
argos = parser.parse_args(arg.split())
current_branch = bash('git symbolic-ref HEAD').value().decode('utf-8')
current_branch = current_branch.replace('refs/heads/', '').strip()
if current_branch in argos.branches:
return ("Branch '{0}' is blocked from being "
"committed to.".format(current_branch))
def get_parser():
parser = argparse.ArgumentParser()
parser.add_argument('branches', metavar='B', nargs='+',
help='a branch to block commits to')
return parser
## Instruction:
Remove decode from block branches check
## Code After:
import argparse
from .utils import bash
CHECK_NAME = 'block_branch'
def run(files, temp_folder, arg=None):
"Check we're not committing to a blocked branch"
parser = get_parser()
argos = parser.parse_args(arg.split())
current_branch = bash('git symbolic-ref HEAD').value()
current_branch = current_branch.replace('refs/heads/', '').strip()
if current_branch in argos.branches:
return ("Branch '{0}' is blocked from being "
"committed to.".format(current_branch))
def get_parser():
parser = argparse.ArgumentParser()
parser.add_argument('branches', metavar='B', nargs='+',
help='a branch to block commits to')
return parser
| ...
argos = parser.parse_args(arg.split())
current_branch = bash('git symbolic-ref HEAD').value()
current_branch = current_branch.replace('refs/heads/', '').strip()
if current_branch in argos.branches:
... |
e82225201772794bf347c6e768d25f24a61b9b54 | migrations/schematic_settings.py | migrations/schematic_settings.py | import sys
import os
# This only works if you're running schematic from the zamboni root.
sys.path.insert(0, os.path.realpath('.'))
# Set up zamboni.
import manage
from django.conf import settings
config = settings.DATABASES['default']
config['HOST'] = config.get('HOST', 'localhost')
config['PORT'] = config.get('PORT', '3306')
if config['HOST'].endswith('.sock'):
""" Oh you meant 'localhost'! """
config['HOST'] = 'localhost'
s = 'mysql --silent {NAME} -h{HOST} -P{PORT} -u{USER}'
if config['PASSWORD']:
s += ' -p{PASSWORD}'
else:
del config['PASSWORD']
db = s.format(**config)
table = 'schema_version'
| import sys
import os
sys.path.insert(0, os.path.dirname(os.path.dirname(os.path.abspath(__file__))))
# Set up zamboni.
import manage
from django.conf import settings
config = settings.DATABASES['default']
config['HOST'] = config.get('HOST', 'localhost')
config['PORT'] = config.get('PORT', '3306')
if not config['HOST'] or config['HOST'].endswith('.sock'):
""" Oh you meant 'localhost'! """
config['HOST'] = 'localhost'
s = 'mysql --silent {NAME} -h{HOST} -u{USER}'
if config['PASSWORD']:
s += ' -p{PASSWORD}'
else:
del config['PASSWORD']
if config['PORT']:
s += ' -P{PORT}'
else:
del config['PORT']
db = s.format(**config)
table = 'schema_version'
| Make the settings work when there's no port, and fix up the path manipulation | Make the settings work when there's no port, and fix up the path manipulation
| Python | bsd-3-clause | kumar303/zamboni,kmaglione/olympia,Prashant-Surya/addons-server,jamesthechamp/zamboni,yfdyh000/olympia,aviarypl/mozilla-l10n-addons-server,Joergen/zamboni,muffinresearch/addons-server,Jobava/zamboni,koehlermichael/olympia,clouserw/zamboni,kmaglione/olympia,mstriemer/addons-server,Hitechverma/zamboni,mstriemer/olympia,psiinon/addons-server,mozilla/addons-server,lavish205/olympia,Nolski/olympia,spasovski/zamboni,jasonthomas/zamboni,jpetto/olympia,kumar303/olympia,beni55/olympia,crdoconnor/olympia,diox/zamboni,harry-7/addons-server,andymckay/addons-server,clouserw/zamboni,andymckay/olympia,andymckay/zamboni,Nolski/olympia,kmaglione/olympia,SuriyaaKudoIsc/olympia,magopian/olympia,psiinon/addons-server,kumar303/zamboni,mstriemer/zamboni,Joergen/olympia,shahbaz17/zamboni,ayushagrawal288/zamboni,Jobava/zamboni,jbalogh/zamboni,Hitechverma/zamboni,magopian/olympia,atiqueahmedziad/addons-server,Joergen/olympia,tsl143/zamboni,elysium001/zamboni,eviljeff/zamboni,andymckay/olympia,wagnerand/olympia,mozilla/zamboni,johancz/olympia,shahbaz17/zamboni,kmaglione/olympia,SuriyaaKudoIsc/olympia,atiqueahmedziad/addons-server,luckylavish/zamboni,bqbn/addons-server,johancz/olympia,psiinon/addons-server,jamesthechamp/zamboni,robhudson/zamboni,mozilla/olympia,diox/zamboni,anaran/olympia,crdoconnor/olympia,andymckay/zamboni,mudithkr/zamboni,harry-7/addons-server,mdaif/olympia,shahbaz17/zamboni,mozilla/addons-server,atiqueahmedziad/addons-server,mozilla/olympia,Witia1/olympia,ingenioustechie/zamboni,wagnerand/olympia,spasovski/zamboni,Nolski/olympia,robhudson/zamboni,kumar303/addons-server,koehlermichael/olympia,lavish205/olympia,luckylavish/zamboni,muffinresearch/olympia,jasonthomas/zamboni,clouserw/zamboni,ddurst/zamboni,mstriemer/addons-server,diox/zamboni,elysium001/zamboni,shahbaz17/zamboni,ayushagrawal288/zamboni,koehlermichael/olympia,eviljeff/zamboni,ngokevin/zamboni,beni55/olympia,SuriyaaKudoIsc/olympia,atiqueahmedziad/addons-server,kumar303/olympia,Prashant-Surya/addons-server,ngokevin/zamboni,wagnerand/zamboni,robhudson/zamboni,andymckay/addons-server,eviljeff/zamboni,bqbn/addons-server,mrrrgn/olympia,jamesthechamp/zamboni,johancz/olympia,crdoconnor/olympia,crdoconnor/olympia,Nolski/olympia,luckylavish/zamboni,muffinresearch/olympia,robhudson/zamboni,ingenioustechie/zamboni,jpetto/olympia,Jobava/zamboni,muffinresearch/addons-server,jasonthomas/zamboni,spasovski/zamboni,Prashant-Surya/addons-server,mstriemer/zamboni,washort/zamboni,Joergen/zamboni,Prashant-Surya/addons-server,Joergen/zamboni,wagnerand/addons-server,mozilla/olympia,eviljeff/olympia,mozilla/zamboni,Witia1/olympia,tsl143/addons-server,eviljeff/olympia,jbalogh/zamboni,kumar303/addons-server,andymckay/addons-server,Witia1/olympia,wagnerand/zamboni,magopian/olympia,Joergen/zamboni,jbalogh/zamboni,mstriemer/zamboni,washort/zamboni,jpetto/olympia,andymckay/olympia,ddurst/zamboni,elysium001/zamboni,Revanth47/addons-server,aviarypl/mozilla-l10n-addons-server,spasovski/zamboni,wagnerand/zamboni,kumar303/zamboni,diox/olympia,ayushagrawal288/zamboni,lavish205/olympia,mozilla/olympia,mstriemer/olympia,yfdyh000/olympia,harikishen/addons-server,andymckay/olympia,washort/zamboni,Nolski/olympia,ddurst/zamboni,Joergen/olympia,tsl143/zamboni,harikishen/addons-server,wagnerand/olympia,muffinresearch/addons-server,tsl143/addons-server,diox/olympia,washort/zamboni,clouserw/zamboni,luckylavish/zamboni,mstriemer/addons-server,anaran/olympia,beni55/olympia,jasonthomas/zamboni,mudithkr/zamboni,mrrrgn/olympia,crdoconnor/olympia,mrrrgn/olympia,kumar303/addons-server,tsl143/addons-server,magopian/olympia,andymckay/zamboni,johancz/olympia,lavish205/olympia,tsl143/addons-server,ngokevin/zamboni,mozilla/addons-server,aviarypl/mozilla-l10n-addons-server,mozilla/zamboni,Witia1/olympia,Jobava/zamboni,beni55/olympia,muffinresearch/olympia,mrrrgn/olympia,mstriemer/olympia,Joergen/olympia,muffinresearch/addons-server,wagnerand/addons-server,Revanth47/addons-server,SuriyaaKudoIsc/olympia,beni55/olympia,diox/zamboni,Witia1/olympia,andymckay/addons-server,mstriemer/zamboni,mozilla/zamboni,mdaif/olympia,Hitechverma/zamboni,Revanth47/addons-server,diox/olympia,mudithkr/zamboni,ddurst/zamboni,anaran/olympia,ingenioustechie/zamboni,tsl143/zamboni,Joergen/zamboni,tsl143/zamboni,diox/olympia,yfdyh000/olympia,magopian/olympia,mdaif/olympia,kumar303/zamboni,jpetto/olympia,Joergen/zamboni,yfdyh000/olympia,ayushagrawal288/zamboni,kumar303/olympia,wagnerand/addons-server,mdaif/olympia,muffinresearch/olympia,harry-7/addons-server,kmaglione/olympia,elysium001/zamboni,jamesthechamp/zamboni,koehlermichael/olympia,psiinon/addons-server,johancz/olympia,yfdyh000/olympia,mudithkr/zamboni,eviljeff/olympia,mrrrgn/olympia,ingenioustechie/zamboni,anaran/olympia,wagnerand/olympia,bqbn/addons-server,wagnerand/addons-server,kumar303/olympia,eviljeff/zamboni,aviarypl/mozilla-l10n-addons-server,Revanth47/addons-server,harry-7/addons-server,mstriemer/addons-server,wagnerand/zamboni,harikishen/addons-server,mozilla/addons-server,Joergen/olympia,kumar303/addons-server,eviljeff/olympia,jbalogh/zamboni,Hitechverma/zamboni,mdaif/olympia,mstriemer/olympia,harikishen/addons-server,koehlermichael/olympia,muffinresearch/olympia,bqbn/addons-server,muffinresearch/addons-server | import sys
import os
+ sys.path.insert(0, os.path.dirname(os.path.dirname(os.path.abspath(__file__))))
- # This only works if you're running schematic from the zamboni root.
- sys.path.insert(0, os.path.realpath('.'))
# Set up zamboni.
import manage
from django.conf import settings
config = settings.DATABASES['default']
config['HOST'] = config.get('HOST', 'localhost')
config['PORT'] = config.get('PORT', '3306')
- if config['HOST'].endswith('.sock'):
+ if not config['HOST'] or config['HOST'].endswith('.sock'):
""" Oh you meant 'localhost'! """
config['HOST'] = 'localhost'
- s = 'mysql --silent {NAME} -h{HOST} -P{PORT} -u{USER}'
+ s = 'mysql --silent {NAME} -h{HOST} -u{USER}'
if config['PASSWORD']:
s += ' -p{PASSWORD}'
else:
del config['PASSWORD']
+ if config['PORT']:
+ s += ' -P{PORT}'
+ else:
+ del config['PORT']
db = s.format(**config)
table = 'schema_version'
| Make the settings work when there's no port, and fix up the path manipulation | ## Code Before:
import sys
import os
# This only works if you're running schematic from the zamboni root.
sys.path.insert(0, os.path.realpath('.'))
# Set up zamboni.
import manage
from django.conf import settings
config = settings.DATABASES['default']
config['HOST'] = config.get('HOST', 'localhost')
config['PORT'] = config.get('PORT', '3306')
if config['HOST'].endswith('.sock'):
""" Oh you meant 'localhost'! """
config['HOST'] = 'localhost'
s = 'mysql --silent {NAME} -h{HOST} -P{PORT} -u{USER}'
if config['PASSWORD']:
s += ' -p{PASSWORD}'
else:
del config['PASSWORD']
db = s.format(**config)
table = 'schema_version'
## Instruction:
Make the settings work when there's no port, and fix up the path manipulation
## Code After:
import sys
import os
sys.path.insert(0, os.path.dirname(os.path.dirname(os.path.abspath(__file__))))
# Set up zamboni.
import manage
from django.conf import settings
config = settings.DATABASES['default']
config['HOST'] = config.get('HOST', 'localhost')
config['PORT'] = config.get('PORT', '3306')
if not config['HOST'] or config['HOST'].endswith('.sock'):
""" Oh you meant 'localhost'! """
config['HOST'] = 'localhost'
s = 'mysql --silent {NAME} -h{HOST} -u{USER}'
if config['PASSWORD']:
s += ' -p{PASSWORD}'
else:
del config['PASSWORD']
if config['PORT']:
s += ' -P{PORT}'
else:
del config['PORT']
db = s.format(**config)
table = 'schema_version'
| ...
import os
sys.path.insert(0, os.path.dirname(os.path.dirname(os.path.abspath(__file__))))
# Set up zamboni.
...
config['PORT'] = config.get('PORT', '3306')
if not config['HOST'] or config['HOST'].endswith('.sock'):
""" Oh you meant 'localhost'! """
config['HOST'] = 'localhost'
s = 'mysql --silent {NAME} -h{HOST} -u{USER}'
if config['PASSWORD']:
...
else:
del config['PASSWORD']
if config['PORT']:
s += ' -P{PORT}'
else:
del config['PORT']
db = s.format(**config)
... |
79c9ee6107b841986054915c23f8456c80097c5b | osgtest/tests/test_13_gridftp.py | osgtest/tests/test_13_gridftp.py | import os
import osgtest.library.core as core
import unittest
class TestStartGridFTP(unittest.TestCase):
def test_01_start_gridftp(self):
core.config['gridftp.pid-file'] = '/var/run/globus-gridftp-server.pid'
core.state['gridftp.started-server'] = False
if not core.rpm_is_installed('globus-gridftp-server-progs'):
core.skip('not installed')
return
if os.path.exists(core.config['gridftp.pid-file']):
core.skip('apparently running')
return
command = ('service', 'globus-gridftp-server', 'start')
stdout, _, fail = core.check_system(command, 'Start GridFTP server')
self.assert_(stdout.find('FAILED') == -1, fail)
self.assert_(os.path.exists(core.config['gridftp.pid-file']),
'GridFTP server PID file missing')
core.state['gridftp.started-server'] = True
| import os
from osgtest.library import core, osgunittest
import unittest
class TestStartGridFTP(osgunittest.OSGTestCase):
def test_01_start_gridftp(self):
core.config['gridftp.pid-file'] = '/var/run/globus-gridftp-server.pid'
core.state['gridftp.started-server'] = False
core.skip_ok_unless_installed('globus-gridftp-server-progs')
self.skip_ok_if(os.path.exists(core.config['gridftp.pid-file']), 'already running')
command = ('service', 'globus-gridftp-server', 'start')
stdout, _, fail = core.check_system(command, 'Start GridFTP server')
self.assert_(stdout.find('FAILED') == -1, fail)
self.assert_(os.path.exists(core.config['gridftp.pid-file']),
'GridFTP server PID file missing')
core.state['gridftp.started-server'] = True
| Update 13_gridftp to use OkSkip functionality | Update 13_gridftp to use OkSkip functionality
git-svn-id: 884a03e47e2adb735d896e55bb5ad6bc3421ba19@16527 4e558342-562e-0410-864c-e07659590f8c
| Python | apache-2.0 | efajardo/osg-test,efajardo/osg-test | import os
- import osgtest.library.core as core
+ from osgtest.library import core, osgunittest
import unittest
- class TestStartGridFTP(unittest.TestCase):
+ class TestStartGridFTP(osgunittest.OSGTestCase):
def test_01_start_gridftp(self):
core.config['gridftp.pid-file'] = '/var/run/globus-gridftp-server.pid'
core.state['gridftp.started-server'] = False
- if not core.rpm_is_installed('globus-gridftp-server-progs'):
+ core.skip_ok_unless_installed('globus-gridftp-server-progs')
- core.skip('not installed')
- return
- if os.path.exists(core.config['gridftp.pid-file']):
+ self.skip_ok_if(os.path.exists(core.config['gridftp.pid-file']), 'already running')
- core.skip('apparently running')
- return
command = ('service', 'globus-gridftp-server', 'start')
stdout, _, fail = core.check_system(command, 'Start GridFTP server')
self.assert_(stdout.find('FAILED') == -1, fail)
self.assert_(os.path.exists(core.config['gridftp.pid-file']),
'GridFTP server PID file missing')
core.state['gridftp.started-server'] = True
| Update 13_gridftp to use OkSkip functionality | ## Code Before:
import os
import osgtest.library.core as core
import unittest
class TestStartGridFTP(unittest.TestCase):
def test_01_start_gridftp(self):
core.config['gridftp.pid-file'] = '/var/run/globus-gridftp-server.pid'
core.state['gridftp.started-server'] = False
if not core.rpm_is_installed('globus-gridftp-server-progs'):
core.skip('not installed')
return
if os.path.exists(core.config['gridftp.pid-file']):
core.skip('apparently running')
return
command = ('service', 'globus-gridftp-server', 'start')
stdout, _, fail = core.check_system(command, 'Start GridFTP server')
self.assert_(stdout.find('FAILED') == -1, fail)
self.assert_(os.path.exists(core.config['gridftp.pid-file']),
'GridFTP server PID file missing')
core.state['gridftp.started-server'] = True
## Instruction:
Update 13_gridftp to use OkSkip functionality
## Code After:
import os
from osgtest.library import core, osgunittest
import unittest
class TestStartGridFTP(osgunittest.OSGTestCase):
def test_01_start_gridftp(self):
core.config['gridftp.pid-file'] = '/var/run/globus-gridftp-server.pid'
core.state['gridftp.started-server'] = False
core.skip_ok_unless_installed('globus-gridftp-server-progs')
self.skip_ok_if(os.path.exists(core.config['gridftp.pid-file']), 'already running')
command = ('service', 'globus-gridftp-server', 'start')
stdout, _, fail = core.check_system(command, 'Start GridFTP server')
self.assert_(stdout.find('FAILED') == -1, fail)
self.assert_(os.path.exists(core.config['gridftp.pid-file']),
'GridFTP server PID file missing')
core.state['gridftp.started-server'] = True
| ...
import os
from osgtest.library import core, osgunittest
import unittest
class TestStartGridFTP(osgunittest.OSGTestCase):
def test_01_start_gridftp(self):
...
core.state['gridftp.started-server'] = False
core.skip_ok_unless_installed('globus-gridftp-server-progs')
self.skip_ok_if(os.path.exists(core.config['gridftp.pid-file']), 'already running')
command = ('service', 'globus-gridftp-server', 'start')
... |
0b28fe44514969470db926c6f38615a8a5478bf6 | smoke_signal/__init__.py | smoke_signal/__init__.py | from flask import Flask, g
from .main.views import main
from .nojs.views import nojs
from sqlalchemy import create_engine
from smoke_signal.database.models import Base
from sqlalchemy.orm import sessionmaker
app = Flask(__name__, instance_relative_config=True)
app.config.from_object("config")
app.config.from_pyfile("config.py")
app.register_blueprint(main)
app.register_blueprint(nojs)
@app.before_request
def init_db():
engine = create_engine(app.config["DATABASE_PATH"])
if not engine.dialect.has_table(engine.connect(), "feed"):
Base.metadata.create_all(engine)
Session = sessionmaker(bind=engine)
g.db = Session()
@app.teardown_appcontext
def shutdown_session(exception=None):
db = getattr(g, 'db', None)
if db is not None:
g.db.close()
| from flask import Flask, g
from .main.views import main
from sqlalchemy import create_engine
from smoke_signal.database.models import Base
from sqlalchemy.orm import sessionmaker
app = Flask(__name__, instance_relative_config=True)
app.config.from_object("config")
app.config.from_pyfile("config.py")
app.register_blueprint(main)
@app.before_request
def init_db():
engine = create_engine(app.config["DATABASE_PATH"])
if not engine.dialect.has_table(engine.connect(), "feed"):
Base.metadata.create_all(engine)
Session = sessionmaker(bind=engine)
g.db = Session()
@app.teardown_appcontext
def shutdown_session(exception=None):
db = getattr(g, 'db', None)
if db is not None:
g.db.close()
| Remove the no-JS version from the app | Remove the no-JS version from the app
I haven't looked into it for a long while.
| Python | mit | flacerdk/smoke-signal,flacerdk/smoke-signal,flacerdk/smoke-signal | from flask import Flask, g
from .main.views import main
- from .nojs.views import nojs
from sqlalchemy import create_engine
from smoke_signal.database.models import Base
from sqlalchemy.orm import sessionmaker
app = Flask(__name__, instance_relative_config=True)
app.config.from_object("config")
app.config.from_pyfile("config.py")
app.register_blueprint(main)
- app.register_blueprint(nojs)
@app.before_request
def init_db():
engine = create_engine(app.config["DATABASE_PATH"])
if not engine.dialect.has_table(engine.connect(), "feed"):
Base.metadata.create_all(engine)
Session = sessionmaker(bind=engine)
g.db = Session()
@app.teardown_appcontext
def shutdown_session(exception=None):
db = getattr(g, 'db', None)
if db is not None:
g.db.close()
| Remove the no-JS version from the app | ## Code Before:
from flask import Flask, g
from .main.views import main
from .nojs.views import nojs
from sqlalchemy import create_engine
from smoke_signal.database.models import Base
from sqlalchemy.orm import sessionmaker
app = Flask(__name__, instance_relative_config=True)
app.config.from_object("config")
app.config.from_pyfile("config.py")
app.register_blueprint(main)
app.register_blueprint(nojs)
@app.before_request
def init_db():
engine = create_engine(app.config["DATABASE_PATH"])
if not engine.dialect.has_table(engine.connect(), "feed"):
Base.metadata.create_all(engine)
Session = sessionmaker(bind=engine)
g.db = Session()
@app.teardown_appcontext
def shutdown_session(exception=None):
db = getattr(g, 'db', None)
if db is not None:
g.db.close()
## Instruction:
Remove the no-JS version from the app
## Code After:
from flask import Flask, g
from .main.views import main
from sqlalchemy import create_engine
from smoke_signal.database.models import Base
from sqlalchemy.orm import sessionmaker
app = Flask(__name__, instance_relative_config=True)
app.config.from_object("config")
app.config.from_pyfile("config.py")
app.register_blueprint(main)
@app.before_request
def init_db():
engine = create_engine(app.config["DATABASE_PATH"])
if not engine.dialect.has_table(engine.connect(), "feed"):
Base.metadata.create_all(engine)
Session = sessionmaker(bind=engine)
g.db = Session()
@app.teardown_appcontext
def shutdown_session(exception=None):
db = getattr(g, 'db', None)
if db is not None:
g.db.close()
| // ... existing code ...
from flask import Flask, g
from .main.views import main
from sqlalchemy import create_engine
from smoke_signal.database.models import Base
// ... modified code ...
app.config.from_pyfile("config.py")
app.register_blueprint(main)
// ... rest of the code ... |
3039b00e761f02eb0586dad51049377a31329491 | reggae/reflect.py | reggae/reflect.py | from __future__ import (unicode_literals, division,
absolute_import, print_function)
from reggae.build import Build, DefaultOptions
from inspect import getmembers
def get_build(module):
builds = [v for n, v in getmembers(module) if isinstance(v, Build)]
assert len(builds) == 1
return builds[0]
def get_default_options(module):
opts = [v for n, v in getmembers(module) if isinstance(v, DefaultOptions)]
assert len(opts) == 1 or len(opts) == 0
return opts[0] if len(opts) else None
def get_dependencies(module):
from modulefinder import ModuleFinder
import os
finder = ModuleFinder()
finder.run_script(module)
all_module_paths = [m.__file__ for m in finder.modules.values()]
def is_in_same_path(p):
return p and os.path.dirname(p).startswith(os.path.dirname(module))
return [x for x in all_module_paths if is_in_same_path(x) and x != module]
| from __future__ import (unicode_literals, division,
absolute_import, print_function)
from reggae.build import Build, DefaultOptions
from inspect import getmembers
def get_build(module):
builds = [v for n, v in getmembers(module) if isinstance(v, Build)]
assert len(builds) == 1
return builds[0]
def get_default_options(module):
opts = [v for n, v in getmembers(module) if isinstance(v, DefaultOptions)]
assert len(opts) == 1 or len(opts) == 0
return opts[0] if len(opts) else None
def get_dependencies(module):
from modulefinder import ModuleFinder
import os
finder = ModuleFinder()
finder.run_script(module)
all_module_paths = [os.path.abspath(m.__file__) for
m in finder.modules.values() if m.__file__ is not None]
def is_in_same_path(p):
return p and os.path.dirname(p).startswith(os.path.dirname(module))
return [x for x in all_module_paths if is_in_same_path(x) and x != module]
| Use absolute paths for dependencies | Use absolute paths for dependencies
| Python | bsd-3-clause | atilaneves/reggae-python | from __future__ import (unicode_literals, division,
absolute_import, print_function)
from reggae.build import Build, DefaultOptions
from inspect import getmembers
def get_build(module):
builds = [v for n, v in getmembers(module) if isinstance(v, Build)]
assert len(builds) == 1
return builds[0]
def get_default_options(module):
opts = [v for n, v in getmembers(module) if isinstance(v, DefaultOptions)]
assert len(opts) == 1 or len(opts) == 0
return opts[0] if len(opts) else None
def get_dependencies(module):
from modulefinder import ModuleFinder
import os
finder = ModuleFinder()
finder.run_script(module)
- all_module_paths = [m.__file__ for m in finder.modules.values()]
+ all_module_paths = [os.path.abspath(m.__file__) for
+ m in finder.modules.values() if m.__file__ is not None]
def is_in_same_path(p):
return p and os.path.dirname(p).startswith(os.path.dirname(module))
return [x for x in all_module_paths if is_in_same_path(x) and x != module]
| Use absolute paths for dependencies | ## Code Before:
from __future__ import (unicode_literals, division,
absolute_import, print_function)
from reggae.build import Build, DefaultOptions
from inspect import getmembers
def get_build(module):
builds = [v for n, v in getmembers(module) if isinstance(v, Build)]
assert len(builds) == 1
return builds[0]
def get_default_options(module):
opts = [v for n, v in getmembers(module) if isinstance(v, DefaultOptions)]
assert len(opts) == 1 or len(opts) == 0
return opts[0] if len(opts) else None
def get_dependencies(module):
from modulefinder import ModuleFinder
import os
finder = ModuleFinder()
finder.run_script(module)
all_module_paths = [m.__file__ for m in finder.modules.values()]
def is_in_same_path(p):
return p and os.path.dirname(p).startswith(os.path.dirname(module))
return [x for x in all_module_paths if is_in_same_path(x) and x != module]
## Instruction:
Use absolute paths for dependencies
## Code After:
from __future__ import (unicode_literals, division,
absolute_import, print_function)
from reggae.build import Build, DefaultOptions
from inspect import getmembers
def get_build(module):
builds = [v for n, v in getmembers(module) if isinstance(v, Build)]
assert len(builds) == 1
return builds[0]
def get_default_options(module):
opts = [v for n, v in getmembers(module) if isinstance(v, DefaultOptions)]
assert len(opts) == 1 or len(opts) == 0
return opts[0] if len(opts) else None
def get_dependencies(module):
from modulefinder import ModuleFinder
import os
finder = ModuleFinder()
finder.run_script(module)
all_module_paths = [os.path.abspath(m.__file__) for
m in finder.modules.values() if m.__file__ is not None]
def is_in_same_path(p):
return p and os.path.dirname(p).startswith(os.path.dirname(module))
return [x for x in all_module_paths if is_in_same_path(x) and x != module]
| ...
finder = ModuleFinder()
finder.run_script(module)
all_module_paths = [os.path.abspath(m.__file__) for
m in finder.modules.values() if m.__file__ is not None]
def is_in_same_path(p):
... |
194a84b4559449f0b0e3e9cc9e7026392822c0af | questions/urls.py | questions/urls.py | from django.conf.urls.defaults import *
from spenglr.education.models import *
# Uncomment the next two lines to enable the admin:
from django.contrib import admin
admin.autodiscover()
urlpatterns = patterns('',
(r'^m/(?P<object_id>\d+)$', 'django.views.generic.list_detail.object_detail', { 'queryset': Module.objects.all(), 'template_name': 'education/module_questions.html' } ),
)
| from django.conf.urls.defaults import *
from spenglr.education.models import *
# Uncomment the next two lines to enable the admin:
from django.contrib import admin
admin.autodiscover()
urlpatterns = patterns('',
(r'^m/(?P<object_id>\d+)$', 'django.views.generic.list_detail.object_detail', { 'queryset': Module.objects.all(), 'template_name': 'questions/question_list.html' } ),
)
| Change to questions template location. | Change to questions template location.
| Python | bsd-3-clause | mfitzp/smrtr,mfitzp/smrtr | from django.conf.urls.defaults import *
from spenglr.education.models import *
# Uncomment the next two lines to enable the admin:
from django.contrib import admin
admin.autodiscover()
urlpatterns = patterns('',
- (r'^m/(?P<object_id>\d+)$', 'django.views.generic.list_detail.object_detail', { 'queryset': Module.objects.all(), 'template_name': 'education/module_questions.html' } ),
+ (r'^m/(?P<object_id>\d+)$', 'django.views.generic.list_detail.object_detail', { 'queryset': Module.objects.all(), 'template_name': 'questions/question_list.html' } ),
)
| Change to questions template location. | ## Code Before:
from django.conf.urls.defaults import *
from spenglr.education.models import *
# Uncomment the next two lines to enable the admin:
from django.contrib import admin
admin.autodiscover()
urlpatterns = patterns('',
(r'^m/(?P<object_id>\d+)$', 'django.views.generic.list_detail.object_detail', { 'queryset': Module.objects.all(), 'template_name': 'education/module_questions.html' } ),
)
## Instruction:
Change to questions template location.
## Code After:
from django.conf.urls.defaults import *
from spenglr.education.models import *
# Uncomment the next two lines to enable the admin:
from django.contrib import admin
admin.autodiscover()
urlpatterns = patterns('',
(r'^m/(?P<object_id>\d+)$', 'django.views.generic.list_detail.object_detail', { 'queryset': Module.objects.all(), 'template_name': 'questions/question_list.html' } ),
)
| ...
urlpatterns = patterns('',
(r'^m/(?P<object_id>\d+)$', 'django.views.generic.list_detail.object_detail', { 'queryset': Module.objects.all(), 'template_name': 'questions/question_list.html' } ),
)
... |
56e764835e75035452a6a1ea06c386ec61dbe872 | src/rinoh/stylesheets/__init__.py | src/rinoh/stylesheets/__init__.py |
import inspect
import os
import sys
from .. import DATA_PATH
from ..style import StyleSheetFile
from .matcher import matcher
__all__ = ['matcher', 'sphinx', 'sphinx_base14']
STYLESHEETS_PATH = os.path.join(DATA_PATH, 'stylesheets')
def path(filename):
return os.path.join(STYLESHEETS_PATH, filename)
sphinx = StyleSheetFile(path('sphinx.rts'))
sphinx_article = StyleSheetFile(path('sphinx_article.rts'))
sphinx_base14 = StyleSheetFile(path('base14.rts'))
# generate docstrings for the StyleSheet instances
for name, stylesheet in inspect.getmembers(sys.modules[__name__]):
if not isinstance(stylesheet, StyleSheetFile):
continue
stylesheet.__doc__ = (':entry point name: ``{}``\n\n{}'
.format(stylesheet, stylesheet.description))
|
import inspect
import os
import sys
from .. import DATA_PATH
from ..style import StyleSheetFile
from .matcher import matcher
__all__ = ['matcher', 'sphinx', 'sphinx_base14']
STYLESHEETS_PATH = os.path.join(DATA_PATH, 'stylesheets')
def path(filename):
return os.path.join(STYLESHEETS_PATH, filename)
sphinx = StyleSheetFile(path('sphinx.rts'))
sphinx_article = StyleSheetFile(path('sphinx_article.rts'))
sphinx_base14 = StyleSheetFile(path('base14.rts'))
# generate docstrings for the StyleSheet instances
for name, stylesheet in inspect.getmembers(sys.modules[__name__]):
if not isinstance(stylesheet, StyleSheetFile):
continue
stylesheet.__doc__ = ('{}\n\nEntry point name: ``{}``'
.format(stylesheet.description, stylesheet))
| Fix the auto-generated docstrings of style sheets | Fix the auto-generated docstrings of style sheets
| Python | agpl-3.0 | brechtm/rinohtype,brechtm/rinohtype,brechtm/rinohtype |
import inspect
import os
import sys
from .. import DATA_PATH
from ..style import StyleSheetFile
from .matcher import matcher
__all__ = ['matcher', 'sphinx', 'sphinx_base14']
STYLESHEETS_PATH = os.path.join(DATA_PATH, 'stylesheets')
def path(filename):
return os.path.join(STYLESHEETS_PATH, filename)
sphinx = StyleSheetFile(path('sphinx.rts'))
sphinx_article = StyleSheetFile(path('sphinx_article.rts'))
sphinx_base14 = StyleSheetFile(path('base14.rts'))
# generate docstrings for the StyleSheet instances
for name, stylesheet in inspect.getmembers(sys.modules[__name__]):
if not isinstance(stylesheet, StyleSheetFile):
continue
- stylesheet.__doc__ = (':entry point name: ``{}``\n\n{}'
+ stylesheet.__doc__ = ('{}\n\nEntry point name: ``{}``'
- .format(stylesheet, stylesheet.description))
+ .format(stylesheet.description, stylesheet))
| Fix the auto-generated docstrings of style sheets | ## Code Before:
import inspect
import os
import sys
from .. import DATA_PATH
from ..style import StyleSheetFile
from .matcher import matcher
__all__ = ['matcher', 'sphinx', 'sphinx_base14']
STYLESHEETS_PATH = os.path.join(DATA_PATH, 'stylesheets')
def path(filename):
return os.path.join(STYLESHEETS_PATH, filename)
sphinx = StyleSheetFile(path('sphinx.rts'))
sphinx_article = StyleSheetFile(path('sphinx_article.rts'))
sphinx_base14 = StyleSheetFile(path('base14.rts'))
# generate docstrings for the StyleSheet instances
for name, stylesheet in inspect.getmembers(sys.modules[__name__]):
if not isinstance(stylesheet, StyleSheetFile):
continue
stylesheet.__doc__ = (':entry point name: ``{}``\n\n{}'
.format(stylesheet, stylesheet.description))
## Instruction:
Fix the auto-generated docstrings of style sheets
## Code After:
import inspect
import os
import sys
from .. import DATA_PATH
from ..style import StyleSheetFile
from .matcher import matcher
__all__ = ['matcher', 'sphinx', 'sphinx_base14']
STYLESHEETS_PATH = os.path.join(DATA_PATH, 'stylesheets')
def path(filename):
return os.path.join(STYLESHEETS_PATH, filename)
sphinx = StyleSheetFile(path('sphinx.rts'))
sphinx_article = StyleSheetFile(path('sphinx_article.rts'))
sphinx_base14 = StyleSheetFile(path('base14.rts'))
# generate docstrings for the StyleSheet instances
for name, stylesheet in inspect.getmembers(sys.modules[__name__]):
if not isinstance(stylesheet, StyleSheetFile):
continue
stylesheet.__doc__ = ('{}\n\nEntry point name: ``{}``'
.format(stylesheet.description, stylesheet))
| ...
if not isinstance(stylesheet, StyleSheetFile):
continue
stylesheet.__doc__ = ('{}\n\nEntry point name: ``{}``'
.format(stylesheet.description, stylesheet))
... |
d7a227ae5f0f53b5c620864df08c7b883402e968 | netmiko/brocade/brocade_nos_ssh.py | netmiko/brocade/brocade_nos_ssh.py | """Support for Brocade NOS/VDX."""
from __future__ import unicode_literals
import time
from netmiko.cisco_base_connection import CiscoSSHConnection
class BrocadeNosSSH(CiscoSSHConnection):
"""Support for Brocade NOS/VDX."""
def enable(self, *args, **kwargs):
"""No enable mode on Brocade VDX."""
pass
def exit_enable_mode(self, *args, **kwargs):
"""No enable mode on Brocade VDX."""
pass
def special_login_handler(self, delay_factor=1):
"""Adding a delay after login."""
delay_factor = self.select_delay_factor(delay_factor)
self.write_channel(self.RETURN)
time.sleep(1 * delay_factor)
| """Support for Brocade NOS/VDX."""
from __future__ import unicode_literals
import time
from netmiko.cisco_base_connection import CiscoSSHConnection
class BrocadeNosSSH(CiscoSSHConnection):
"""Support for Brocade NOS/VDX."""
def enable(self, *args, **kwargs):
"""No enable mode on Brocade VDX."""
pass
def exit_enable_mode(self, *args, **kwargs):
"""No enable mode on Brocade VDX."""
pass
def special_login_handler(self, delay_factor=1):
"""Adding a delay after login."""
delay_factor = self.select_delay_factor(delay_factor)
self.write_channel(self.RETURN)
time.sleep(1 * delay_factor)
def save_config(self):
"""Save Config for Brocade VDX."""
self.send_command('copy running-config startup-config', '[Y/N]')
self.send_command('y')
| Add save_config for brocade VDX | Add save_config for brocade VDX
| Python | mit | ktbyers/netmiko,ktbyers/netmiko | """Support for Brocade NOS/VDX."""
from __future__ import unicode_literals
import time
from netmiko.cisco_base_connection import CiscoSSHConnection
class BrocadeNosSSH(CiscoSSHConnection):
"""Support for Brocade NOS/VDX."""
def enable(self, *args, **kwargs):
"""No enable mode on Brocade VDX."""
pass
def exit_enable_mode(self, *args, **kwargs):
"""No enable mode on Brocade VDX."""
pass
def special_login_handler(self, delay_factor=1):
"""Adding a delay after login."""
delay_factor = self.select_delay_factor(delay_factor)
self.write_channel(self.RETURN)
time.sleep(1 * delay_factor)
+ def save_config(self):
+ """Save Config for Brocade VDX."""
+ self.send_command('copy running-config startup-config', '[Y/N]')
+ self.send_command('y')
+ | Add save_config for brocade VDX | ## Code Before:
"""Support for Brocade NOS/VDX."""
from __future__ import unicode_literals
import time
from netmiko.cisco_base_connection import CiscoSSHConnection
class BrocadeNosSSH(CiscoSSHConnection):
"""Support for Brocade NOS/VDX."""
def enable(self, *args, **kwargs):
"""No enable mode on Brocade VDX."""
pass
def exit_enable_mode(self, *args, **kwargs):
"""No enable mode on Brocade VDX."""
pass
def special_login_handler(self, delay_factor=1):
"""Adding a delay after login."""
delay_factor = self.select_delay_factor(delay_factor)
self.write_channel(self.RETURN)
time.sleep(1 * delay_factor)
## Instruction:
Add save_config for brocade VDX
## Code After:
"""Support for Brocade NOS/VDX."""
from __future__ import unicode_literals
import time
from netmiko.cisco_base_connection import CiscoSSHConnection
class BrocadeNosSSH(CiscoSSHConnection):
"""Support for Brocade NOS/VDX."""
def enable(self, *args, **kwargs):
"""No enable mode on Brocade VDX."""
pass
def exit_enable_mode(self, *args, **kwargs):
"""No enable mode on Brocade VDX."""
pass
def special_login_handler(self, delay_factor=1):
"""Adding a delay after login."""
delay_factor = self.select_delay_factor(delay_factor)
self.write_channel(self.RETURN)
time.sleep(1 * delay_factor)
def save_config(self):
"""Save Config for Brocade VDX."""
self.send_command('copy running-config startup-config', '[Y/N]')
self.send_command('y')
| ...
self.write_channel(self.RETURN)
time.sleep(1 * delay_factor)
def save_config(self):
"""Save Config for Brocade VDX."""
self.send_command('copy running-config startup-config', '[Y/N]')
self.send_command('y')
... |
81b1cf6973dde3ca23bbe5ac071d3decad81079a | pydsa/sleep_sort.py | pydsa/sleep_sort.py | from time import sleep
from threading import Timer
# Sleep Sort ;)
# Complexity: O(max(input)+n)
def sleep_sort(a):
"""
Sorts the list 'a' using Sleep sort algorithm
>>> from pydsa import sleep_sort
>>> a = [3, 4, 2]
>>> sleep_sort(a)
[2, 3, 4]
"""
sleep_sort.result = []
def add1(x):
sleep_sort.result.append(x)
mx = a[0]
for v in a:
if mx < v: mx = v
Timer(v, add1, [v]).start()
sleep(mx + 1)
return sleep_sort.result
| from time import sleep
from threading import Timer
# Sleep Sort ;)
# Complexity: O(max(input)+n)
def sleep_sort(a):
"""
Sorts the list 'a' using Sleep sort algorithm
>>> from pydsa import sleep_sort
>>> a = [3, 4, 2]
>>> sleep_sort(a)
[2, 3, 4]
"""
sleep_sort.result = []
def add1(x):
sleep_sort.result.append(x)
mx = a[0]
for v in a:
if mx < v:
mx = v
Timer(v, add1, [v]).start()
sleep(mx + 1)
return sleep_sort.result
| Format code according to PEP8 | Format code according to PEP8
| Python | bsd-3-clause | rehassachdeva/pydsa,aktech/pydsa | from time import sleep
from threading import Timer
# Sleep Sort ;)
# Complexity: O(max(input)+n)
+
def sleep_sort(a):
"""
Sorts the list 'a' using Sleep sort algorithm
>>> from pydsa import sleep_sort
>>> a = [3, 4, 2]
>>> sleep_sort(a)
[2, 3, 4]
"""
sleep_sort.result = []
+
def add1(x):
sleep_sort.result.append(x)
+
mx = a[0]
for v in a:
- if mx < v: mx = v
+ if mx < v:
+ mx = v
Timer(v, add1, [v]).start()
sleep(mx + 1)
return sleep_sort.result
| Format code according to PEP8 | ## Code Before:
from time import sleep
from threading import Timer
# Sleep Sort ;)
# Complexity: O(max(input)+n)
def sleep_sort(a):
"""
Sorts the list 'a' using Sleep sort algorithm
>>> from pydsa import sleep_sort
>>> a = [3, 4, 2]
>>> sleep_sort(a)
[2, 3, 4]
"""
sleep_sort.result = []
def add1(x):
sleep_sort.result.append(x)
mx = a[0]
for v in a:
if mx < v: mx = v
Timer(v, add1, [v]).start()
sleep(mx + 1)
return sleep_sort.result
## Instruction:
Format code according to PEP8
## Code After:
from time import sleep
from threading import Timer
# Sleep Sort ;)
# Complexity: O(max(input)+n)
def sleep_sort(a):
"""
Sorts the list 'a' using Sleep sort algorithm
>>> from pydsa import sleep_sort
>>> a = [3, 4, 2]
>>> sleep_sort(a)
[2, 3, 4]
"""
sleep_sort.result = []
def add1(x):
sleep_sort.result.append(x)
mx = a[0]
for v in a:
if mx < v:
mx = v
Timer(v, add1, [v]).start()
sleep(mx + 1)
return sleep_sort.result
| ...
# Sleep Sort ;)
# Complexity: O(max(input)+n)
def sleep_sort(a):
...
"""
sleep_sort.result = []
def add1(x):
sleep_sort.result.append(x)
mx = a[0]
for v in a:
if mx < v:
mx = v
Timer(v, add1, [v]).start()
sleep(mx + 1)
... |
39c777d6fc5555534628113190bb543c6225c07e | uncurl/bin.py | uncurl/bin.py | from __future__ import print_function
import sys
from .api import parse
def main():
result = parse(sys.argv[1])
print(result)
| from __future__ import print_function
import sys
from .api import parse
def main():
if sys.stdin.isatty():
result = parse(sys.argv[1])
else:
result = parse(sys.stdin.read())
print(result)
| Read from stdin if available. | Read from stdin if available.
| Python | apache-2.0 | weinerjm/uncurl,spulec/uncurl | from __future__ import print_function
import sys
from .api import parse
def main():
+ if sys.stdin.isatty():
- result = parse(sys.argv[1])
+ result = parse(sys.argv[1])
+ else:
+ result = parse(sys.stdin.read())
print(result)
| Read from stdin if available. | ## Code Before:
from __future__ import print_function
import sys
from .api import parse
def main():
result = parse(sys.argv[1])
print(result)
## Instruction:
Read from stdin if available.
## Code After:
from __future__ import print_function
import sys
from .api import parse
def main():
if sys.stdin.isatty():
result = parse(sys.argv[1])
else:
result = parse(sys.stdin.read())
print(result)
| // ... existing code ...
def main():
if sys.stdin.isatty():
result = parse(sys.argv[1])
else:
result = parse(sys.stdin.read())
print(result)
// ... rest of the code ... |
6807e5a5966f1f37f69a54e255a9981918cc8fb6 | tests/test_cmd.py | tests/test_cmd.py | import base64
import os
from distutils.core import Command
class TestCommand(Command):
description = "Launch all tests under fusion_tables app"
user_options = []
def initialize_options(self):
pass
def finalize_options(self):
pass
def create_client_secret_file(self):
client_secret = open("/tmp/client_secret.json", "w")
data = os.environ.get("CLIENT_SECRET")
client_secret.write(base64.b64decode(data))
client_secret.close()
def configure_settings(self):
from django.conf import settings
settings.configure(
DATABASES={
"default": {
"NAME": ":memory:",
"ENGINE": "django.db.backends.sqlite3",
"TEST": {
"NAME": ":memory:"
}
}
},
INSTALLED_APPS=(
"django.contrib.contenttypes",
"fusion_tables",
),
ROOT_URLCONF="tests.urls",
MODELS_TO_SYNC=("fusion_tables.SampleModel", ),
CLIENT_SECRET_JSON_FILEPATH="/tmp/client_secret.json",
LOCATION_FIELDS=("TextField", )
)
def run(self):
import django
from django.core.management import call_command
self.create_client_secret_file()
self.configure_settings()
django.setup()
call_command("test", "fusion_tables")
| import base64
import os
from distutils.core import Command
class TestCommand(Command):
description = "Launch all tests under fusion_tables app"
user_options = []
def initialize_options(self):
pass
def finalize_options(self):
pass
def create_client_secret_file(self):
client_secret = open("/tmp/client_secret.json", "w")
data = os.environ.get("CLIENT_SECRET").decode("utf-8")
client_secret.write(base64.b64decode(data))
client_secret.close()
def configure_settings(self):
from django.conf import settings
settings.configure(
DATABASES={
"default": {
"NAME": ":memory:",
"ENGINE": "django.db.backends.sqlite3",
"TEST": {
"NAME": ":memory:"
}
}
},
INSTALLED_APPS=(
"django.contrib.contenttypes",
"fusion_tables",
),
ROOT_URLCONF="tests.urls",
MODELS_TO_SYNC=("fusion_tables.SampleModel", ),
CLIENT_SECRET_JSON_FILEPATH="/tmp/client_secret.json",
LOCATION_FIELDS=("TextField", )
)
def run(self):
import django
from django.core.management import call_command
self.create_client_secret_file()
self.configure_settings()
django.setup()
call_command("test", "fusion_tables")
| Fix unit test python3 compatibility. | Fix unit test python3 compatibility.
| Python | mit | bsvetchine/django-fusion-tables | import base64
import os
from distutils.core import Command
class TestCommand(Command):
description = "Launch all tests under fusion_tables app"
user_options = []
def initialize_options(self):
pass
def finalize_options(self):
pass
def create_client_secret_file(self):
client_secret = open("/tmp/client_secret.json", "w")
- data = os.environ.get("CLIENT_SECRET")
+ data = os.environ.get("CLIENT_SECRET").decode("utf-8")
client_secret.write(base64.b64decode(data))
client_secret.close()
def configure_settings(self):
from django.conf import settings
settings.configure(
DATABASES={
"default": {
"NAME": ":memory:",
"ENGINE": "django.db.backends.sqlite3",
"TEST": {
"NAME": ":memory:"
}
}
},
INSTALLED_APPS=(
"django.contrib.contenttypes",
"fusion_tables",
),
ROOT_URLCONF="tests.urls",
MODELS_TO_SYNC=("fusion_tables.SampleModel", ),
CLIENT_SECRET_JSON_FILEPATH="/tmp/client_secret.json",
LOCATION_FIELDS=("TextField", )
)
def run(self):
import django
from django.core.management import call_command
self.create_client_secret_file()
self.configure_settings()
django.setup()
call_command("test", "fusion_tables")
| Fix unit test python3 compatibility. | ## Code Before:
import base64
import os
from distutils.core import Command
class TestCommand(Command):
description = "Launch all tests under fusion_tables app"
user_options = []
def initialize_options(self):
pass
def finalize_options(self):
pass
def create_client_secret_file(self):
client_secret = open("/tmp/client_secret.json", "w")
data = os.environ.get("CLIENT_SECRET")
client_secret.write(base64.b64decode(data))
client_secret.close()
def configure_settings(self):
from django.conf import settings
settings.configure(
DATABASES={
"default": {
"NAME": ":memory:",
"ENGINE": "django.db.backends.sqlite3",
"TEST": {
"NAME": ":memory:"
}
}
},
INSTALLED_APPS=(
"django.contrib.contenttypes",
"fusion_tables",
),
ROOT_URLCONF="tests.urls",
MODELS_TO_SYNC=("fusion_tables.SampleModel", ),
CLIENT_SECRET_JSON_FILEPATH="/tmp/client_secret.json",
LOCATION_FIELDS=("TextField", )
)
def run(self):
import django
from django.core.management import call_command
self.create_client_secret_file()
self.configure_settings()
django.setup()
call_command("test", "fusion_tables")
## Instruction:
Fix unit test python3 compatibility.
## Code After:
import base64
import os
from distutils.core import Command
class TestCommand(Command):
description = "Launch all tests under fusion_tables app"
user_options = []
def initialize_options(self):
pass
def finalize_options(self):
pass
def create_client_secret_file(self):
client_secret = open("/tmp/client_secret.json", "w")
data = os.environ.get("CLIENT_SECRET").decode("utf-8")
client_secret.write(base64.b64decode(data))
client_secret.close()
def configure_settings(self):
from django.conf import settings
settings.configure(
DATABASES={
"default": {
"NAME": ":memory:",
"ENGINE": "django.db.backends.sqlite3",
"TEST": {
"NAME": ":memory:"
}
}
},
INSTALLED_APPS=(
"django.contrib.contenttypes",
"fusion_tables",
),
ROOT_URLCONF="tests.urls",
MODELS_TO_SYNC=("fusion_tables.SampleModel", ),
CLIENT_SECRET_JSON_FILEPATH="/tmp/client_secret.json",
LOCATION_FIELDS=("TextField", )
)
def run(self):
import django
from django.core.management import call_command
self.create_client_secret_file()
self.configure_settings()
django.setup()
call_command("test", "fusion_tables")
| ...
def create_client_secret_file(self):
client_secret = open("/tmp/client_secret.json", "w")
data = os.environ.get("CLIENT_SECRET").decode("utf-8")
client_secret.write(base64.b64decode(data))
client_secret.close()
... |
d89747e26371b1986b4cec5a7514ba2c99480487 | tests/test_codec.py | tests/test_codec.py | from .common import *
from av.codec import Codec, Encoder, Decoder
class TestCodecs(TestCase):
def test_codec_mpeg4(self):
for cls in (Encoder, Decoder):
c = cls('mpeg4')
self.assertEqual(c.name, 'mpeg4')
self.assertEqual(c.long_name, 'MPEG-4 part 2')
self.assertEqual(c.type, 'video')
self.assertEqual(c.id, 13)
formats = c.video_formats
self.assertEqual(len(formats), 1)
self.assertEqual(formats[0].name, 'yuv420p')
| from .common import *
from av.codec import Codec, Encoder, Decoder
class TestCodecs(TestCase):
def test_codec_mpeg4(self):
for cls in (Encoder, Decoder):
c = cls('mpeg4')
self.assertEqual(c.name, 'mpeg4')
self.assertEqual(c.long_name, 'MPEG-4 part 2')
self.assertEqual(c.type, 'video')
self.assertEqual(c.id, 13)
formats = c.video_formats
self.assertTrue(formats)
self.assertTrue(any(f.name == 'yuv420p' for f in formats))
| Allow codec test to have more than just the one format | Allow codec test to have more than just the one format
| Python | bsd-3-clause | mcpv/PyAV,danielballan/PyAV,pupil-labs/PyAV,PyAV-Org/PyAV,markreidvfx/PyAV,PyAV-Org/PyAV,xxr3376/PyAV,pupil-labs/PyAV,pupil-labs/PyAV,xxr3376/PyAV,mikeboers/PyAV,markreidvfx/PyAV,danielballan/PyAV,danielballan/PyAV,xxr3376/PyAV,markreidvfx/PyAV,mcpv/PyAV,mikeboers/PyAV,mcpv/PyAV,pupil-labs/PyAV | from .common import *
from av.codec import Codec, Encoder, Decoder
class TestCodecs(TestCase):
def test_codec_mpeg4(self):
for cls in (Encoder, Decoder):
c = cls('mpeg4')
self.assertEqual(c.name, 'mpeg4')
self.assertEqual(c.long_name, 'MPEG-4 part 2')
self.assertEqual(c.type, 'video')
self.assertEqual(c.id, 13)
formats = c.video_formats
- self.assertEqual(len(formats), 1)
+ self.assertTrue(formats)
- self.assertEqual(formats[0].name, 'yuv420p')
+ self.assertTrue(any(f.name == 'yuv420p' for f in formats))
| Allow codec test to have more than just the one format | ## Code Before:
from .common import *
from av.codec import Codec, Encoder, Decoder
class TestCodecs(TestCase):
def test_codec_mpeg4(self):
for cls in (Encoder, Decoder):
c = cls('mpeg4')
self.assertEqual(c.name, 'mpeg4')
self.assertEqual(c.long_name, 'MPEG-4 part 2')
self.assertEqual(c.type, 'video')
self.assertEqual(c.id, 13)
formats = c.video_formats
self.assertEqual(len(formats), 1)
self.assertEqual(formats[0].name, 'yuv420p')
## Instruction:
Allow codec test to have more than just the one format
## Code After:
from .common import *
from av.codec import Codec, Encoder, Decoder
class TestCodecs(TestCase):
def test_codec_mpeg4(self):
for cls in (Encoder, Decoder):
c = cls('mpeg4')
self.assertEqual(c.name, 'mpeg4')
self.assertEqual(c.long_name, 'MPEG-4 part 2')
self.assertEqual(c.type, 'video')
self.assertEqual(c.id, 13)
formats = c.video_formats
self.assertTrue(formats)
self.assertTrue(any(f.name == 'yuv420p' for f in formats))
| // ... existing code ...
formats = c.video_formats
self.assertTrue(formats)
self.assertTrue(any(f.name == 'yuv420p' for f in formats))
// ... rest of the code ... |
0c2a7bfebbb6d427ffea66f4a8df534c5b8be974 | timed/subscription/admin.py | timed/subscription/admin.py | from django.contrib import admin
from . import models
@admin.register(models.Package)
class PackageAdmin(admin.ModelAdmin):
list_display = ['billing_type', 'duration', 'price']
| from django import forms
from django.contrib import admin
from django.utils.translation import ugettext_lazy as _
from timed.forms import DurationInHoursField
from . import models
class PackageForm(forms.ModelForm):
model = models.Package
duration = DurationInHoursField(
label=_('Duration in hours'),
required=True,
)
@admin.register(models.Package)
class PackageAdmin(admin.ModelAdmin):
list_display = ['billing_type', 'duration', 'price']
form = PackageForm
| Configure duration field on subscription package in hours | Configure duration field on subscription package in hours
| Python | agpl-3.0 | adfinis-sygroup/timed-backend,adfinis-sygroup/timed-backend,adfinis-sygroup/timed-backend | + from django import forms
from django.contrib import admin
+ from django.utils.translation import ugettext_lazy as _
+
+ from timed.forms import DurationInHoursField
from . import models
+
+
+ class PackageForm(forms.ModelForm):
+ model = models.Package
+ duration = DurationInHoursField(
+ label=_('Duration in hours'),
+ required=True,
+ )
@admin.register(models.Package)
class PackageAdmin(admin.ModelAdmin):
list_display = ['billing_type', 'duration', 'price']
+ form = PackageForm
| Configure duration field on subscription package in hours | ## Code Before:
from django.contrib import admin
from . import models
@admin.register(models.Package)
class PackageAdmin(admin.ModelAdmin):
list_display = ['billing_type', 'duration', 'price']
## Instruction:
Configure duration field on subscription package in hours
## Code After:
from django import forms
from django.contrib import admin
from django.utils.translation import ugettext_lazy as _
from timed.forms import DurationInHoursField
from . import models
class PackageForm(forms.ModelForm):
model = models.Package
duration = DurationInHoursField(
label=_('Duration in hours'),
required=True,
)
@admin.register(models.Package)
class PackageAdmin(admin.ModelAdmin):
list_display = ['billing_type', 'duration', 'price']
form = PackageForm
| // ... existing code ...
from django import forms
from django.contrib import admin
from django.utils.translation import ugettext_lazy as _
from timed.forms import DurationInHoursField
from . import models
class PackageForm(forms.ModelForm):
model = models.Package
duration = DurationInHoursField(
label=_('Duration in hours'),
required=True,
)
// ... modified code ...
class PackageAdmin(admin.ModelAdmin):
list_display = ['billing_type', 'duration', 'price']
form = PackageForm
// ... rest of the code ... |
d187a8434c9d64171f76efa3055bdc06afbc8981 | scripts/pystart.py | scripts/pystart.py | import os,sys,re
from time import sleep
from pprint import pprint
home = os.path.expanduser('~')
from math import log,ceil
def clog2(num):
return int(ceil(log(num,2)))
if (sys.version_info > (3, 0)):
# Python 3 code in this block
exec(open(home+'/homedir/scripts/hexecho.py').read())
else:
# Python 2 code in this block
execfile(home+'/homedir/scripts/hexecho.py')
hexoff
print ("Imported os,sys,re,sleep,pprint. Defined clog2,hexon/hexoff")
| import os,sys,re
from time import sleep
from pprint import pprint
home = os.path.expanduser('~')
from math import log,ceil
sys.ps1 = '\001\033[96m\002>>> \001\033[0m\002'
sys.ps2 = '\001\033[96m\002... \001\033[0m\002'
def clog2(num):
return int(ceil(log(num,2)))
if (sys.version_info > (3, 0)):
# Python 3 code in this block
exec(open(home+'/homedir/scripts/hexecho.py').read())
else:
# Python 2 code in this block
execfile(home+'/homedir/scripts/hexecho.py')
hexoff
print ("Imported os,sys,re,sleep,pprint. Defined clog2,hexon/hexoff")
| Add color to python prompt | Add color to python prompt
| Python | mit | jdanders/homedir,jdanders/homedir,jdanders/homedir,jdanders/homedir | import os,sys,re
from time import sleep
from pprint import pprint
home = os.path.expanduser('~')
from math import log,ceil
+ sys.ps1 = '\001\033[96m\002>>> \001\033[0m\002'
+ sys.ps2 = '\001\033[96m\002... \001\033[0m\002'
def clog2(num):
return int(ceil(log(num,2)))
if (sys.version_info > (3, 0)):
# Python 3 code in this block
exec(open(home+'/homedir/scripts/hexecho.py').read())
else:
# Python 2 code in this block
execfile(home+'/homedir/scripts/hexecho.py')
hexoff
print ("Imported os,sys,re,sleep,pprint. Defined clog2,hexon/hexoff")
| Add color to python prompt | ## Code Before:
import os,sys,re
from time import sleep
from pprint import pprint
home = os.path.expanduser('~')
from math import log,ceil
def clog2(num):
return int(ceil(log(num,2)))
if (sys.version_info > (3, 0)):
# Python 3 code in this block
exec(open(home+'/homedir/scripts/hexecho.py').read())
else:
# Python 2 code in this block
execfile(home+'/homedir/scripts/hexecho.py')
hexoff
print ("Imported os,sys,re,sleep,pprint. Defined clog2,hexon/hexoff")
## Instruction:
Add color to python prompt
## Code After:
import os,sys,re
from time import sleep
from pprint import pprint
home = os.path.expanduser('~')
from math import log,ceil
sys.ps1 = '\001\033[96m\002>>> \001\033[0m\002'
sys.ps2 = '\001\033[96m\002... \001\033[0m\002'
def clog2(num):
return int(ceil(log(num,2)))
if (sys.version_info > (3, 0)):
# Python 3 code in this block
exec(open(home+'/homedir/scripts/hexecho.py').read())
else:
# Python 2 code in this block
execfile(home+'/homedir/scripts/hexecho.py')
hexoff
print ("Imported os,sys,re,sleep,pprint. Defined clog2,hexon/hexoff")
| ...
home = os.path.expanduser('~')
from math import log,ceil
sys.ps1 = '\001\033[96m\002>>> \001\033[0m\002'
sys.ps2 = '\001\033[96m\002... \001\033[0m\002'
def clog2(num):
return int(ceil(log(num,2)))
... |
2f140327c24a8efab5482a975793dddedd0ebfc4 | nucleus/wsgi.py | nucleus/wsgi.py | # newrelic.agent must be imported and initialized first
# https://docs.newrelic.com/docs/agents/python-agent/installation/python-agent-advanced-integration#manual-integration
import newrelic.agent
newrelic.agent.initialize('newrelic.ini')
import os
os.environ.setdefault('DJANGO_SETTINGS_MODULE', 'nucleus.settings') # NOQA
from django.core.handlers.wsgi import WSGIRequest
from django.core.wsgi import get_wsgi_application
from decouple import config
IS_HTTPS = config('HTTPS', default='off', cast=bool)
class WSGIHTTPSRequest(WSGIRequest):
def _get_scheme(self):
if IS_HTTPS:
return 'https'
return super(WSGIHTTPSRequest, self)._get_scheme()
application = get_wsgi_application()
application.request_class = WSGIHTTPSRequest
if config('SENTRY_DSN', None):
from raven.contrib.django.raven_compat.middleware.wsgi import Sentry
application = Sentry(application)
newrelic_license_key = config('NEW_RELIC_LICENSE_KEY', default=None)
if newrelic_license_key:
application = newrelic.agent.WSGIApplicationWrapper(application)
| import newrelic.agent
newrelic.agent.initialize('newrelic.ini')
import os
os.environ.setdefault('DJANGO_SETTINGS_MODULE', 'nucleus.settings') # NOQA
from django.core.handlers.wsgi import WSGIRequest
from django.core.wsgi import get_wsgi_application
from decouple import config
IS_HTTPS = config('HTTPS', default='off', cast=bool)
class WSGIHTTPSRequest(WSGIRequest):
def _get_scheme(self):
if IS_HTTPS:
return 'https'
return super(WSGIHTTPSRequest, self)._get_scheme()
application = get_wsgi_application()
application.request_class = WSGIHTTPSRequest
if config('SENTRY_DSN', None):
from raven.contrib.django.raven_compat.middleware.wsgi import Sentry
application = Sentry(application)
newrelic_license_key = config('NEW_RELIC_LICENSE_KEY', default=None)
if newrelic_license_key:
application = newrelic.agent.WSGIApplicationWrapper(application)
| Remove old docstring with link to old django docs | Remove old docstring with link to old django docs
| Python | mpl-2.0 | mozilla/nucleus,mozilla/nucleus,mozilla/nucleus,mozilla/nucleus | - # newrelic.agent must be imported and initialized first
- # https://docs.newrelic.com/docs/agents/python-agent/installation/python-agent-advanced-integration#manual-integration
import newrelic.agent
newrelic.agent.initialize('newrelic.ini')
import os
os.environ.setdefault('DJANGO_SETTINGS_MODULE', 'nucleus.settings') # NOQA
from django.core.handlers.wsgi import WSGIRequest
from django.core.wsgi import get_wsgi_application
from decouple import config
IS_HTTPS = config('HTTPS', default='off', cast=bool)
class WSGIHTTPSRequest(WSGIRequest):
def _get_scheme(self):
if IS_HTTPS:
return 'https'
return super(WSGIHTTPSRequest, self)._get_scheme()
application = get_wsgi_application()
application.request_class = WSGIHTTPSRequest
if config('SENTRY_DSN', None):
from raven.contrib.django.raven_compat.middleware.wsgi import Sentry
application = Sentry(application)
newrelic_license_key = config('NEW_RELIC_LICENSE_KEY', default=None)
if newrelic_license_key:
application = newrelic.agent.WSGIApplicationWrapper(application)
| Remove old docstring with link to old django docs | ## Code Before:
# newrelic.agent must be imported and initialized first
# https://docs.newrelic.com/docs/agents/python-agent/installation/python-agent-advanced-integration#manual-integration
import newrelic.agent
newrelic.agent.initialize('newrelic.ini')
import os
os.environ.setdefault('DJANGO_SETTINGS_MODULE', 'nucleus.settings') # NOQA
from django.core.handlers.wsgi import WSGIRequest
from django.core.wsgi import get_wsgi_application
from decouple import config
IS_HTTPS = config('HTTPS', default='off', cast=bool)
class WSGIHTTPSRequest(WSGIRequest):
def _get_scheme(self):
if IS_HTTPS:
return 'https'
return super(WSGIHTTPSRequest, self)._get_scheme()
application = get_wsgi_application()
application.request_class = WSGIHTTPSRequest
if config('SENTRY_DSN', None):
from raven.contrib.django.raven_compat.middleware.wsgi import Sentry
application = Sentry(application)
newrelic_license_key = config('NEW_RELIC_LICENSE_KEY', default=None)
if newrelic_license_key:
application = newrelic.agent.WSGIApplicationWrapper(application)
## Instruction:
Remove old docstring with link to old django docs
## Code After:
import newrelic.agent
newrelic.agent.initialize('newrelic.ini')
import os
os.environ.setdefault('DJANGO_SETTINGS_MODULE', 'nucleus.settings') # NOQA
from django.core.handlers.wsgi import WSGIRequest
from django.core.wsgi import get_wsgi_application
from decouple import config
IS_HTTPS = config('HTTPS', default='off', cast=bool)
class WSGIHTTPSRequest(WSGIRequest):
def _get_scheme(self):
if IS_HTTPS:
return 'https'
return super(WSGIHTTPSRequest, self)._get_scheme()
application = get_wsgi_application()
application.request_class = WSGIHTTPSRequest
if config('SENTRY_DSN', None):
from raven.contrib.django.raven_compat.middleware.wsgi import Sentry
application = Sentry(application)
newrelic_license_key = config('NEW_RELIC_LICENSE_KEY', default=None)
if newrelic_license_key:
application = newrelic.agent.WSGIApplicationWrapper(application)
| ...
import newrelic.agent
newrelic.agent.initialize('newrelic.ini')
... |
df227a375c1cf5fdd0ad23505799e7c6f7177b9c | InvenTree/InvenTree/validators.py | InvenTree/InvenTree/validators.py |
from django.core.exceptions import ValidationError
from django.utils.translation import gettext_lazy as _
def validate_part_name(value):
# Prevent some illegal characters in part names
for c in ['/', '\\', '|', '#', '$']:
if c in str(value):
raise ValidationError(
_('Invalid character in part name')
)
|
from django.core.exceptions import ValidationError
from django.utils.translation import gettext_lazy as _
def validate_part_name(value):
# Prevent some illegal characters in part names
for c in ['|', '#', '$']:
if c in str(value):
raise ValidationError(
_('Invalid character in part name')
)
| Allow some more chars in part names | Allow some more chars in part names
| Python | mit | inventree/InvenTree,inventree/InvenTree,SchrodingersGat/InvenTree,inventree/InvenTree,SchrodingersGat/InvenTree,inventree/InvenTree,SchrodingersGat/InvenTree,SchrodingersGat/InvenTree |
from django.core.exceptions import ValidationError
from django.utils.translation import gettext_lazy as _
def validate_part_name(value):
# Prevent some illegal characters in part names
- for c in ['/', '\\', '|', '#', '$']:
+ for c in ['|', '#', '$']:
if c in str(value):
raise ValidationError(
_('Invalid character in part name')
)
| Allow some more chars in part names | ## Code Before:
from django.core.exceptions import ValidationError
from django.utils.translation import gettext_lazy as _
def validate_part_name(value):
# Prevent some illegal characters in part names
for c in ['/', '\\', '|', '#', '$']:
if c in str(value):
raise ValidationError(
_('Invalid character in part name')
)
## Instruction:
Allow some more chars in part names
## Code After:
from django.core.exceptions import ValidationError
from django.utils.translation import gettext_lazy as _
def validate_part_name(value):
# Prevent some illegal characters in part names
for c in ['|', '#', '$']:
if c in str(value):
raise ValidationError(
_('Invalid character in part name')
)
| // ... existing code ...
def validate_part_name(value):
# Prevent some illegal characters in part names
for c in ['|', '#', '$']:
if c in str(value):
raise ValidationError(
// ... rest of the code ... |
7d8283b2d233a8fbee97de122f0b4ba293cf788d | app/emails.py | app/emails.py |
from flask import render_template,g
from flask.ext.mail import Message
from app import mail, db
from .models import User
from config import MAIL_SENDER
# Wrapper function for sending mails using flask-mail plugin
def send_email(subject, sender, recipients, text_body):
msg = Message(subject, sender=sender, recipients=recipients)
msg.body = text_body
mail.send(msg)
def add_movie_notification(movie):
users = User.query.all()
you_user = False
for cur_user in users:
# Check if the cur_user is the logged user who added the movie
# in order to change the mail text
if cur_user.id==g.user.id:
you_user=True
send_email('[Cineapp] - Ajout d\'un film' , MAIL_SENDER,[ cur_user.email ] ,
render_template('add_movie_notification.txt', dest_user=cur_user, add_user=g.user,movie=movie,you_user=you_user))
|
from flask import render_template,g
from flask.ext.mail import Message
from app import mail, db
from .models import User
from config import MAIL_SENDER
from threading import Thread
from app import app
# Send mail into a dedicated thread in order to avoir the web app to wait
def send_async_email(app, msg):
with app.app_context():
mail.send(msg)
# Wrapper function for sending mails using flask-mail plugin
def send_email(subject, sender, recipients, text_body):
msg = Message(subject, sender=sender, recipients=recipients)
msg.body = text_body
thr = Thread(target=send_async_email, args=[app, msg])
thr.start()
# Function which sends notifications to users when a movie is added
def add_movie_notification(movie):
users = User.query.all()
you_user = False
for cur_user in users:
# Check if the cur_user is the logged user who added the movie
# in order to change the mail text
if cur_user.id==g.user.id:
you_user=True
send_email('[Cineapp] - Ajout d\'un film' , MAIL_SENDER,[ cur_user.email ] ,
render_template('add_movie_notification.txt', dest_user=cur_user, add_user=g.user,movie=movie,you_user=you_user))
| Send email notifications in asynchronous mode | Send email notifications in asynchronous mode
Each mail notification is done on a thread in order to not block the main thread
of the web app.
| Python | mit | ptitoliv/cineapp,ptitoliv/cineapp,ptitoliv/cineapp |
from flask import render_template,g
from flask.ext.mail import Message
from app import mail, db
from .models import User
from config import MAIL_SENDER
+ from threading import Thread
+ from app import app
+
+ # Send mail into a dedicated thread in order to avoir the web app to wait
+ def send_async_email(app, msg):
+ with app.app_context():
+ mail.send(msg)
# Wrapper function for sending mails using flask-mail plugin
def send_email(subject, sender, recipients, text_body):
msg = Message(subject, sender=sender, recipients=recipients)
msg.body = text_body
- mail.send(msg)
+ thr = Thread(target=send_async_email, args=[app, msg])
+ thr.start()
+ # Function which sends notifications to users when a movie is added
def add_movie_notification(movie):
users = User.query.all()
you_user = False
for cur_user in users:
# Check if the cur_user is the logged user who added the movie
# in order to change the mail text
if cur_user.id==g.user.id:
you_user=True
send_email('[Cineapp] - Ajout d\'un film' , MAIL_SENDER,[ cur_user.email ] ,
render_template('add_movie_notification.txt', dest_user=cur_user, add_user=g.user,movie=movie,you_user=you_user))
| Send email notifications in asynchronous mode | ## Code Before:
from flask import render_template,g
from flask.ext.mail import Message
from app import mail, db
from .models import User
from config import MAIL_SENDER
# Wrapper function for sending mails using flask-mail plugin
def send_email(subject, sender, recipients, text_body):
msg = Message(subject, sender=sender, recipients=recipients)
msg.body = text_body
mail.send(msg)
def add_movie_notification(movie):
users = User.query.all()
you_user = False
for cur_user in users:
# Check if the cur_user is the logged user who added the movie
# in order to change the mail text
if cur_user.id==g.user.id:
you_user=True
send_email('[Cineapp] - Ajout d\'un film' , MAIL_SENDER,[ cur_user.email ] ,
render_template('add_movie_notification.txt', dest_user=cur_user, add_user=g.user,movie=movie,you_user=you_user))
## Instruction:
Send email notifications in asynchronous mode
## Code After:
from flask import render_template,g
from flask.ext.mail import Message
from app import mail, db
from .models import User
from config import MAIL_SENDER
from threading import Thread
from app import app
# Send mail into a dedicated thread in order to avoir the web app to wait
def send_async_email(app, msg):
with app.app_context():
mail.send(msg)
# Wrapper function for sending mails using flask-mail plugin
def send_email(subject, sender, recipients, text_body):
msg = Message(subject, sender=sender, recipients=recipients)
msg.body = text_body
thr = Thread(target=send_async_email, args=[app, msg])
thr.start()
# Function which sends notifications to users when a movie is added
def add_movie_notification(movie):
users = User.query.all()
you_user = False
for cur_user in users:
# Check if the cur_user is the logged user who added the movie
# in order to change the mail text
if cur_user.id==g.user.id:
you_user=True
send_email('[Cineapp] - Ajout d\'un film' , MAIL_SENDER,[ cur_user.email ] ,
render_template('add_movie_notification.txt', dest_user=cur_user, add_user=g.user,movie=movie,you_user=you_user))
| // ... existing code ...
from .models import User
from config import MAIL_SENDER
from threading import Thread
from app import app
# Send mail into a dedicated thread in order to avoir the web app to wait
def send_async_email(app, msg):
with app.app_context():
mail.send(msg)
# Wrapper function for sending mails using flask-mail plugin
// ... modified code ...
msg = Message(subject, sender=sender, recipients=recipients)
msg.body = text_body
thr = Thread(target=send_async_email, args=[app, msg])
thr.start()
# Function which sends notifications to users when a movie is added
def add_movie_notification(movie):
users = User.query.all()
// ... rest of the code ... |
be964b02036159567efcaecce5b5d905f23985af | deduper/scanfiles.py | deduper/scanfiles.py | import os
from sqlalchemy.ext.declarative import declarative_base
Base = declarative_base()
from .models import ImageFile
from .util import Util
def ScanFiles(session, FOLDER):
for root, dirs, files in os.walk(FOLDER):
if root.split('/')[-1].startswith('.'):
continue
for count, filename in enumerate(files, start=1):
if filename.startswith('.'):
continue
fullpath = os.path.join(root, filename)
if Util.file_record_exists(session, fullpath):
print('{count} of {length}: Skipping {filename}'.format(
count=count, length=len(files), filename=filename))
else:
print('{count} of {length}: Processing {filename}'.format(
count=count, length=len(files), filename=filename))
new_file = ImageFile(name=filename, fullpath=fullpath,
filehash=Util.hash_file(fullpath))
session.add(new_file)
session.commit()
session.close()
| import os
from sqlalchemy.ext.declarative import declarative_base
Base = declarative_base()
from .models import ImageFile
from .util import Util
def ScanFiles(session, FOLDER):
for root, dirs, files in os.walk(FOLDER):
if root.split('/')[-1].startswith('.'):
continue
for count, filename in enumerate(files, start=1):
if filename.startswith('.'):
continue
fullpath = os.path.join(root, filename)
if not os.path.isfile(fullpath):
continue
if Util.file_record_exists(session, fullpath):
print('{count} of {length}: Skipping {filename}'.format(
count=count, length=len(files), filename=filename))
else:
print('{count} of {length}: Processing {filename}'.format(
count=count, length=len(files), filename=filename))
new_file = ImageFile(name=filename, fullpath=fullpath,
filehash=Util.hash_file(fullpath))
session.add(new_file)
session.commit()
session.close()
| Check that fullpath is a regular file before continuing | Check that fullpath is a regular file before continuing
| Python | bsd-3-clause | cgspeck/filededuper | import os
from sqlalchemy.ext.declarative import declarative_base
Base = declarative_base()
from .models import ImageFile
from .util import Util
def ScanFiles(session, FOLDER):
for root, dirs, files in os.walk(FOLDER):
if root.split('/')[-1].startswith('.'):
continue
for count, filename in enumerate(files, start=1):
if filename.startswith('.'):
continue
fullpath = os.path.join(root, filename)
+ if not os.path.isfile(fullpath):
+ continue
+
if Util.file_record_exists(session, fullpath):
print('{count} of {length}: Skipping {filename}'.format(
count=count, length=len(files), filename=filename))
else:
print('{count} of {length}: Processing {filename}'.format(
count=count, length=len(files), filename=filename))
new_file = ImageFile(name=filename, fullpath=fullpath,
filehash=Util.hash_file(fullpath))
session.add(new_file)
session.commit()
session.close()
| Check that fullpath is a regular file before continuing | ## Code Before:
import os
from sqlalchemy.ext.declarative import declarative_base
Base = declarative_base()
from .models import ImageFile
from .util import Util
def ScanFiles(session, FOLDER):
for root, dirs, files in os.walk(FOLDER):
if root.split('/')[-1].startswith('.'):
continue
for count, filename in enumerate(files, start=1):
if filename.startswith('.'):
continue
fullpath = os.path.join(root, filename)
if Util.file_record_exists(session, fullpath):
print('{count} of {length}: Skipping {filename}'.format(
count=count, length=len(files), filename=filename))
else:
print('{count} of {length}: Processing {filename}'.format(
count=count, length=len(files), filename=filename))
new_file = ImageFile(name=filename, fullpath=fullpath,
filehash=Util.hash_file(fullpath))
session.add(new_file)
session.commit()
session.close()
## Instruction:
Check that fullpath is a regular file before continuing
## Code After:
import os
from sqlalchemy.ext.declarative import declarative_base
Base = declarative_base()
from .models import ImageFile
from .util import Util
def ScanFiles(session, FOLDER):
for root, dirs, files in os.walk(FOLDER):
if root.split('/')[-1].startswith('.'):
continue
for count, filename in enumerate(files, start=1):
if filename.startswith('.'):
continue
fullpath = os.path.join(root, filename)
if not os.path.isfile(fullpath):
continue
if Util.file_record_exists(session, fullpath):
print('{count} of {length}: Skipping {filename}'.format(
count=count, length=len(files), filename=filename))
else:
print('{count} of {length}: Processing {filename}'.format(
count=count, length=len(files), filename=filename))
new_file = ImageFile(name=filename, fullpath=fullpath,
filehash=Util.hash_file(fullpath))
session.add(new_file)
session.commit()
session.close()
| // ... existing code ...
fullpath = os.path.join(root, filename)
if not os.path.isfile(fullpath):
continue
if Util.file_record_exists(session, fullpath):
print('{count} of {length}: Skipping {filename}'.format(
// ... rest of the code ... |
c5239c6bbb40ede4279b33b965c5ded26a78b2ae | app/tests/manual/test_twitter_api.py | app/tests/manual/test_twitter_api.py | from __future__ import absolute_import
from unittest import TestCase
from lib.twitter_api import authentication
class TestAuth(TestCase):
def test_generateAppAccessToken(self):
auth = authentication._generateAppAccessToken()
def test_getTweepyConnection(self):
auth = authentication._generateAppAccessToken()
api = authentication._getTweepyConnection(auth)
def test_getAPIConnection(self):
"""
Test that App Access token can be used to connect to Twitter API.
"""
api = authentication.getAPIConnection(userFlow=False)
def test_getAppOnlyConnection(self):
"""
Test App-only token.
"""
api = authentication.getAppOnlyConnection()
| from __future__ import absolute_import
import os
import sys
import unittest
from unittest import TestCase
# Allow imports to be done when executing this file directly.
sys.path.insert(0, os.path.abspath(os.path.join(
os.path.dirname(__file__), os.path.pardir, os.path.pardir)
))
from lib.twitter_api import authentication
class TestAuth(TestCase):
def test_generateAppAccessToken(self):
auth = authentication._generateAppAccessToken()
def test_getTweepyConnection(self):
auth = authentication._generateAppAccessToken()
api = authentication._getTweepyConnection(auth)
def test_getAPIConnection(self):
"""
Test that App Access token can be used to connect to Twitter API.
"""
api = authentication.getAPIConnection(userFlow=False)
def test_getAppOnlyConnection(self):
"""
Test App-only token.
"""
api = authentication.getAppOnlyConnection()
if __name__ == '__main__':
unittest.main()
| Update Twitter auth test to run directly | test: Update Twitter auth test to run directly
| Python | mit | MichaelCurrin/twitterverse,MichaelCurrin/twitterverse | from __future__ import absolute_import
+ import os
+ import sys
+ import unittest
from unittest import TestCase
+
+ # Allow imports to be done when executing this file directly.
+ sys.path.insert(0, os.path.abspath(os.path.join(
+ os.path.dirname(__file__), os.path.pardir, os.path.pardir)
+ ))
+
from lib.twitter_api import authentication
class TestAuth(TestCase):
def test_generateAppAccessToken(self):
auth = authentication._generateAppAccessToken()
def test_getTweepyConnection(self):
auth = authentication._generateAppAccessToken()
api = authentication._getTweepyConnection(auth)
def test_getAPIConnection(self):
"""
Test that App Access token can be used to connect to Twitter API.
"""
api = authentication.getAPIConnection(userFlow=False)
def test_getAppOnlyConnection(self):
"""
Test App-only token.
"""
api = authentication.getAppOnlyConnection()
+
+ if __name__ == '__main__':
+ unittest.main()
+ | Update Twitter auth test to run directly | ## Code Before:
from __future__ import absolute_import
from unittest import TestCase
from lib.twitter_api import authentication
class TestAuth(TestCase):
def test_generateAppAccessToken(self):
auth = authentication._generateAppAccessToken()
def test_getTweepyConnection(self):
auth = authentication._generateAppAccessToken()
api = authentication._getTweepyConnection(auth)
def test_getAPIConnection(self):
"""
Test that App Access token can be used to connect to Twitter API.
"""
api = authentication.getAPIConnection(userFlow=False)
def test_getAppOnlyConnection(self):
"""
Test App-only token.
"""
api = authentication.getAppOnlyConnection()
## Instruction:
Update Twitter auth test to run directly
## Code After:
from __future__ import absolute_import
import os
import sys
import unittest
from unittest import TestCase
# Allow imports to be done when executing this file directly.
sys.path.insert(0, os.path.abspath(os.path.join(
os.path.dirname(__file__), os.path.pardir, os.path.pardir)
))
from lib.twitter_api import authentication
class TestAuth(TestCase):
def test_generateAppAccessToken(self):
auth = authentication._generateAppAccessToken()
def test_getTweepyConnection(self):
auth = authentication._generateAppAccessToken()
api = authentication._getTweepyConnection(auth)
def test_getAPIConnection(self):
"""
Test that App Access token can be used to connect to Twitter API.
"""
api = authentication.getAPIConnection(userFlow=False)
def test_getAppOnlyConnection(self):
"""
Test App-only token.
"""
api = authentication.getAppOnlyConnection()
if __name__ == '__main__':
unittest.main()
| // ... existing code ...
from __future__ import absolute_import
import os
import sys
import unittest
from unittest import TestCase
# Allow imports to be done when executing this file directly.
sys.path.insert(0, os.path.abspath(os.path.join(
os.path.dirname(__file__), os.path.pardir, os.path.pardir)
))
from lib.twitter_api import authentication
// ... modified code ...
"""
api = authentication.getAppOnlyConnection()
if __name__ == '__main__':
unittest.main()
// ... rest of the code ... |
8b538c452242050e468b71ca937e3d4feb57887b | mopidy/backends/stream/__init__.py | mopidy/backends/stream/__init__.py | from __future__ import unicode_literals
import mopidy
from mopidy import ext
__doc__ = """A backend for playing music for streaming music.
This backend will handle streaming of URIs in
:attr:`mopidy.settings.STREAM_PROTOCOLS` assuming the right plugins are
installed.
**Issues:**
https://github.com/mopidy/mopidy/issues?labels=Stream+backend
**Dependencies:**
- None
**Settings:**
- :attr:`mopidy.settings.STREAM_PROTOCOLS`
"""
class Extension(ext.Extension):
name = 'Mopidy-Stream'
version = mopidy.__version__
def get_default_config(self):
return '[ext.stream]'
def validate_config(self, config):
pass
def validate_environment(self):
pass
def get_backend_classes(self):
from .actor import StreamBackend
return [StreamBackend]
| from __future__ import unicode_literals
import mopidy
from mopidy import ext
from mopidy.utils import config, formatting
default_config = """
[ext.stream]
# If the stream extension should be enabled or not
enabled = true
# Whitelist of URI schemas to support streaming from
protocols =
http
https
mms
rtmp
rtmps
rtsp
"""
__doc__ = """A backend for playing music for streaming music.
This backend will handle streaming of URIs in
:attr:`mopidy.settings.STREAM_PROTOCOLS` assuming the right plugins are
installed.
**Issues:**
https://github.com/mopidy/mopidy/issues?labels=Stream+backend
**Dependencies:**
- None
**Default config:**
.. code-block:: ini
%(config)s
""" % {'config': formatting.indent(default_config)}
class Extension(ext.Extension):
name = 'Mopidy-Stream'
version = mopidy.__version__
def get_default_config(self):
return default_config
def get_config_schema(self):
schema = config.ExtensionConfigSchema()
schema['protocols'] = config.List()
return schema
def validate_environment(self):
pass
def get_backend_classes(self):
from .actor import StreamBackend
return [StreamBackend]
| Add default config and config schema | stream: Add default config and config schema
| Python | apache-2.0 | tkem/mopidy,jcass77/mopidy,jmarsik/mopidy,ZenithDK/mopidy,swak/mopidy,vrs01/mopidy,diandiankan/mopidy,quartz55/mopidy,adamcik/mopidy,abarisain/mopidy,liamw9534/mopidy,vrs01/mopidy,tkem/mopidy,dbrgn/mopidy,liamw9534/mopidy,SuperStarPL/mopidy,diandiankan/mopidy,abarisain/mopidy,glogiotatidis/mopidy,hkariti/mopidy,mopidy/mopidy,mokieyue/mopidy,mopidy/mopidy,vrs01/mopidy,ali/mopidy,ZenithDK/mopidy,tkem/mopidy,SuperStarPL/mopidy,hkariti/mopidy,ali/mopidy,glogiotatidis/mopidy,adamcik/mopidy,jmarsik/mopidy,kingosticks/mopidy,ZenithDK/mopidy,kingosticks/mopidy,pacificIT/mopidy,glogiotatidis/mopidy,priestd09/mopidy,rawdlite/mopidy,bacontext/mopidy,diandiankan/mopidy,SuperStarPL/mopidy,hkariti/mopidy,mopidy/mopidy,kingosticks/mopidy,diandiankan/mopidy,mokieyue/mopidy,rawdlite/mopidy,swak/mopidy,priestd09/mopidy,ali/mopidy,bacontext/mopidy,bencevans/mopidy,jodal/mopidy,quartz55/mopidy,mokieyue/mopidy,jodal/mopidy,bencevans/mopidy,quartz55/mopidy,quartz55/mopidy,dbrgn/mopidy,mokieyue/mopidy,jmarsik/mopidy,rawdlite/mopidy,bencevans/mopidy,vrs01/mopidy,ZenithDK/mopidy,tkem/mopidy,dbrgn/mopidy,hkariti/mopidy,rawdlite/mopidy,priestd09/mopidy,woutervanwijk/mopidy,jmarsik/mopidy,SuperStarPL/mopidy,bacontext/mopidy,dbrgn/mopidy,jodal/mopidy,swak/mopidy,jcass77/mopidy,pacificIT/mopidy,adamcik/mopidy,bencevans/mopidy,ali/mopidy,bacontext/mopidy,swak/mopidy,pacificIT/mopidy,glogiotatidis/mopidy,pacificIT/mopidy,jcass77/mopidy,woutervanwijk/mopidy | from __future__ import unicode_literals
import mopidy
from mopidy import ext
+ from mopidy.utils import config, formatting
+
+ default_config = """
+ [ext.stream]
+
+ # If the stream extension should be enabled or not
+ enabled = true
+
+ # Whitelist of URI schemas to support streaming from
+ protocols =
+ http
+ https
+ mms
+ rtmp
+ rtmps
+ rtsp
+ """
__doc__ = """A backend for playing music for streaming music.
This backend will handle streaming of URIs in
:attr:`mopidy.settings.STREAM_PROTOCOLS` assuming the right plugins are
installed.
**Issues:**
https://github.com/mopidy/mopidy/issues?labels=Stream+backend
**Dependencies:**
- None
- **Settings:**
+ **Default config:**
- - :attr:`mopidy.settings.STREAM_PROTOCOLS`
- """
+ .. code-block:: ini
+
+ %(config)s
+ """ % {'config': formatting.indent(default_config)}
class Extension(ext.Extension):
name = 'Mopidy-Stream'
version = mopidy.__version__
def get_default_config(self):
- return '[ext.stream]'
+ return default_config
- def validate_config(self, config):
- pass
+ def get_config_schema(self):
+ schema = config.ExtensionConfigSchema()
+ schema['protocols'] = config.List()
+ return schema
def validate_environment(self):
pass
def get_backend_classes(self):
from .actor import StreamBackend
return [StreamBackend]
| Add default config and config schema | ## Code Before:
from __future__ import unicode_literals
import mopidy
from mopidy import ext
__doc__ = """A backend for playing music for streaming music.
This backend will handle streaming of URIs in
:attr:`mopidy.settings.STREAM_PROTOCOLS` assuming the right plugins are
installed.
**Issues:**
https://github.com/mopidy/mopidy/issues?labels=Stream+backend
**Dependencies:**
- None
**Settings:**
- :attr:`mopidy.settings.STREAM_PROTOCOLS`
"""
class Extension(ext.Extension):
name = 'Mopidy-Stream'
version = mopidy.__version__
def get_default_config(self):
return '[ext.stream]'
def validate_config(self, config):
pass
def validate_environment(self):
pass
def get_backend_classes(self):
from .actor import StreamBackend
return [StreamBackend]
## Instruction:
Add default config and config schema
## Code After:
from __future__ import unicode_literals
import mopidy
from mopidy import ext
from mopidy.utils import config, formatting
default_config = """
[ext.stream]
# If the stream extension should be enabled or not
enabled = true
# Whitelist of URI schemas to support streaming from
protocols =
http
https
mms
rtmp
rtmps
rtsp
"""
__doc__ = """A backend for playing music for streaming music.
This backend will handle streaming of URIs in
:attr:`mopidy.settings.STREAM_PROTOCOLS` assuming the right plugins are
installed.
**Issues:**
https://github.com/mopidy/mopidy/issues?labels=Stream+backend
**Dependencies:**
- None
**Default config:**
.. code-block:: ini
%(config)s
""" % {'config': formatting.indent(default_config)}
class Extension(ext.Extension):
name = 'Mopidy-Stream'
version = mopidy.__version__
def get_default_config(self):
return default_config
def get_config_schema(self):
schema = config.ExtensionConfigSchema()
schema['protocols'] = config.List()
return schema
def validate_environment(self):
pass
def get_backend_classes(self):
from .actor import StreamBackend
return [StreamBackend]
| // ... existing code ...
import mopidy
from mopidy import ext
from mopidy.utils import config, formatting
default_config = """
[ext.stream]
# If the stream extension should be enabled or not
enabled = true
# Whitelist of URI schemas to support streaming from
protocols =
http
https
mms
rtmp
rtmps
rtsp
"""
__doc__ = """A backend for playing music for streaming music.
// ... modified code ...
- None
**Default config:**
.. code-block:: ini
%(config)s
""" % {'config': formatting.indent(default_config)}
...
def get_default_config(self):
return default_config
def get_config_schema(self):
schema = config.ExtensionConfigSchema()
schema['protocols'] = config.List()
return schema
def validate_environment(self):
// ... rest of the code ... |
7b05ce75c0dd16944b26f2c53f1508aa3f771d27 | migrations/versions/0177_add_virus_scan_statuses.py | migrations/versions/0177_add_virus_scan_statuses.py | from alembic import op
revision = '0176_alter_billing_columns'
down_revision = '0175_drop_job_statistics_table'
def upgrade():
op.execute("INSERT INTO notification_status_types (name) VALUES ('pending-virus-check')")
op.execute("INSERT INTO notification_status_types (name) VALUES ('virus-scan-failed')")
def downgrade():
op.execute("UPDATE notifications SET notification_status = 'created' WHERE notification_status = 'pending-virus-check'")
op.execute("UPDATE notification_history SET notification_status = 'created' WHERE notification_status = 'pending-virus-check'")
op.execute("UPDATE notifications SET notification_status = 'permanent-failure' WHERE notification_status = 'virus-scan-failed'")
op.execute("UPDATE notification_history SET notification_status = 'permanent-failure' WHERE notification_status = 'virus-scan-failed'")
op.execute("DELETE FROM notification_status_types WHERE name in ('pending-virus-check', 'virus-scan-failed')")
| from alembic import op
revision = '0177_add_virus_scan_statuses'
down_revision = '0176_alter_billing_columns'
def upgrade():
op.execute("INSERT INTO notification_status_types (name) VALUES ('pending-virus-check')")
op.execute("INSERT INTO notification_status_types (name) VALUES ('virus-scan-failed')")
def downgrade():
op.execute("UPDATE notifications SET notification_status = 'created' WHERE notification_status = 'pending-virus-check'")
op.execute("UPDATE notification_history SET notification_status = 'created' WHERE notification_status = 'pending-virus-check'")
op.execute("UPDATE notifications SET notification_status = 'permanent-failure' WHERE notification_status = 'virus-scan-failed'")
op.execute("UPDATE notification_history SET notification_status = 'permanent-failure' WHERE notification_status = 'virus-scan-failed'")
op.execute("DELETE FROM notification_status_types WHERE name in ('pending-virus-check', 'virus-scan-failed')")
| Fix revision numbers in migration 0177 | Fix revision numbers in migration 0177
| Python | mit | alphagov/notifications-api,alphagov/notifications-api | from alembic import op
+ revision = '0177_add_virus_scan_statuses'
- revision = '0176_alter_billing_columns'
+ down_revision = '0176_alter_billing_columns'
- down_revision = '0175_drop_job_statistics_table'
def upgrade():
op.execute("INSERT INTO notification_status_types (name) VALUES ('pending-virus-check')")
op.execute("INSERT INTO notification_status_types (name) VALUES ('virus-scan-failed')")
def downgrade():
op.execute("UPDATE notifications SET notification_status = 'created' WHERE notification_status = 'pending-virus-check'")
op.execute("UPDATE notification_history SET notification_status = 'created' WHERE notification_status = 'pending-virus-check'")
op.execute("UPDATE notifications SET notification_status = 'permanent-failure' WHERE notification_status = 'virus-scan-failed'")
op.execute("UPDATE notification_history SET notification_status = 'permanent-failure' WHERE notification_status = 'virus-scan-failed'")
op.execute("DELETE FROM notification_status_types WHERE name in ('pending-virus-check', 'virus-scan-failed')")
| Fix revision numbers in migration 0177 | ## Code Before:
from alembic import op
revision = '0176_alter_billing_columns'
down_revision = '0175_drop_job_statistics_table'
def upgrade():
op.execute("INSERT INTO notification_status_types (name) VALUES ('pending-virus-check')")
op.execute("INSERT INTO notification_status_types (name) VALUES ('virus-scan-failed')")
def downgrade():
op.execute("UPDATE notifications SET notification_status = 'created' WHERE notification_status = 'pending-virus-check'")
op.execute("UPDATE notification_history SET notification_status = 'created' WHERE notification_status = 'pending-virus-check'")
op.execute("UPDATE notifications SET notification_status = 'permanent-failure' WHERE notification_status = 'virus-scan-failed'")
op.execute("UPDATE notification_history SET notification_status = 'permanent-failure' WHERE notification_status = 'virus-scan-failed'")
op.execute("DELETE FROM notification_status_types WHERE name in ('pending-virus-check', 'virus-scan-failed')")
## Instruction:
Fix revision numbers in migration 0177
## Code After:
from alembic import op
revision = '0177_add_virus_scan_statuses'
down_revision = '0176_alter_billing_columns'
def upgrade():
op.execute("INSERT INTO notification_status_types (name) VALUES ('pending-virus-check')")
op.execute("INSERT INTO notification_status_types (name) VALUES ('virus-scan-failed')")
def downgrade():
op.execute("UPDATE notifications SET notification_status = 'created' WHERE notification_status = 'pending-virus-check'")
op.execute("UPDATE notification_history SET notification_status = 'created' WHERE notification_status = 'pending-virus-check'")
op.execute("UPDATE notifications SET notification_status = 'permanent-failure' WHERE notification_status = 'virus-scan-failed'")
op.execute("UPDATE notification_history SET notification_status = 'permanent-failure' WHERE notification_status = 'virus-scan-failed'")
op.execute("DELETE FROM notification_status_types WHERE name in ('pending-virus-check', 'virus-scan-failed')")
| # ... existing code ...
revision = '0177_add_virus_scan_statuses'
down_revision = '0176_alter_billing_columns'
# ... rest of the code ... |
38b4af0b3c1c6105d68ff453d86107758ef9d751 | preconditions.py | preconditions.py | class PreconditionError (TypeError):
pass
def preconditions(*precs):
def decorate(f):
def g(*a, **kw):
return f(*a, **kw)
return g
return decorate
| import inspect
class PreconditionError (TypeError):
pass
def preconditions(*precs):
precinfo = []
for p in precs:
spec = inspect.getargspec(p)
if spec.varargs or spec.keywords:
raise PreconditionError(
'Precondition {!r} must not accept * nor ** args.'.format(p))
i = -len(spec.defaults)
appargs, closureargs = spec.args[:i], spec.args[i:]
precinfo.append( (appargs, closureargs, p) )
def decorate(f):
def g(*a, **kw):
return f(*a, **kw)
return g
return decorate
| Implement two of the "early" InvalidPreconditionTests which can be checked prior to seeing the wrapping function. | Implement two of the "early" InvalidPreconditionTests which can be checked prior to seeing the wrapping function.
| Python | mit | nejucomo/preconditions | + import inspect
+
+
class PreconditionError (TypeError):
pass
def preconditions(*precs):
+
+ precinfo = []
+ for p in precs:
+ spec = inspect.getargspec(p)
+ if spec.varargs or spec.keywords:
+ raise PreconditionError(
+ 'Precondition {!r} must not accept * nor ** args.'.format(p))
+
+ i = -len(spec.defaults)
+ appargs, closureargs = spec.args[:i], spec.args[i:]
+ precinfo.append( (appargs, closureargs, p) )
+
def decorate(f):
def g(*a, **kw):
return f(*a, **kw)
return g
return decorate
| Implement two of the "early" InvalidPreconditionTests which can be checked prior to seeing the wrapping function. | ## Code Before:
class PreconditionError (TypeError):
pass
def preconditions(*precs):
def decorate(f):
def g(*a, **kw):
return f(*a, **kw)
return g
return decorate
## Instruction:
Implement two of the "early" InvalidPreconditionTests which can be checked prior to seeing the wrapping function.
## Code After:
import inspect
class PreconditionError (TypeError):
pass
def preconditions(*precs):
precinfo = []
for p in precs:
spec = inspect.getargspec(p)
if spec.varargs or spec.keywords:
raise PreconditionError(
'Precondition {!r} must not accept * nor ** args.'.format(p))
i = -len(spec.defaults)
appargs, closureargs = spec.args[:i], spec.args[i:]
precinfo.append( (appargs, closureargs, p) )
def decorate(f):
def g(*a, **kw):
return f(*a, **kw)
return g
return decorate
| ...
import inspect
class PreconditionError (TypeError):
pass
...
def preconditions(*precs):
precinfo = []
for p in precs:
spec = inspect.getargspec(p)
if spec.varargs or spec.keywords:
raise PreconditionError(
'Precondition {!r} must not accept * nor ** args.'.format(p))
i = -len(spec.defaults)
appargs, closureargs = spec.args[:i], spec.args[i:]
precinfo.append( (appargs, closureargs, p) )
def decorate(f):
def g(*a, **kw):
... |
45116fc996b097176bcfa2dcd7fb8c9710f6d66e | tests/test_basics.py | tests/test_basics.py |
import os
from xml.etree import ElementTree
from utils import with_app, pretty_print_xml
#=============================================================================
# Tests
@with_app(buildername="xml", srcdir="basics")
def test_basics(app, status, warning):
app.build()
tree = ElementTree.parse(app.outdir / "index.xml")
pretty_print_xml(tree.getroot())
# Verify that 2 traceables are found.
assert len(tree.findall(".//target")) == 2
assert len(tree.findall(".//index")) == 2
assert len(tree.findall(".//admonition")) == 2
assert len(tree.findall(".//admonition")) == 2
# Verify that child-parent relationship are made.
assert len(tree.findall(".//field_list")) == 2
parent_fields, child_fields = tree.findall(".//field_list")
for field in parent_fields:
field_name = field.findall("./field_name")[0]
if field_name.text == "child":
break
else:
assert False, "Parent's child field not found!"
for field in child_fields:
field_name = field.findall("./field_name")[0]
if field_name.text == "parent":
break
else:
assert False, "Child's parent field not found!"
# Verify that a warning is emitted for unknown traceable tag.
assert (warning.getvalue().find(
"WARNING: Traceables: no traceable with tag"
" 'NONEXISTENT' found!") > 0)
|
import os
from xml.etree import ElementTree
from utils import with_app, pretty_print_xml
#=============================================================================
# Tests
@with_app(buildername="xml", srcdir="basics")
def test_basics(app, status, warning):
app.build()
tree = ElementTree.parse(app.outdir / "index.xml")
# Verify that 2 traceables are found.
assert len(tree.findall(".//target")) == 2
assert len(tree.findall(".//index")) == 2
assert len(tree.findall(".//admonition")) == 2
assert len(tree.findall(".//admonition")) == 2
# Verify that child-parent relationship are made.
assert len(tree.findall(".//field_list")) == 2
parent_fields, child_fields = tree.findall(".//field_list")
for field in parent_fields:
field_name = field.findall("./field_name")[0]
if field_name.text == "child":
break
else:
assert False, "Parent's child field not found!"
for field in child_fields:
field_name = field.findall("./field_name")[0]
if field_name.text == "parent":
break
else:
assert False, "Child's parent field not found!"
# Verify that a warning is emitted for unknown traceable tag.
assert (warning.getvalue().find(
"WARNING: Traceables: no traceable with tag"
" 'NONEXISTENT' found!") > 0)
| Remove debug printing from test case | Remove debug printing from test case
| Python | apache-2.0 | t4ngo/sphinxcontrib-traceables |
import os
from xml.etree import ElementTree
from utils import with_app, pretty_print_xml
#=============================================================================
# Tests
@with_app(buildername="xml", srcdir="basics")
def test_basics(app, status, warning):
app.build()
tree = ElementTree.parse(app.outdir / "index.xml")
- pretty_print_xml(tree.getroot())
# Verify that 2 traceables are found.
assert len(tree.findall(".//target")) == 2
assert len(tree.findall(".//index")) == 2
assert len(tree.findall(".//admonition")) == 2
assert len(tree.findall(".//admonition")) == 2
# Verify that child-parent relationship are made.
assert len(tree.findall(".//field_list")) == 2
parent_fields, child_fields = tree.findall(".//field_list")
for field in parent_fields:
field_name = field.findall("./field_name")[0]
if field_name.text == "child":
break
else:
assert False, "Parent's child field not found!"
for field in child_fields:
field_name = field.findall("./field_name")[0]
if field_name.text == "parent":
break
else:
assert False, "Child's parent field not found!"
# Verify that a warning is emitted for unknown traceable tag.
assert (warning.getvalue().find(
"WARNING: Traceables: no traceable with tag"
" 'NONEXISTENT' found!") > 0)
| Remove debug printing from test case | ## Code Before:
import os
from xml.etree import ElementTree
from utils import with_app, pretty_print_xml
#=============================================================================
# Tests
@with_app(buildername="xml", srcdir="basics")
def test_basics(app, status, warning):
app.build()
tree = ElementTree.parse(app.outdir / "index.xml")
pretty_print_xml(tree.getroot())
# Verify that 2 traceables are found.
assert len(tree.findall(".//target")) == 2
assert len(tree.findall(".//index")) == 2
assert len(tree.findall(".//admonition")) == 2
assert len(tree.findall(".//admonition")) == 2
# Verify that child-parent relationship are made.
assert len(tree.findall(".//field_list")) == 2
parent_fields, child_fields = tree.findall(".//field_list")
for field in parent_fields:
field_name = field.findall("./field_name")[0]
if field_name.text == "child":
break
else:
assert False, "Parent's child field not found!"
for field in child_fields:
field_name = field.findall("./field_name")[0]
if field_name.text == "parent":
break
else:
assert False, "Child's parent field not found!"
# Verify that a warning is emitted for unknown traceable tag.
assert (warning.getvalue().find(
"WARNING: Traceables: no traceable with tag"
" 'NONEXISTENT' found!") > 0)
## Instruction:
Remove debug printing from test case
## Code After:
import os
from xml.etree import ElementTree
from utils import with_app, pretty_print_xml
#=============================================================================
# Tests
@with_app(buildername="xml", srcdir="basics")
def test_basics(app, status, warning):
app.build()
tree = ElementTree.parse(app.outdir / "index.xml")
# Verify that 2 traceables are found.
assert len(tree.findall(".//target")) == 2
assert len(tree.findall(".//index")) == 2
assert len(tree.findall(".//admonition")) == 2
assert len(tree.findall(".//admonition")) == 2
# Verify that child-parent relationship are made.
assert len(tree.findall(".//field_list")) == 2
parent_fields, child_fields = tree.findall(".//field_list")
for field in parent_fields:
field_name = field.findall("./field_name")[0]
if field_name.text == "child":
break
else:
assert False, "Parent's child field not found!"
for field in child_fields:
field_name = field.findall("./field_name")[0]
if field_name.text == "parent":
break
else:
assert False, "Child's parent field not found!"
# Verify that a warning is emitted for unknown traceable tag.
assert (warning.getvalue().find(
"WARNING: Traceables: no traceable with tag"
" 'NONEXISTENT' found!") > 0)
| ...
app.build()
tree = ElementTree.parse(app.outdir / "index.xml")
# Verify that 2 traceables are found.
... |
8932d0717bf57c86b81b6744353d6387821b8b15 | wsgi/setup.py | wsgi/setup.py |
import subprocess
import sys
import setup_util
import os
def start(args):
subprocess.Popen("gunicorn hello:app -b 0.0.0.0:8080 -w " + str((args.max_threads * 2)) + " --log-level=critical", shell=True, cwd="wsgi")
return 0
def stop():
p = subprocess.Popen(['ps', 'aux'], stdout=subprocess.PIPE)
out, err = p.communicate()
for line in out.splitlines():
if 'gunicorn' in line:
try:
pid = int(line.split(None, 2)[1])
os.kill(pid, 9)
except OSError:
pass
return 0 |
import subprocess
import sys
import setup_util
import os
def start(args):
subprocess.Popen('gunicorn hello:app --worker-class="egg:meinheld#gunicorn_worker" -b 0.0.0.0:8080 -w '
+ str((args.max_threads * 2)) + " --log-level=critical", shell=True, cwd="wsgi")
return 0
def stop():
p = subprocess.Popen(['ps', 'aux'], stdout=subprocess.PIPE)
out, err = p.communicate()
for line in out.splitlines():
if 'gunicorn' in line:
try:
pid = int(line.split(None, 2)[1])
os.kill(pid, 9)
except OSError:
pass
return 0
| Use meinheld worker (same as other Python Frameworks) | wsgi: Use meinheld worker (same as other Python Frameworks)
| Python | bsd-3-clause | jamming/FrameworkBenchmarks,s-ludwig/FrameworkBenchmarks,Ocramius/FrameworkBenchmarks,diablonhn/FrameworkBenchmarks,ashawnbandy-te-tfb/FrameworkBenchmarks,zloster/FrameworkBenchmarks,donovanmuller/FrameworkBenchmarks,zdanek/FrameworkBenchmarks,markkolich/FrameworkBenchmarks,circlespainter/FrameworkBenchmarks,saturday06/FrameworkBenchmarks,sgml/FrameworkBenchmarks,mfirry/FrameworkBenchmarks,nathana1/FrameworkBenchmarks,s-ludwig/FrameworkBenchmarks,victorbriz/FrameworkBenchmarks,ratpack/FrameworkBenchmarks,diablonhn/FrameworkBenchmarks,s-ludwig/FrameworkBenchmarks,jetty-project/FrameworkBenchmarks,ashawnbandy-te-tfb/FrameworkBenchmarks,nathana1/FrameworkBenchmarks,leafo/FrameworkBenchmarks,zane-techempower/FrameworkBenchmarks,greg-hellings/FrameworkBenchmarks,Synchro/FrameworkBenchmarks,F3Community/FrameworkBenchmarks,knewmanTE/FrameworkBenchmarks,sgml/FrameworkBenchmarks,circlespainter/FrameworkBenchmarks,xitrum-framework/FrameworkBenchmarks,ratpack/FrameworkBenchmarks,marko-asplund/FrameworkBenchmarks,nathana1/FrameworkBenchmarks,F3Community/FrameworkBenchmarks,kbrock/FrameworkBenchmarks,nbrady-techempower/FrameworkBenchmarks,jeevatkm/FrameworkBenchmarks,julienschmidt/FrameworkBenchmarks,alubbe/FrameworkBenchmarks,joshk/FrameworkBenchmarks,zdanek/FrameworkBenchmarks,alubbe/FrameworkBenchmarks,saturday06/FrameworkBenchmarks,youprofit/FrameworkBenchmarks,actframework/FrameworkBenchmarks,doom369/FrameworkBenchmarks,saturday06/FrameworkBenchmarks,joshk/FrameworkBenchmarks,greenlaw110/FrameworkBenchmarks,denkab/FrameworkBenchmarks,martin-g/FrameworkBenchmarks,waiteb3/FrameworkBenchmarks,marko-asplund/FrameworkBenchmarks,Eyepea/FrameworkBenchmarks,zloster/FrameworkBenchmarks,PermeAgility/FrameworkBenchmarks,Ocramius/FrameworkBenchmarks,RockinRoel/FrameworkBenchmarks,knewmanTE/FrameworkBenchmarks,jetty-project/FrameworkBenchmarks,nbrady-techempower/FrameworkBenchmarks,stefanocasazza/FrameworkBenchmarks,fabianmurariu/FrameworkBenchmarks,stefanocasazza/FrameworkBenchmarks,lcp0578/FrameworkBenchmarks,zloster/FrameworkBenchmarks,marko-asplund/FrameworkBenchmarks,sxend/FrameworkBenchmarks,jamming/FrameworkBenchmarks,zdanek/FrameworkBenchmarks,Eyepea/FrameworkBenchmarks,herloct/FrameworkBenchmarks,F3Community/FrameworkBenchmarks,grob/FrameworkBenchmarks,grob/FrameworkBenchmarks,hamiltont/FrameworkBenchmarks,mfirry/FrameworkBenchmarks,raziel057/FrameworkBenchmarks,khellang/FrameworkBenchmarks,stefanocasazza/FrameworkBenchmarks,sagenschneider/FrameworkBenchmarks,herloct/FrameworkBenchmarks,donovanmuller/FrameworkBenchmarks,hamiltont/FrameworkBenchmarks,F3Community/FrameworkBenchmarks,zapov/FrameworkBenchmarks,Eyepea/FrameworkBenchmarks,actframework/FrameworkBenchmarks,Ocramius/FrameworkBenchmarks,zapov/FrameworkBenchmarks,knewmanTE/FrameworkBenchmarks,greenlaw110/FrameworkBenchmarks,k-r-g/FrameworkBenchmarks,doom369/FrameworkBenchmarks,steveklabnik/FrameworkBenchmarks,saturday06/FrameworkBenchmarks,sgml/FrameworkBenchmarks,denkab/FrameworkBenchmarks,testn/FrameworkBenchmarks,denkab/FrameworkBenchmarks,psfblair/FrameworkBenchmarks,xitrum-framework/FrameworkBenchmarks,PermeAgility/FrameworkBenchmarks,nbrady-techempower/FrameworkBenchmarks,circlespainter/FrameworkBenchmarks,testn/FrameworkBenchmarks,s-ludwig/FrameworkBenchmarks,ashawnbandy-te-tfb/FrameworkBenchmarks,markkolich/FrameworkBenchmarks,jaguililla/FrameworkBenchmarks,zdanek/FrameworkBenchmarks,Verber/FrameworkBenchmarks,zloster/FrameworkBenchmarks,psfblair/FrameworkBenchmarks,zdanek/FrameworkBenchmarks,kostya-sh/FrameworkBenchmarks,greenlaw110/FrameworkBenchmarks,torhve/FrameworkBenchmarks,nathana1/FrameworkBenchmarks,Rayne/FrameworkBenchmarks,youprofit/FrameworkBenchmarks,diablonhn/FrameworkBenchmarks,raziel057/FrameworkBenchmarks,Jesterovskiy/FrameworkBenchmarks,grob/FrameworkBenchmarks,testn/FrameworkBenchmarks,Verber/FrameworkBenchmarks,nkasvosve/FrameworkBenchmarks,nkasvosve/FrameworkBenchmarks,methane/FrameworkBenchmarks,k-r-g/FrameworkBenchmarks,seem-sky/FrameworkBenchmarks,testn/FrameworkBenchmarks,Jesterovskiy/FrameworkBenchmarks,Rayne/FrameworkBenchmarks,Verber/FrameworkBenchmarks,psfblair/FrameworkBenchmarks,Eyepea/FrameworkBenchmarks,zloster/FrameworkBenchmarks,sanjoydesk/FrameworkBenchmarks,jebbstewart/FrameworkBenchmarks,alubbe/FrameworkBenchmarks,dmacd/FB-try1,kellabyte/FrameworkBenchmarks,nkasvosve/FrameworkBenchmarks,nathana1/FrameworkBenchmarks,actframework/FrameworkBenchmarks,knewmanTE/FrameworkBenchmarks,zane-techempower/FrameworkBenchmarks,marko-asplund/FrameworkBenchmarks,seem-sky/FrameworkBenchmarks,zapov/FrameworkBenchmarks,RockinRoel/FrameworkBenchmarks,diablonhn/FrameworkBenchmarks,zloster/FrameworkBenchmarks,steveklabnik/FrameworkBenchmarks,dmacd/FB-try1,alubbe/FrameworkBenchmarks,Rayne/FrameworkBenchmarks,grob/FrameworkBenchmarks,zloster/FrameworkBenchmarks,Dith3r/FrameworkBenchmarks,lcp0578/FrameworkBenchmarks,RockinRoel/FrameworkBenchmarks,sagenschneider/FrameworkBenchmarks,donovanmuller/FrameworkBenchmarks,saturday06/FrameworkBenchmarks,steveklabnik/FrameworkBenchmarks,PermeAgility/FrameworkBenchmarks,jetty-project/FrameworkBenchmarks,Verber/FrameworkBenchmarks,denkab/FrameworkBenchmarks,mfirry/FrameworkBenchmarks,seem-sky/FrameworkBenchmarks,ratpack/FrameworkBenchmarks,nbrady-techempower/FrameworkBenchmarks,zapov/FrameworkBenchmarks,Eyepea/FrameworkBenchmarks,stefanocasazza/FrameworkBenchmarks,testn/FrameworkBenchmarks,lcp0578/FrameworkBenchmarks,valyala/FrameworkBenchmarks,julienschmidt/FrameworkBenchmarks,torhve/FrameworkBenchmarks,valyala/FrameworkBenchmarks,ashawnbandy-te-tfb/FrameworkBenchmarks,joshk/FrameworkBenchmarks,MTDdk/FrameworkBenchmarks,denkab/FrameworkBenchmarks,Jesterovskiy/FrameworkBenchmarks,thousandsofthem/FrameworkBenchmarks,steveklabnik/FrameworkBenchmarks,diablonhn/FrameworkBenchmarks,steveklabnik/FrameworkBenchmarks,kellabyte/FrameworkBenchmarks,dmacd/FB-try1,valyala/FrameworkBenchmarks,doom369/FrameworkBenchmarks,steveklabnik/FrameworkBenchmarks,jeevatkm/FrameworkBenchmarks,herloct/FrameworkBenchmarks,waiteb3/FrameworkBenchmarks,markkolich/FrameworkBenchmarks,nathana1/FrameworkBenchmarks,waiteb3/FrameworkBenchmarks,jamming/FrameworkBenchmarks,mfirry/FrameworkBenchmarks,MTDdk/FrameworkBenchmarks,greg-hellings/FrameworkBenchmarks,zhuochenKIDD/FrameworkBenchmarks,jamming/FrameworkBenchmarks,diablonhn/FrameworkBenchmarks,Ocramius/FrameworkBenchmarks,denkab/FrameworkBenchmarks,nkasvosve/FrameworkBenchmarks,zapov/FrameworkBenchmarks,donovanmuller/FrameworkBenchmarks,psfblair/FrameworkBenchmarks,jaguililla/FrameworkBenchmarks,greenlaw110/FrameworkBenchmarks,Rydgel/FrameworkBenchmarks,alubbe/FrameworkBenchmarks,yunspace/FrameworkBenchmarks,jaguililla/FrameworkBenchmarks,hperadin/FrameworkBenchmarks,hperadin/FrameworkBenchmarks,nbrady-techempower/FrameworkBenchmarks,nathana1/FrameworkBenchmarks,jebbstewart/FrameworkBenchmarks,Ocramius/FrameworkBenchmarks,nkasvosve/FrameworkBenchmarks,denkab/FrameworkBenchmarks,greg-hellings/FrameworkBenchmarks,jebbstewart/FrameworkBenchmarks,xitrum-framework/FrameworkBenchmarks,PermeAgility/FrameworkBenchmarks,markkolich/FrameworkBenchmarks,hamiltont/FrameworkBenchmarks,Dith3r/FrameworkBenchmarks,xitrum-framework/FrameworkBenchmarks,testn/FrameworkBenchmarks,sxend/FrameworkBenchmarks,kbrock/FrameworkBenchmarks,kostya-sh/FrameworkBenchmarks,doom369/FrameworkBenchmarks,nathana1/FrameworkBenchmarks,raziel057/FrameworkBenchmarks,knewmanTE/FrameworkBenchmarks,hamiltont/FrameworkBenchmarks,psfblair/FrameworkBenchmarks,raziel057/FrameworkBenchmarks,alubbe/FrameworkBenchmarks,Ocramius/FrameworkBenchmarks,alubbe/FrameworkBenchmarks,PermeAgility/FrameworkBenchmarks,k-r-g/FrameworkBenchmarks,methane/FrameworkBenchmarks,waiteb3/FrameworkBenchmarks,RockinRoel/FrameworkBenchmarks,jetty-project/FrameworkBenchmarks,circlespainter/FrameworkBenchmarks,F3Community/FrameworkBenchmarks,nathana1/FrameworkBenchmarks,k-r-g/FrameworkBenchmarks,jaguililla/FrameworkBenchmarks,lcp0578/FrameworkBenchmarks,Ocramius/FrameworkBenchmarks,ashawnbandy-te-tfb/FrameworkBenchmarks,kellabyte/FrameworkBenchmarks,s-ludwig/FrameworkBenchmarks,jaguililla/FrameworkBenchmarks,martin-g/FrameworkBenchmarks,khellang/FrameworkBenchmarks,khellang/FrameworkBenchmarks,grob/FrameworkBenchmarks,greg-hellings/FrameworkBenchmarks,methane/FrameworkBenchmarks,jeevatkm/FrameworkBenchmarks,khellang/FrameworkBenchmarks,seem-sky/FrameworkBenchmarks,hamiltont/FrameworkBenchmarks,sanjoydesk/FrameworkBenchmarks,dmacd/FB-try1,sanjoydesk/FrameworkBenchmarks,victorbriz/FrameworkBenchmarks,Eyepea/FrameworkBenchmarks,sxend/FrameworkBenchmarks,testn/FrameworkBenchmarks,victorbriz/FrameworkBenchmarks,fabianmurariu/FrameworkBenchmarks,alubbe/FrameworkBenchmarks,greg-hellings/FrameworkBenchmarks,diablonhn/FrameworkBenchmarks,Dith3r/FrameworkBenchmarks,yunspace/FrameworkBenchmarks,Synchro/FrameworkBenchmarks,grob/FrameworkBenchmarks,donovanmuller/FrameworkBenchmarks,herloct/FrameworkBenchmarks,donovanmuller/FrameworkBenchmarks,yunspace/FrameworkBenchmarks,jeevatkm/FrameworkBenchmarks,steveklabnik/FrameworkBenchmarks,Rayne/FrameworkBenchmarks,sanjoydesk/FrameworkBenchmarks,nkasvosve/FrameworkBenchmarks,sagenschneider/FrameworkBenchmarks,herloct/FrameworkBenchmarks,zhuochenKIDD/FrameworkBenchmarks,psfblair/FrameworkBenchmarks,marko-asplund/FrameworkBenchmarks,thousandsofthem/FrameworkBenchmarks,zloster/FrameworkBenchmarks,jetty-project/FrameworkBenchmarks,hamiltont/FrameworkBenchmarks,zane-techempower/FrameworkBenchmarks,torhve/FrameworkBenchmarks,herloct/FrameworkBenchmarks,knewmanTE/FrameworkBenchmarks,zapov/FrameworkBenchmarks,victorbriz/FrameworkBenchmarks,kellabyte/FrameworkBenchmarks,sagenschneider/FrameworkBenchmarks,zhuochenKIDD/FrameworkBenchmarks,diablonhn/FrameworkBenchmarks,zhuochenKIDD/FrameworkBenchmarks,F3Community/FrameworkBenchmarks,jeevatkm/FrameworkBenchmarks,ashawnbandy-te-tfb/FrameworkBenchmarks,jetty-project/FrameworkBenchmarks,zane-techempower/FrameworkBenchmarks,leafo/FrameworkBenchmarks,nathana1/FrameworkBenchmarks,sagenschneider/FrameworkBenchmarks,greg-hellings/FrameworkBenchmarks,zloster/FrameworkBenchmarks,Eyepea/FrameworkBenchmarks,Rydgel/FrameworkBenchmarks,PermeAgility/FrameworkBenchmarks,ashawnbandy-te-tfb/FrameworkBenchmarks,sagenschneider/FrameworkBenchmarks,Verber/FrameworkBenchmarks,martin-g/FrameworkBenchmarks,sxend/FrameworkBenchmarks,yunspace/FrameworkBenchmarks,stefanocasazza/FrameworkBenchmarks,methane/FrameworkBenchmarks,julienschmidt/FrameworkBenchmarks,hperadin/FrameworkBenchmarks,seem-sky/FrameworkBenchmarks,ashawnbandy-te-tfb/FrameworkBenchmarks,zane-techempower/FrameworkBenchmarks,s-ludwig/FrameworkBenchmarks,zloster/FrameworkBenchmarks,F3Community/FrameworkBenchmarks,Dith3r/FrameworkBenchmarks,thousandsofthem/FrameworkBenchmarks,mfirry/FrameworkBenchmarks,MTDdk/FrameworkBenchmarks,raziel057/FrameworkBenchmarks,Synchro/FrameworkBenchmarks,Verber/FrameworkBenchmarks,kostya-sh/FrameworkBenchmarks,ratpack/FrameworkBenchmarks,stefanocasazza/FrameworkBenchmarks,greenlaw110/FrameworkBenchmarks,kostya-sh/FrameworkBenchmarks,fabianmurariu/FrameworkBenchmarks,knewmanTE/FrameworkBenchmarks,xitrum-framework/FrameworkBenchmarks,stefanocasazza/FrameworkBenchmarks,Synchro/FrameworkBenchmarks,nbrady-techempower/FrameworkBenchmarks,joshk/FrameworkBenchmarks,ratpack/FrameworkBenchmarks,mfirry/FrameworkBenchmarks,torhve/FrameworkBenchmarks,raziel057/FrameworkBenchmarks,victorbriz/FrameworkBenchmarks,Dith3r/FrameworkBenchmarks,xitrum-framework/FrameworkBenchmarks,kostya-sh/FrameworkBenchmarks,Eyepea/FrameworkBenchmarks,waiteb3/FrameworkBenchmarks,zane-techempower/FrameworkBenchmarks,actframework/FrameworkBenchmarks,steveklabnik/FrameworkBenchmarks,seem-sky/FrameworkBenchmarks,martin-g/FrameworkBenchmarks,zdanek/FrameworkBenchmarks,stefanocasazza/FrameworkBenchmarks,jetty-project/FrameworkBenchmarks,jeevatkm/FrameworkBenchmarks,diablonhn/FrameworkBenchmarks,RockinRoel/FrameworkBenchmarks,sxend/FrameworkBenchmarks,jeevatkm/FrameworkBenchmarks,Synchro/FrameworkBenchmarks,julienschmidt/FrameworkBenchmarks,jeevatkm/FrameworkBenchmarks,jeevatkm/FrameworkBenchmarks,hperadin/FrameworkBenchmarks,jamming/FrameworkBenchmarks,MTDdk/FrameworkBenchmarks,waiteb3/FrameworkBenchmarks,fabianmurariu/FrameworkBenchmarks,k-r-g/FrameworkBenchmarks,xitrum-framework/FrameworkBenchmarks,sgml/FrameworkBenchmarks,martin-g/FrameworkBenchmarks,PermeAgility/FrameworkBenchmarks,RockinRoel/FrameworkBenchmarks,RockinRoel/FrameworkBenchmarks,leafo/FrameworkBenchmarks,zdanek/FrameworkBenchmarks,methane/FrameworkBenchmarks,fabianmurariu/FrameworkBenchmarks,Ocramius/FrameworkBenchmarks,youprofit/FrameworkBenchmarks,seem-sky/FrameworkBenchmarks,herloct/FrameworkBenchmarks,sxend/FrameworkBenchmarks,nkasvosve/FrameworkBenchmarks,ratpack/FrameworkBenchmarks,testn/FrameworkBenchmarks,stefanocasazza/FrameworkBenchmarks,knewmanTE/FrameworkBenchmarks,nbrady-techempower/FrameworkBenchmarks,nbrady-techempower/FrameworkBenchmarks,zane-techempower/FrameworkBenchmarks,MTDdk/FrameworkBenchmarks,RockinRoel/FrameworkBenchmarks,herloct/FrameworkBenchmarks,zdanek/FrameworkBenchmarks,jeevatkm/FrameworkBenchmarks,psfblair/FrameworkBenchmarks,saturday06/FrameworkBenchmarks,doom369/FrameworkBenchmarks,sxend/FrameworkBenchmarks,saturday06/FrameworkBenchmarks,grob/FrameworkBenchmarks,Ocramius/FrameworkBenchmarks,Jesterovskiy/FrameworkBenchmarks,lcp0578/FrameworkBenchmarks,kostya-sh/FrameworkBenchmarks,actframework/FrameworkBenchmarks,ratpack/FrameworkBenchmarks,s-ludwig/FrameworkBenchmarks,circlespainter/FrameworkBenchmarks,PermeAgility/FrameworkBenchmarks,saturday06/FrameworkBenchmarks,thousandsofthem/FrameworkBenchmarks,sanjoydesk/FrameworkBenchmarks,yunspace/FrameworkBenchmarks,markkolich/FrameworkBenchmarks,nkasvosve/FrameworkBenchmarks,Jesterovskiy/FrameworkBenchmarks,waiteb3/FrameworkBenchmarks,sgml/FrameworkBenchmarks,jeevatkm/FrameworkBenchmarks,victorbriz/FrameworkBenchmarks,F3Community/FrameworkBenchmarks,herloct/FrameworkBenchmarks,greenlaw110/FrameworkBenchmarks,waiteb3/FrameworkBenchmarks,Dith3r/FrameworkBenchmarks,Eyepea/FrameworkBenchmarks,Rydgel/FrameworkBenchmarks,actframework/FrameworkBenchmarks,Rydgel/FrameworkBenchmarks,sxend/FrameworkBenchmarks,yunspace/FrameworkBenchmarks,youprofit/FrameworkBenchmarks,youprofit/FrameworkBenchmarks,knewmanTE/FrameworkBenchmarks,hperadin/FrameworkBenchmarks,zloster/FrameworkBenchmarks,fabianmurariu/FrameworkBenchmarks,lcp0578/FrameworkBenchmarks,Verber/FrameworkBenchmarks,sgml/FrameworkBenchmarks,donovanmuller/FrameworkBenchmarks,kostya-sh/FrameworkBenchmarks,Eyepea/FrameworkBenchmarks,doom369/FrameworkBenchmarks,julienschmidt/FrameworkBenchmarks,testn/FrameworkBenchmarks,Eyepea/FrameworkBenchmarks,Jesterovskiy/FrameworkBenchmarks,Verber/FrameworkBenchmarks,grob/FrameworkBenchmarks,PermeAgility/FrameworkBenchmarks,herloct/FrameworkBenchmarks,donovanmuller/FrameworkBenchmarks,markkolich/FrameworkBenchmarks,Rayne/FrameworkBenchmarks,ashawnbandy-te-tfb/FrameworkBenchmarks,methane/FrameworkBenchmarks,Verber/FrameworkBenchmarks,denkab/FrameworkBenchmarks,markkolich/FrameworkBenchmarks,nkasvosve/FrameworkBenchmarks,saturday06/FrameworkBenchmarks,mfirry/FrameworkBenchmarks,khellang/FrameworkBenchmarks,Synchro/FrameworkBenchmarks,k-r-g/FrameworkBenchmarks,valyala/FrameworkBenchmarks,youprofit/FrameworkBenchmarks,Rayne/FrameworkBenchmarks,seem-sky/FrameworkBenchmarks,julienschmidt/FrameworkBenchmarks,Rayne/FrameworkBenchmarks,jebbstewart/FrameworkBenchmarks,zapov/FrameworkBenchmarks,kbrock/FrameworkBenchmarks,donovanmuller/FrameworkBenchmarks,kbrock/FrameworkBenchmarks,raziel057/FrameworkBenchmarks,diablonhn/FrameworkBenchmarks,greg-hellings/FrameworkBenchmarks,hperadin/FrameworkBenchmarks,kostya-sh/FrameworkBenchmarks,youprofit/FrameworkBenchmarks,testn/FrameworkBenchmarks,testn/FrameworkBenchmarks,psfblair/FrameworkBenchmarks,Dith3r/FrameworkBenchmarks,stefanocasazza/FrameworkBenchmarks,jaguililla/FrameworkBenchmarks,F3Community/FrameworkBenchmarks,torhve/FrameworkBenchmarks,jetty-project/FrameworkBenchmarks,kellabyte/FrameworkBenchmarks,Rydgel/FrameworkBenchmarks,martin-g/FrameworkBenchmarks,zapov/FrameworkBenchmarks,circlespainter/FrameworkBenchmarks,marko-asplund/FrameworkBenchmarks,xitrum-framework/FrameworkBenchmarks,stefanocasazza/FrameworkBenchmarks,seem-sky/FrameworkBenchmarks,mfirry/FrameworkBenchmarks,alubbe/FrameworkBenchmarks,khellang/FrameworkBenchmarks,methane/FrameworkBenchmarks,s-ludwig/FrameworkBenchmarks,Rydgel/FrameworkBenchmarks,raziel057/FrameworkBenchmarks,leafo/FrameworkBenchmarks,mfirry/FrameworkBenchmarks,kellabyte/FrameworkBenchmarks,thousandsofthem/FrameworkBenchmarks,sagenschneider/FrameworkBenchmarks,khellang/FrameworkBenchmarks,jebbstewart/FrameworkBenchmarks,thousandsofthem/FrameworkBenchmarks,nkasvosve/FrameworkBenchmarks,khellang/FrameworkBenchmarks,dmacd/FB-try1,martin-g/FrameworkBenchmarks,yunspace/FrameworkBenchmarks,Rydgel/FrameworkBenchmarks,nkasvosve/FrameworkBenchmarks,sagenschneider/FrameworkBenchmarks,jamming/FrameworkBenchmarks,jetty-project/FrameworkBenchmarks,ratpack/FrameworkBenchmarks,markkolich/FrameworkBenchmarks,greenlaw110/FrameworkBenchmarks,nbrady-techempower/FrameworkBenchmarks,greenlaw110/FrameworkBenchmarks,actframework/FrameworkBenchmarks,Rydgel/FrameworkBenchmarks,xitrum-framework/FrameworkBenchmarks,raziel057/FrameworkBenchmarks,Eyepea/FrameworkBenchmarks,RockinRoel/FrameworkBenchmarks,stefanocasazza/FrameworkBenchmarks,mfirry/FrameworkBenchmarks,martin-g/FrameworkBenchmarks,joshk/FrameworkBenchmarks,denkab/FrameworkBenchmarks,psfblair/FrameworkBenchmarks,circlespainter/FrameworkBenchmarks,youprofit/FrameworkBenchmarks,thousandsofthem/FrameworkBenchmarks,seem-sky/FrameworkBenchmarks,marko-asplund/FrameworkBenchmarks,jaguililla/FrameworkBenchmarks,sanjoydesk/FrameworkBenchmarks,khellang/FrameworkBenchmarks,yunspace/FrameworkBenchmarks,sxend/FrameworkBenchmarks,Jesterovskiy/FrameworkBenchmarks,leafo/FrameworkBenchmarks,MTDdk/FrameworkBenchmarks,dmacd/FB-try1,kostya-sh/FrameworkBenchmarks,PermeAgility/FrameworkBenchmarks,khellang/FrameworkBenchmarks,psfblair/FrameworkBenchmarks,sanjoydesk/FrameworkBenchmarks,methane/FrameworkBenchmarks,jaguililla/FrameworkBenchmarks,kbrock/FrameworkBenchmarks,markkolich/FrameworkBenchmarks,Dith3r/FrameworkBenchmarks,s-ludwig/FrameworkBenchmarks,julienschmidt/FrameworkBenchmarks,grob/FrameworkBenchmarks,PermeAgility/FrameworkBenchmarks,zhuochenKIDD/FrameworkBenchmarks,jebbstewart/FrameworkBenchmarks,torhve/FrameworkBenchmarks,MTDdk/FrameworkBenchmarks,jaguililla/FrameworkBenchmarks,Jesterovskiy/FrameworkBenchmarks,marko-asplund/FrameworkBenchmarks,sanjoydesk/FrameworkBenchmarks,sgml/FrameworkBenchmarks,sgml/FrameworkBenchmarks,circlespainter/FrameworkBenchmarks,markkolich/FrameworkBenchmarks,grob/FrameworkBenchmarks,nkasvosve/FrameworkBenchmarks,zhuochenKIDD/FrameworkBenchmarks,joshk/FrameworkBenchmarks,martin-g/FrameworkBenchmarks,F3Community/FrameworkBenchmarks,seem-sky/FrameworkBenchmarks,kostya-sh/FrameworkBenchmarks,k-r-g/FrameworkBenchmarks,jetty-project/FrameworkBenchmarks,yunspace/FrameworkBenchmarks,s-ludwig/FrameworkBenchmarks,s-ludwig/FrameworkBenchmarks,Ocramius/FrameworkBenchmarks,herloct/FrameworkBenchmarks,knewmanTE/FrameworkBenchmarks,alubbe/FrameworkBenchmarks,steveklabnik/FrameworkBenchmarks,zane-techempower/FrameworkBenchmarks,Verber/FrameworkBenchmarks,knewmanTE/FrameworkBenchmarks,martin-g/FrameworkBenchmarks,k-r-g/FrameworkBenchmarks,k-r-g/FrameworkBenchmarks,markkolich/FrameworkBenchmarks,jebbstewart/FrameworkBenchmarks,donovanmuller/FrameworkBenchmarks,alubbe/FrameworkBenchmarks,torhve/FrameworkBenchmarks,Ocramius/FrameworkBenchmarks,Ocramius/FrameworkBenchmarks,leafo/FrameworkBenchmarks,diablonhn/FrameworkBenchmarks,k-r-g/FrameworkBenchmarks,sgml/FrameworkBenchmarks,kellabyte/FrameworkBenchmarks,xitrum-framework/FrameworkBenchmarks,nathana1/FrameworkBenchmarks,stefanocasazza/FrameworkBenchmarks,zane-techempower/FrameworkBenchmarks,hperadin/FrameworkBenchmarks,saturday06/FrameworkBenchmarks,RockinRoel/FrameworkBenchmarks,zhuochenKIDD/FrameworkBenchmarks,Rayne/FrameworkBenchmarks,kbrock/FrameworkBenchmarks,hperadin/FrameworkBenchmarks,actframework/FrameworkBenchmarks,hperadin/FrameworkBenchmarks,MTDdk/FrameworkBenchmarks,valyala/FrameworkBenchmarks,doom369/FrameworkBenchmarks,Rydgel/FrameworkBenchmarks,xitrum-framework/FrameworkBenchmarks,markkolich/FrameworkBenchmarks,jamming/FrameworkBenchmarks,zhuochenKIDD/FrameworkBenchmarks,hamiltont/FrameworkBenchmarks,MTDdk/FrameworkBenchmarks,marko-asplund/FrameworkBenchmarks,jebbstewart/FrameworkBenchmarks,thousandsofthem/FrameworkBenchmarks,Dith3r/FrameworkBenchmarks,victorbriz/FrameworkBenchmarks,greg-hellings/FrameworkBenchmarks,RockinRoel/FrameworkBenchmarks,kbrock/FrameworkBenchmarks,ratpack/FrameworkBenchmarks,valyala/FrameworkBenchmarks,victorbriz/FrameworkBenchmarks,grob/FrameworkBenchmarks,khellang/FrameworkBenchmarks,torhve/FrameworkBenchmarks,MTDdk/FrameworkBenchmarks,zloster/FrameworkBenchmarks,zdanek/FrameworkBenchmarks,zapov/FrameworkBenchmarks,doom369/FrameworkBenchmarks,doom369/FrameworkBenchmarks,psfblair/FrameworkBenchmarks,stefanocasazza/FrameworkBenchmarks,Rayne/FrameworkBenchmarks,sxend/FrameworkBenchmarks,zapov/FrameworkBenchmarks,ratpack/FrameworkBenchmarks,mfirry/FrameworkBenchmarks,raziel057/FrameworkBenchmarks,knewmanTE/FrameworkBenchmarks,ashawnbandy-te-tfb/FrameworkBenchmarks,zloster/FrameworkBenchmarks,martin-g/FrameworkBenchmarks,zhuochenKIDD/FrameworkBenchmarks,xitrum-framework/FrameworkBenchmarks,sagenschneider/FrameworkBenchmarks,nathana1/FrameworkBenchmarks,xitrum-framework/FrameworkBenchmarks,greg-hellings/FrameworkBenchmarks,joshk/FrameworkBenchmarks,thousandsofthem/FrameworkBenchmarks,raziel057/FrameworkBenchmarks,steveklabnik/FrameworkBenchmarks,Eyepea/FrameworkBenchmarks,Synchro/FrameworkBenchmarks,julienschmidt/FrameworkBenchmarks,doom369/FrameworkBenchmarks,circlespainter/FrameworkBenchmarks,youprofit/FrameworkBenchmarks,zhuochenKIDD/FrameworkBenchmarks,Eyepea/FrameworkBenchmarks,kbrock/FrameworkBenchmarks,victorbriz/FrameworkBenchmarks,k-r-g/FrameworkBenchmarks,doom369/FrameworkBenchmarks,nathana1/FrameworkBenchmarks,nbrady-techempower/FrameworkBenchmarks,Rydgel/FrameworkBenchmarks,marko-asplund/FrameworkBenchmarks,Jesterovskiy/FrameworkBenchmarks,nathana1/FrameworkBenchmarks,sxend/FrameworkBenchmarks,Eyepea/FrameworkBenchmarks,Rydgel/FrameworkBenchmarks,stefanocasazza/FrameworkBenchmarks,valyala/FrameworkBenchmarks,s-ludwig/FrameworkBenchmarks,waiteb3/FrameworkBenchmarks,jebbstewart/FrameworkBenchmarks,k-r-g/FrameworkBenchmarks,fabianmurariu/FrameworkBenchmarks,sagenschneider/FrameworkBenchmarks,zapov/FrameworkBenchmarks,zdanek/FrameworkBenchmarks,methane/FrameworkBenchmarks,sxend/FrameworkBenchmarks,jaguililla/FrameworkBenchmarks,kbrock/FrameworkBenchmarks,jebbstewart/FrameworkBenchmarks,jetty-project/FrameworkBenchmarks,Dith3r/FrameworkBenchmarks,kostya-sh/FrameworkBenchmarks,Synchro/FrameworkBenchmarks,zane-techempower/FrameworkBenchmarks,torhve/FrameworkBenchmarks,sxend/FrameworkBenchmarks,fabianmurariu/FrameworkBenchmarks,grob/FrameworkBenchmarks,F3Community/FrameworkBenchmarks,jeevatkm/FrameworkBenchmarks,Rydgel/FrameworkBenchmarks,kellabyte/FrameworkBenchmarks,actframework/FrameworkBenchmarks,donovanmuller/FrameworkBenchmarks,jeevatkm/FrameworkBenchmarks,marko-asplund/FrameworkBenchmarks,nbrady-techempower/FrameworkBenchmarks,nathana1/FrameworkBenchmarks,thousandsofthem/FrameworkBenchmarks,alubbe/FrameworkBenchmarks,actframework/FrameworkBenchmarks,PermeAgility/FrameworkBenchmarks,MTDdk/FrameworkBenchmarks,dmacd/FB-try1,Synchro/FrameworkBenchmarks,julienschmidt/FrameworkBenchmarks,zdanek/FrameworkBenchmarks,ashawnbandy-te-tfb/FrameworkBenchmarks,psfblair/FrameworkBenchmarks,zane-techempower/FrameworkBenchmarks,kellabyte/FrameworkBenchmarks,khellang/FrameworkBenchmarks,MTDdk/FrameworkBenchmarks,jamming/FrameworkBenchmarks,sgml/FrameworkBenchmarks,doom369/FrameworkBenchmarks,circlespainter/FrameworkBenchmarks,greenlaw110/FrameworkBenchmarks,grob/FrameworkBenchmarks,sanjoydesk/FrameworkBenchmarks,methane/FrameworkBenchmarks,doom369/FrameworkBenchmarks,denkab/FrameworkBenchmarks,hperadin/FrameworkBenchmarks,sanjoydesk/FrameworkBenchmarks,hperadin/FrameworkBenchmarks,sagenschneider/FrameworkBenchmarks,fabianmurariu/FrameworkBenchmarks,jamming/FrameworkBenchmarks,Rydgel/FrameworkBenchmarks,jetty-project/FrameworkBenchmarks,leafo/FrameworkBenchmarks,jaguililla/FrameworkBenchmarks,kellabyte/FrameworkBenchmarks,greenlaw110/FrameworkBenchmarks,kellabyte/FrameworkBenchmarks,doom369/FrameworkBenchmarks,jeevatkm/FrameworkBenchmarks,victorbriz/FrameworkBenchmarks,kbrock/FrameworkBenchmarks,youprofit/FrameworkBenchmarks,knewmanTE/FrameworkBenchmarks,mfirry/FrameworkBenchmarks,jeevatkm/FrameworkBenchmarks,jeevatkm/FrameworkBenchmarks,kostya-sh/FrameworkBenchmarks,fabianmurariu/FrameworkBenchmarks,mfirry/FrameworkBenchmarks,waiteb3/FrameworkBenchmarks,zdanek/FrameworkBenchmarks,PermeAgility/FrameworkBenchmarks,methane/FrameworkBenchmarks,Verber/FrameworkBenchmarks,jetty-project/FrameworkBenchmarks,julienschmidt/FrameworkBenchmarks,sgml/FrameworkBenchmarks,RockinRoel/FrameworkBenchmarks,ratpack/FrameworkBenchmarks,dmacd/FB-try1,jaguililla/FrameworkBenchmarks,joshk/FrameworkBenchmarks,jetty-project/FrameworkBenchmarks,jetty-project/FrameworkBenchmarks,kellabyte/FrameworkBenchmarks,zapov/FrameworkBenchmarks,lcp0578/FrameworkBenchmarks,nbrady-techempower/FrameworkBenchmarks,herloct/FrameworkBenchmarks,steveklabnik/FrameworkBenchmarks,jetty-project/FrameworkBenchmarks,kellabyte/FrameworkBenchmarks,nkasvosve/FrameworkBenchmarks,sanjoydesk/FrameworkBenchmarks,RockinRoel/FrameworkBenchmarks,sxend/FrameworkBenchmarks,actframework/FrameworkBenchmarks,mfirry/FrameworkBenchmarks,nbrady-techempower/FrameworkBenchmarks,marko-asplund/FrameworkBenchmarks,diablonhn/FrameworkBenchmarks,greenlaw110/FrameworkBenchmarks,sxend/FrameworkBenchmarks,markkolich/FrameworkBenchmarks,yunspace/FrameworkBenchmarks,ashawnbandy-te-tfb/FrameworkBenchmarks,fabianmurariu/FrameworkBenchmarks,denkab/FrameworkBenchmarks,sanjoydesk/FrameworkBenchmarks,doom369/FrameworkBenchmarks,Verber/FrameworkBenchmarks,sxend/FrameworkBenchmarks,marko-asplund/FrameworkBenchmarks,actframework/FrameworkBenchmarks,markkolich/FrameworkBenchmarks,circlespainter/FrameworkBenchmarks,saturday06/FrameworkBenchmarks,raziel057/FrameworkBenchmarks,raziel057/FrameworkBenchmarks,joshk/FrameworkBenchmarks,MTDdk/FrameworkBenchmarks,waiteb3/FrameworkBenchmarks,donovanmuller/FrameworkBenchmarks,valyala/FrameworkBenchmarks,martin-g/FrameworkBenchmarks,nbrady-techempower/FrameworkBenchmarks,saturday06/FrameworkBenchmarks,s-ludwig/FrameworkBenchmarks,steveklabnik/FrameworkBenchmarks,hamiltont/FrameworkBenchmarks,Verber/FrameworkBenchmarks,xitrum-framework/FrameworkBenchmarks,sgml/FrameworkBenchmarks,kbrock/FrameworkBenchmarks,nkasvosve/FrameworkBenchmarks,RockinRoel/FrameworkBenchmarks,sanjoydesk/FrameworkBenchmarks,actframework/FrameworkBenchmarks,zane-techempower/FrameworkBenchmarks,greg-hellings/FrameworkBenchmarks,diablonhn/FrameworkBenchmarks,Eyepea/FrameworkBenchmarks,youprofit/FrameworkBenchmarks,saturday06/FrameworkBenchmarks,psfblair/FrameworkBenchmarks,greenlaw110/FrameworkBenchmarks,khellang/FrameworkBenchmarks,jamming/FrameworkBenchmarks,zloster/FrameworkBenchmarks,zloster/FrameworkBenchmarks,thousandsofthem/FrameworkBenchmarks,steveklabnik/FrameworkBenchmarks,jeevatkm/FrameworkBenchmarks,s-ludwig/FrameworkBenchmarks,MTDdk/FrameworkBenchmarks,zhuochenKIDD/FrameworkBenchmarks,actframework/FrameworkBenchmarks,Jesterovskiy/FrameworkBenchmarks,zapov/FrameworkBenchmarks,doom369/FrameworkBenchmarks,steveklabnik/FrameworkBenchmarks,joshk/FrameworkBenchmarks,martin-g/FrameworkBenchmarks,steveklabnik/FrameworkBenchmarks,stefanocasazza/FrameworkBenchmarks,Synchro/FrameworkBenchmarks,jaguililla/FrameworkBenchmarks,knewmanTE/FrameworkBenchmarks,kbrock/FrameworkBenchmarks,zhuochenKIDD/FrameworkBenchmarks,testn/FrameworkBenchmarks,donovanmuller/FrameworkBenchmarks,greg-hellings/FrameworkBenchmarks,RockinRoel/FrameworkBenchmarks,denkab/FrameworkBenchmarks,methane/FrameworkBenchmarks,torhve/FrameworkBenchmarks,valyala/FrameworkBenchmarks,greg-hellings/FrameworkBenchmarks,zhuochenKIDD/FrameworkBenchmarks,youprofit/FrameworkBenchmarks,valyala/FrameworkBenchmarks,ashawnbandy-te-tfb/FrameworkBenchmarks,dmacd/FB-try1,leafo/FrameworkBenchmarks,hamiltont/FrameworkBenchmarks,victorbriz/FrameworkBenchmarks,nbrady-techempower/FrameworkBenchmarks,mfirry/FrameworkBenchmarks,MTDdk/FrameworkBenchmarks,nbrady-techempower/FrameworkBenchmarks,dmacd/FB-try1,fabianmurariu/FrameworkBenchmarks,Jesterovskiy/FrameworkBenchmarks,dmacd/FB-try1,jetty-project/FrameworkBenchmarks,leafo/FrameworkBenchmarks,lcp0578/FrameworkBenchmarks,valyala/FrameworkBenchmarks,zhuochenKIDD/FrameworkBenchmarks,fabianmurariu/FrameworkBenchmarks,thousandsofthem/FrameworkBenchmarks,ratpack/FrameworkBenchmarks,waiteb3/FrameworkBenchmarks,victorbriz/FrameworkBenchmarks,zloster/FrameworkBenchmarks,RockinRoel/FrameworkBenchmarks,raziel057/FrameworkBenchmarks,waiteb3/FrameworkBenchmarks,jamming/FrameworkBenchmarks,knewmanTE/FrameworkBenchmarks,alubbe/FrameworkBenchmarks,doom369/FrameworkBenchmarks,sanjoydesk/FrameworkBenchmarks,greg-hellings/FrameworkBenchmarks,greenlaw110/FrameworkBenchmarks,herloct/FrameworkBenchmarks,Jesterovskiy/FrameworkBenchmarks,sagenschneider/FrameworkBenchmarks,hamiltont/FrameworkBenchmarks,lcp0578/FrameworkBenchmarks,Synchro/FrameworkBenchmarks,MTDdk/FrameworkBenchmarks,waiteb3/FrameworkBenchmarks,hperadin/FrameworkBenchmarks,jebbstewart/FrameworkBenchmarks,Synchro/FrameworkBenchmarks,ashawnbandy-te-tfb/FrameworkBenchmarks,actframework/FrameworkBenchmarks,saturday06/FrameworkBenchmarks,thousandsofthem/FrameworkBenchmarks,lcp0578/FrameworkBenchmarks,greg-hellings/FrameworkBenchmarks,seem-sky/FrameworkBenchmarks,khellang/FrameworkBenchmarks,Dith3r/FrameworkBenchmarks,julienschmidt/FrameworkBenchmarks,lcp0578/FrameworkBenchmarks,martin-g/FrameworkBenchmarks,valyala/FrameworkBenchmarks,Dith3r/FrameworkBenchmarks,leafo/FrameworkBenchmarks,RockinRoel/FrameworkBenchmarks,Dith3r/FrameworkBenchmarks,sagenschneider/FrameworkBenchmarks,testn/FrameworkBenchmarks,k-r-g/FrameworkBenchmarks,denkab/FrameworkBenchmarks,actframework/FrameworkBenchmarks,mfirry/FrameworkBenchmarks,k-r-g/FrameworkBenchmarks,ashawnbandy-te-tfb/FrameworkBenchmarks,yunspace/FrameworkBenchmarks,fabianmurariu/FrameworkBenchmarks,marko-asplund/FrameworkBenchmarks,jebbstewart/FrameworkBenchmarks,seem-sky/FrameworkBenchmarks,Rayne/FrameworkBenchmarks,circlespainter/FrameworkBenchmarks,methane/FrameworkBenchmarks,joshk/FrameworkBenchmarks,raziel057/FrameworkBenchmarks,torhve/FrameworkBenchmarks,circlespainter/FrameworkBenchmarks,julienschmidt/FrameworkBenchmarks,Dith3r/FrameworkBenchmarks,k-r-g/FrameworkBenchmarks,jaguililla/FrameworkBenchmarks,PermeAgility/FrameworkBenchmarks,zane-techempower/FrameworkBenchmarks,zapov/FrameworkBenchmarks,Dith3r/FrameworkBenchmarks,dmacd/FB-try1,sxend/FrameworkBenchmarks,zdanek/FrameworkBenchmarks,greenlaw110/FrameworkBenchmarks,sgml/FrameworkBenchmarks,Jesterovskiy/FrameworkBenchmarks,Rydgel/FrameworkBenchmarks,Verber/FrameworkBenchmarks,zloster/FrameworkBenchmarks,kostya-sh/FrameworkBenchmarks,methane/FrameworkBenchmarks,yunspace/FrameworkBenchmarks,steveklabnik/FrameworkBenchmarks,Synchro/FrameworkBenchmarks,victorbriz/FrameworkBenchmarks,hamiltont/FrameworkBenchmarks,testn/FrameworkBenchmarks,joshk/FrameworkBenchmarks,Rayne/FrameworkBenchmarks,MTDdk/FrameworkBenchmarks,F3Community/FrameworkBenchmarks,Rayne/FrameworkBenchmarks,zane-techempower/FrameworkBenchmarks,zapov/FrameworkBenchmarks,stefanocasazza/FrameworkBenchmarks,herloct/FrameworkBenchmarks,herloct/FrameworkBenchmarks,MTDdk/FrameworkBenchmarks,circlespainter/FrameworkBenchmarks,F3Community/FrameworkBenchmarks,nbrady-techempower/FrameworkBenchmarks,zloster/FrameworkBenchmarks,jebbstewart/FrameworkBenchmarks,sagenschneider/FrameworkBenchmarks,jamming/FrameworkBenchmarks,alubbe/FrameworkBenchmarks,steveklabnik/FrameworkBenchmarks,sagenschneider/FrameworkBenchmarks,jaguililla/FrameworkBenchmarks,joshk/FrameworkBenchmarks,sagenschneider/FrameworkBenchmarks,greenlaw110/FrameworkBenchmarks,valyala/FrameworkBenchmarks,sxend/FrameworkBenchmarks,raziel057/FrameworkBenchmarks,sagenschneider/FrameworkBenchmarks,k-r-g/FrameworkBenchmarks,youprofit/FrameworkBenchmarks,Jesterovskiy/FrameworkBenchmarks,Dith3r/FrameworkBenchmarks,k-r-g/FrameworkBenchmarks,PermeAgility/FrameworkBenchmarks,torhve/FrameworkBenchmarks,Rayne/FrameworkBenchmarks,joshk/FrameworkBenchmarks,yunspace/FrameworkBenchmarks,jetty-project/FrameworkBenchmarks,sgml/FrameworkBenchmarks,lcp0578/FrameworkBenchmarks,greenlaw110/FrameworkBenchmarks,victorbriz/FrameworkBenchmarks,psfblair/FrameworkBenchmarks,nbrady-techempower/FrameworkBenchmarks,greenlaw110/FrameworkBenchmarks,yunspace/FrameworkBenchmarks,lcp0578/FrameworkBenchmarks,denkab/FrameworkBenchmarks,stefanocasazza/FrameworkBenchmarks,Jesterovskiy/FrameworkBenchmarks,Rayne/FrameworkBenchmarks,raziel057/FrameworkBenchmarks,jamming/FrameworkBenchmarks,jeevatkm/FrameworkBenchmarks,thousandsofthem/FrameworkBenchmarks,zdanek/FrameworkBenchmarks,donovanmuller/FrameworkBenchmarks,valyala/FrameworkBenchmarks,Synchro/FrameworkBenchmarks,hamiltont/FrameworkBenchmarks,F3Community/FrameworkBenchmarks,lcp0578/FrameworkBenchmarks,kostya-sh/FrameworkBenchmarks,hamiltont/FrameworkBenchmarks,jebbstewart/FrameworkBenchmarks,yunspace/FrameworkBenchmarks,greenlaw110/FrameworkBenchmarks,youprofit/FrameworkBenchmarks,leafo/FrameworkBenchmarks,martin-g/FrameworkBenchmarks,zapov/FrameworkBenchmarks,Jesterovskiy/FrameworkBenchmarks,leafo/FrameworkBenchmarks,diablonhn/FrameworkBenchmarks,jamming/FrameworkBenchmarks,grob/FrameworkBenchmarks,martin-g/FrameworkBenchmarks,Dith3r/FrameworkBenchmarks,s-ludwig/FrameworkBenchmarks,hperadin/FrameworkBenchmarks,s-ludwig/FrameworkBenchmarks |
import subprocess
import sys
import setup_util
import os
def start(args):
- subprocess.Popen("gunicorn hello:app -b 0.0.0.0:8080 -w " + str((args.max_threads * 2)) + " --log-level=critical", shell=True, cwd="wsgi")
+ subprocess.Popen('gunicorn hello:app --worker-class="egg:meinheld#gunicorn_worker" -b 0.0.0.0:8080 -w '
+ + str((args.max_threads * 2)) + " --log-level=critical", shell=True, cwd="wsgi")
return 0
def stop():
p = subprocess.Popen(['ps', 'aux'], stdout=subprocess.PIPE)
out, err = p.communicate()
for line in out.splitlines():
if 'gunicorn' in line:
try:
pid = int(line.split(None, 2)[1])
os.kill(pid, 9)
except OSError:
pass
return 0
+ | Use meinheld worker (same as other Python Frameworks) | ## Code Before:
import subprocess
import sys
import setup_util
import os
def start(args):
subprocess.Popen("gunicorn hello:app -b 0.0.0.0:8080 -w " + str((args.max_threads * 2)) + " --log-level=critical", shell=True, cwd="wsgi")
return 0
def stop():
p = subprocess.Popen(['ps', 'aux'], stdout=subprocess.PIPE)
out, err = p.communicate()
for line in out.splitlines():
if 'gunicorn' in line:
try:
pid = int(line.split(None, 2)[1])
os.kill(pid, 9)
except OSError:
pass
return 0
## Instruction:
Use meinheld worker (same as other Python Frameworks)
## Code After:
import subprocess
import sys
import setup_util
import os
def start(args):
subprocess.Popen('gunicorn hello:app --worker-class="egg:meinheld#gunicorn_worker" -b 0.0.0.0:8080 -w '
+ str((args.max_threads * 2)) + " --log-level=critical", shell=True, cwd="wsgi")
return 0
def stop():
p = subprocess.Popen(['ps', 'aux'], stdout=subprocess.PIPE)
out, err = p.communicate()
for line in out.splitlines():
if 'gunicorn' in line:
try:
pid = int(line.split(None, 2)[1])
os.kill(pid, 9)
except OSError:
pass
return 0
| # ... existing code ...
def start(args):
subprocess.Popen('gunicorn hello:app --worker-class="egg:meinheld#gunicorn_worker" -b 0.0.0.0:8080 -w '
+ str((args.max_threads * 2)) + " --log-level=critical", shell=True, cwd="wsgi")
return 0
def stop():
# ... rest of the code ... |
2046d82addab9ec83dbb85a2d08c727a52065d8b | deckglue/models.py | deckglue/models.py | from django.db import models
# Create your models here.
| from django.contrib.auth.models import Permission
from django.db.models.signals import post_save, pre_delete
from django.dispatch import receiver
from cardbox.card_model import Card
from cardbox.deck_model import Deck
from guardian.shortcuts import assign_perm, get_users_with_perms
from guardian.models import UserObjectPermission
from memorize.models import Practice
from django.contrib.auth.models import User
@receiver(post_save, sender=UserObjectPermission)
def create_practice_objects_for_new_viewers(sender, **kwargs):
if kwargs['instance'].permission_id == Permission.objects.get(codename="view_deck").id:
for card in Card.objects.filter(deck=kwargs['instance'].object_pk):
Practice(item=card, user=User.objects.get(id = kwargs['instance'].user_id)).save()
@receiver(pre_delete, sender=UserObjectPermission)
def delete_practice_objects_for_removed_viewers(sender, **kwargs):
if kwargs['instance'].permission_id == Permission.objects.get(codename="view_deck").id:
for card in Card.objects.filter(deck=kwargs['instance'].object_pk):
Practice.objects.get(object_id=card.ID, user=User.objects.get(id = kwargs['instance'].user_id)).delete()
@receiver(post_save, sender=Card)
def create_practice_objects_for_new_card(sender,update_fields, **kwargs):
"""Creates practice objects for all users with permission to view the card.
"""
perm_users = get_users_with_perms(kwargs['instance'].deck)
for user in perm_users:
practice = Practice(item = kwargs['instance'], user = user)
if Practice.objects.filter(object_id = kwargs['instance'].ID, user=user).count() == 0:
practice.save()
@receiver(pre_delete, sender=Card)
def delete_practice_objects(sender, **kwargs):
"""Deletes all practice objects for a card once it is deleted.
"""
Practice.objects.filter(object_id = kwargs['instance'].ID).delete()
| Add signal hooks to create practice objects | Add signal hooks to create practice objects
| Python | mit | DummyDivision/Tsune,DummyDivision/Tsune,DummyDivision/Tsune | - from django.db import models
+ from django.contrib.auth.models import Permission
+ from django.db.models.signals import post_save, pre_delete
+ from django.dispatch import receiver
+ from cardbox.card_model import Card
+ from cardbox.deck_model import Deck
+ from guardian.shortcuts import assign_perm, get_users_with_perms
+ from guardian.models import UserObjectPermission
+ from memorize.models import Practice
+ from django.contrib.auth.models import User
- # Create your models here.
+ @receiver(post_save, sender=UserObjectPermission)
+ def create_practice_objects_for_new_viewers(sender, **kwargs):
+ if kwargs['instance'].permission_id == Permission.objects.get(codename="view_deck").id:
+ for card in Card.objects.filter(deck=kwargs['instance'].object_pk):
+ Practice(item=card, user=User.objects.get(id = kwargs['instance'].user_id)).save()
+ @receiver(pre_delete, sender=UserObjectPermission)
+ def delete_practice_objects_for_removed_viewers(sender, **kwargs):
+ if kwargs['instance'].permission_id == Permission.objects.get(codename="view_deck").id:
+ for card in Card.objects.filter(deck=kwargs['instance'].object_pk):
+ Practice.objects.get(object_id=card.ID, user=User.objects.get(id = kwargs['instance'].user_id)).delete()
+
+ @receiver(post_save, sender=Card)
+ def create_practice_objects_for_new_card(sender,update_fields, **kwargs):
+ """Creates practice objects for all users with permission to view the card.
+
+ """
+ perm_users = get_users_with_perms(kwargs['instance'].deck)
+ for user in perm_users:
+ practice = Practice(item = kwargs['instance'], user = user)
+ if Practice.objects.filter(object_id = kwargs['instance'].ID, user=user).count() == 0:
+ practice.save()
+
+
+
+ @receiver(pre_delete, sender=Card)
+ def delete_practice_objects(sender, **kwargs):
+ """Deletes all practice objects for a card once it is deleted.
+
+ """
+ Practice.objects.filter(object_id = kwargs['instance'].ID).delete()
+
+ | Add signal hooks to create practice objects | ## Code Before:
from django.db import models
# Create your models here.
## Instruction:
Add signal hooks to create practice objects
## Code After:
from django.contrib.auth.models import Permission
from django.db.models.signals import post_save, pre_delete
from django.dispatch import receiver
from cardbox.card_model import Card
from cardbox.deck_model import Deck
from guardian.shortcuts import assign_perm, get_users_with_perms
from guardian.models import UserObjectPermission
from memorize.models import Practice
from django.contrib.auth.models import User
@receiver(post_save, sender=UserObjectPermission)
def create_practice_objects_for_new_viewers(sender, **kwargs):
if kwargs['instance'].permission_id == Permission.objects.get(codename="view_deck").id:
for card in Card.objects.filter(deck=kwargs['instance'].object_pk):
Practice(item=card, user=User.objects.get(id = kwargs['instance'].user_id)).save()
@receiver(pre_delete, sender=UserObjectPermission)
def delete_practice_objects_for_removed_viewers(sender, **kwargs):
if kwargs['instance'].permission_id == Permission.objects.get(codename="view_deck").id:
for card in Card.objects.filter(deck=kwargs['instance'].object_pk):
Practice.objects.get(object_id=card.ID, user=User.objects.get(id = kwargs['instance'].user_id)).delete()
@receiver(post_save, sender=Card)
def create_practice_objects_for_new_card(sender,update_fields, **kwargs):
"""Creates practice objects for all users with permission to view the card.
"""
perm_users = get_users_with_perms(kwargs['instance'].deck)
for user in perm_users:
practice = Practice(item = kwargs['instance'], user = user)
if Practice.objects.filter(object_id = kwargs['instance'].ID, user=user).count() == 0:
practice.save()
@receiver(pre_delete, sender=Card)
def delete_practice_objects(sender, **kwargs):
"""Deletes all practice objects for a card once it is deleted.
"""
Practice.objects.filter(object_id = kwargs['instance'].ID).delete()
| // ... existing code ...
from django.contrib.auth.models import Permission
from django.db.models.signals import post_save, pre_delete
from django.dispatch import receiver
from cardbox.card_model import Card
from cardbox.deck_model import Deck
from guardian.shortcuts import assign_perm, get_users_with_perms
from guardian.models import UserObjectPermission
from memorize.models import Practice
from django.contrib.auth.models import User
@receiver(post_save, sender=UserObjectPermission)
def create_practice_objects_for_new_viewers(sender, **kwargs):
if kwargs['instance'].permission_id == Permission.objects.get(codename="view_deck").id:
for card in Card.objects.filter(deck=kwargs['instance'].object_pk):
Practice(item=card, user=User.objects.get(id = kwargs['instance'].user_id)).save()
@receiver(pre_delete, sender=UserObjectPermission)
def delete_practice_objects_for_removed_viewers(sender, **kwargs):
if kwargs['instance'].permission_id == Permission.objects.get(codename="view_deck").id:
for card in Card.objects.filter(deck=kwargs['instance'].object_pk):
Practice.objects.get(object_id=card.ID, user=User.objects.get(id = kwargs['instance'].user_id)).delete()
@receiver(post_save, sender=Card)
def create_practice_objects_for_new_card(sender,update_fields, **kwargs):
"""Creates practice objects for all users with permission to view the card.
"""
perm_users = get_users_with_perms(kwargs['instance'].deck)
for user in perm_users:
practice = Practice(item = kwargs['instance'], user = user)
if Practice.objects.filter(object_id = kwargs['instance'].ID, user=user).count() == 0:
practice.save()
@receiver(pre_delete, sender=Card)
def delete_practice_objects(sender, **kwargs):
"""Deletes all practice objects for a card once it is deleted.
"""
Practice.objects.filter(object_id = kwargs['instance'].ID).delete()
// ... rest of the code ... |
db13de154fa44f3ef0bf1e365d2ee0d7a6951700 | cellcounter/accounts/urls.py | cellcounter/accounts/urls.py | from django.conf.urls import patterns, url
from cellcounter.accounts import views
urlpatterns = patterns('',
url('^new/$', views.RegistrationView.as_view(), name='register'),
url('^(?P<pk>[0-9]+)/$', views.UserDetailView.as_view(), name='user-detail'),
url('^(?P<pk>[0-9]+)/delete/$', views.UserDeleteView.as_view(), name='user-delete'),
url('^(?P<pk>[0-9]+)/edit/$', views.UserUpdateView.as_view(), name='user-update'),
url('^password/reset/$', views.PasswordResetView.as_view(),
name='password-reset'),
url('^password/reset/confirm/(?P<uidb64>[0-9A-Za-z_\-]+)/(?P<token>[\d\w\-]+)/$',
views.PasswordResetConfirmView.as_view(),
name='password-reset-confirm'),
url('^password/change/$', views.PasswordChangeView.as_view(), name='change-password'),
) | from django.conf.urls import patterns, url
from cellcounter.accounts import views
urlpatterns = patterns('',
url('^new/$', views.RegistrationView.as_view(), name='register'),
url('^(?P<pk>[0-9]+)/$', views.UserDetailView.as_view(), name='user-detail'),
url('^(?P<pk>[0-9]+)/delete/$', views.UserDeleteView.as_view(), name='user-delete'),
url('^(?P<pk>[0-9]+)/edit/$', views.UserUpdateView.as_view(), name='user-update'),
url('^password/reset/$', views.PasswordResetView.as_view(),
name='password-reset'),
url('^password/reset/confirm/(?P<uidb64>[0-9A-Za-z_\-]+)/(?P<token>[0-9A-Za-z]{1,13}-[0-9A-Za-z]{1,20})/$',
views.PasswordResetConfirmView.as_view(),
name='password-reset-confirm'),
url('^password/change/$', views.PasswordChangeView.as_view(), name='change-password'),
) | Use URL regex as per main Django project | Use URL regex as per main Django project
| Python | mit | cellcounter/cellcounter,haematologic/cellcounter,haematologic/cellcounter,cellcounter/cellcounter,cellcounter/cellcounter,cellcounter/cellcounter,haematologic/cellcounter | from django.conf.urls import patterns, url
from cellcounter.accounts import views
urlpatterns = patterns('',
url('^new/$', views.RegistrationView.as_view(), name='register'),
url('^(?P<pk>[0-9]+)/$', views.UserDetailView.as_view(), name='user-detail'),
url('^(?P<pk>[0-9]+)/delete/$', views.UserDeleteView.as_view(), name='user-delete'),
url('^(?P<pk>[0-9]+)/edit/$', views.UserUpdateView.as_view(), name='user-update'),
url('^password/reset/$', views.PasswordResetView.as_view(),
name='password-reset'),
- url('^password/reset/confirm/(?P<uidb64>[0-9A-Za-z_\-]+)/(?P<token>[\d\w\-]+)/$',
+ url('^password/reset/confirm/(?P<uidb64>[0-9A-Za-z_\-]+)/(?P<token>[0-9A-Za-z]{1,13}-[0-9A-Za-z]{1,20})/$',
views.PasswordResetConfirmView.as_view(),
name='password-reset-confirm'),
url('^password/change/$', views.PasswordChangeView.as_view(), name='change-password'),
) | Use URL regex as per main Django project | ## Code Before:
from django.conf.urls import patterns, url
from cellcounter.accounts import views
urlpatterns = patterns('',
url('^new/$', views.RegistrationView.as_view(), name='register'),
url('^(?P<pk>[0-9]+)/$', views.UserDetailView.as_view(), name='user-detail'),
url('^(?P<pk>[0-9]+)/delete/$', views.UserDeleteView.as_view(), name='user-delete'),
url('^(?P<pk>[0-9]+)/edit/$', views.UserUpdateView.as_view(), name='user-update'),
url('^password/reset/$', views.PasswordResetView.as_view(),
name='password-reset'),
url('^password/reset/confirm/(?P<uidb64>[0-9A-Za-z_\-]+)/(?P<token>[\d\w\-]+)/$',
views.PasswordResetConfirmView.as_view(),
name='password-reset-confirm'),
url('^password/change/$', views.PasswordChangeView.as_view(), name='change-password'),
)
## Instruction:
Use URL regex as per main Django project
## Code After:
from django.conf.urls import patterns, url
from cellcounter.accounts import views
urlpatterns = patterns('',
url('^new/$', views.RegistrationView.as_view(), name='register'),
url('^(?P<pk>[0-9]+)/$', views.UserDetailView.as_view(), name='user-detail'),
url('^(?P<pk>[0-9]+)/delete/$', views.UserDeleteView.as_view(), name='user-delete'),
url('^(?P<pk>[0-9]+)/edit/$', views.UserUpdateView.as_view(), name='user-update'),
url('^password/reset/$', views.PasswordResetView.as_view(),
name='password-reset'),
url('^password/reset/confirm/(?P<uidb64>[0-9A-Za-z_\-]+)/(?P<token>[0-9A-Za-z]{1,13}-[0-9A-Za-z]{1,20})/$',
views.PasswordResetConfirmView.as_view(),
name='password-reset-confirm'),
url('^password/change/$', views.PasswordChangeView.as_view(), name='change-password'),
) | # ... existing code ...
url('^password/reset/$', views.PasswordResetView.as_view(),
name='password-reset'),
url('^password/reset/confirm/(?P<uidb64>[0-9A-Za-z_\-]+)/(?P<token>[0-9A-Za-z]{1,13}-[0-9A-Za-z]{1,20})/$',
views.PasswordResetConfirmView.as_view(),
name='password-reset-confirm'),
# ... rest of the code ... |
079e7cbbd59266e1dc8b161989c90202caa4c5a8 | flaskbb/utils/views.py | flaskbb/utils/views.py | from flask import render_template
from flask.views import View
class RenderableView(View):
def __init__(self, template, view):
self.template = template
self.view = view
def dispatch_request(self, *args, **kwargs):
view_model = self.view(*args, **kwargs)
return render_template(self.template, **view_model)
| from flaskbb.utils.helpers import render_template
from flask.views import View
class RenderableView(View):
def __init__(self, template, view):
self.template = template
self.view = view
def dispatch_request(self, *args, **kwargs):
view_model = self.view(*args, **kwargs)
return render_template(self.template, **view_model)
| Use local render_template than Flask's native | Use local render_template than Flask's native
TODO: Provide a renderer argument at instantation?
| Python | bsd-3-clause | realityone/flaskbb,realityone/flaskbb,dromanow/flaskbb,dromanow/flaskbb,realityone/flaskbb,dromanow/flaskbb | - from flask import render_template
+ from flaskbb.utils.helpers import render_template
from flask.views import View
class RenderableView(View):
def __init__(self, template, view):
self.template = template
self.view = view
def dispatch_request(self, *args, **kwargs):
view_model = self.view(*args, **kwargs)
return render_template(self.template, **view_model)
| Use local render_template than Flask's native | ## Code Before:
from flask import render_template
from flask.views import View
class RenderableView(View):
def __init__(self, template, view):
self.template = template
self.view = view
def dispatch_request(self, *args, **kwargs):
view_model = self.view(*args, **kwargs)
return render_template(self.template, **view_model)
## Instruction:
Use local render_template than Flask's native
## Code After:
from flaskbb.utils.helpers import render_template
from flask.views import View
class RenderableView(View):
def __init__(self, template, view):
self.template = template
self.view = view
def dispatch_request(self, *args, **kwargs):
view_model = self.view(*args, **kwargs)
return render_template(self.template, **view_model)
| ...
from flaskbb.utils.helpers import render_template
from flask.views import View
... |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.