id int32 0 252k | repo stringlengths 7 55 | path stringlengths 4 127 | func_name stringlengths 1 88 | original_string stringlengths 75 19.8k | language stringclasses 1
value | code stringlengths 75 19.8k | code_tokens list | docstring stringlengths 3 17.3k | docstring_tokens list | sha stringlengths 40 40 | url stringlengths 87 242 |
|---|---|---|---|---|---|---|---|---|---|---|---|
242,200 | briancappello/flask-sqlalchemy-bundle | flask_sqlalchemy_bundle/meta/model_meta_factory.py | ModelMetaFactory._get_model_meta_options | def _get_model_meta_options(self) -> List[MetaOption]:
""""
Define fields allowed in the Meta class on end-user models, and the
behavior of each.
Custom ModelMetaOptions classes should override this method to customize
the options supported on class Meta of end-user models.
"""
# we can't use current_app to determine if we're under test, because it
# doesn't exist yet
testing_options = ([] if os.getenv('FLASK_ENV', False) != TEST
else [_TestingMetaOption()])
# when options require another option, its dependent must be listed.
# options in this list are not order-dependent, except where noted.
# all ColumnMetaOptions subclasses require PolymorphicMetaOption
return testing_options + [
AbstractMetaOption(), # required; must be first
LazyMappedMetaOption(),
RelationshipsMetaOption(), # requires lazy_mapped
TableMetaOption(),
MaterializedViewForMetaOption(),
PolymorphicMetaOption(), # must be first of all polymorphic options
PolymorphicOnColumnMetaOption(),
PolymorphicIdentityMetaOption(),
PolymorphicBaseTablenameMetaOption(),
PolymorphicJoinedPkColumnMetaOption(), # requires _BaseTablename
# must be after PolymorphicJoinedPkColumnMetaOption
PrimaryKeyColumnMetaOption(),
CreatedAtColumnMetaOption(),
UpdatedAtColumnMetaOption(),
] | python | def _get_model_meta_options(self) -> List[MetaOption]:
""""
Define fields allowed in the Meta class on end-user models, and the
behavior of each.
Custom ModelMetaOptions classes should override this method to customize
the options supported on class Meta of end-user models.
"""
# we can't use current_app to determine if we're under test, because it
# doesn't exist yet
testing_options = ([] if os.getenv('FLASK_ENV', False) != TEST
else [_TestingMetaOption()])
# when options require another option, its dependent must be listed.
# options in this list are not order-dependent, except where noted.
# all ColumnMetaOptions subclasses require PolymorphicMetaOption
return testing_options + [
AbstractMetaOption(), # required; must be first
LazyMappedMetaOption(),
RelationshipsMetaOption(), # requires lazy_mapped
TableMetaOption(),
MaterializedViewForMetaOption(),
PolymorphicMetaOption(), # must be first of all polymorphic options
PolymorphicOnColumnMetaOption(),
PolymorphicIdentityMetaOption(),
PolymorphicBaseTablenameMetaOption(),
PolymorphicJoinedPkColumnMetaOption(), # requires _BaseTablename
# must be after PolymorphicJoinedPkColumnMetaOption
PrimaryKeyColumnMetaOption(),
CreatedAtColumnMetaOption(),
UpdatedAtColumnMetaOption(),
] | [
"def",
"_get_model_meta_options",
"(",
"self",
")",
"->",
"List",
"[",
"MetaOption",
"]",
":",
"# we can't use current_app to determine if we're under test, because it",
"# doesn't exist yet",
"testing_options",
"=",
"(",
"[",
"]",
"if",
"os",
".",
"getenv",
"(",
"'FLAS... | Define fields allowed in the Meta class on end-user models, and the
behavior of each.
Custom ModelMetaOptions classes should override this method to customize
the options supported on class Meta of end-user models. | [
"Define",
"fields",
"allowed",
"in",
"the",
"Meta",
"class",
"on",
"end",
"-",
"user",
"models",
"and",
"the",
"behavior",
"of",
"each",
"."
] | 8150896787907ef0001839b5a6ef303edccb9b6c | https://github.com/briancappello/flask-sqlalchemy-bundle/blob/8150896787907ef0001839b5a6ef303edccb9b6c/flask_sqlalchemy_bundle/meta/model_meta_factory.py#L30-L63 |
242,201 | rjw57/throw | throw/minus/minus.py | CreateGallery | def CreateGallery():
"""Creates a Gallery on the server. Returns a Gallery object with the
editor_id and reader_id.
"""
url = 'http://min.us/api/CreateGallery'
response = _dopost(url)
_editor_id = response["editor_id"]
_reader_id = response["reader_id"]
return Gallery(_reader_id, editor_id=_editor_id) | python | def CreateGallery():
"""Creates a Gallery on the server. Returns a Gallery object with the
editor_id and reader_id.
"""
url = 'http://min.us/api/CreateGallery'
response = _dopost(url)
_editor_id = response["editor_id"]
_reader_id = response["reader_id"]
return Gallery(_reader_id, editor_id=_editor_id) | [
"def",
"CreateGallery",
"(",
")",
":",
"url",
"=",
"'http://min.us/api/CreateGallery'",
"response",
"=",
"_dopost",
"(",
"url",
")",
"_editor_id",
"=",
"response",
"[",
"\"editor_id\"",
"]",
"_reader_id",
"=",
"response",
"[",
"\"reader_id\"",
"]",
"return",
"Ga... | Creates a Gallery on the server. Returns a Gallery object with the
editor_id and reader_id. | [
"Creates",
"a",
"Gallery",
"on",
"the",
"server",
".",
"Returns",
"a",
"Gallery",
"object",
"with",
"the",
"editor_id",
"and",
"reader_id",
"."
] | 74a7116362ba5b45635ab247472b25cfbdece4ee | https://github.com/rjw57/throw/blob/74a7116362ba5b45635ab247472b25cfbdece4ee/throw/minus/minus.py#L102-L114 |
242,202 | renzon/gaepermission | gaepermission/facade.py | logged_user | def logged_user(request):
"""
Returns a command that retrieves the current logged user based on secure cookie
If there is no logged user, the result from command is None
"""
dct = cookie_facade.retrive_cookie_data(request, USER_COOKIE_NAME).execute().result
if dct is None:
return Command()
return NodeSearch(dct['id']) | python | def logged_user(request):
"""
Returns a command that retrieves the current logged user based on secure cookie
If there is no logged user, the result from command is None
"""
dct = cookie_facade.retrive_cookie_data(request, USER_COOKIE_NAME).execute().result
if dct is None:
return Command()
return NodeSearch(dct['id']) | [
"def",
"logged_user",
"(",
"request",
")",
":",
"dct",
"=",
"cookie_facade",
".",
"retrive_cookie_data",
"(",
"request",
",",
"USER_COOKIE_NAME",
")",
".",
"execute",
"(",
")",
".",
"result",
"if",
"dct",
"is",
"None",
":",
"return",
"Command",
"(",
")",
... | Returns a command that retrieves the current logged user based on secure cookie
If there is no logged user, the result from command is None | [
"Returns",
"a",
"command",
"that",
"retrieves",
"the",
"current",
"logged",
"user",
"based",
"on",
"secure",
"cookie",
"If",
"there",
"is",
"no",
"logged",
"user",
"the",
"result",
"from",
"command",
"is",
"None"
] | 1a3534a7ef150ba31fa8df3bc8445557cab3d79d | https://github.com/renzon/gaepermission/blob/1a3534a7ef150ba31fa8df3bc8445557cab3d79d/gaepermission/facade.py#L57-L65 |
242,203 | renzon/gaepermission | gaepermission/facade.py | find_users_by_email_starting_with | def find_users_by_email_starting_with(email_prefix=None, cursor=None, page_size=30):
"""
Returns a command that retrieves users by its email_prefix, ordered by email.
It returns a max number of users defined by page_size arg. Next result can be retrieved using cursor, in
a next call. It is provided in cursor attribute from command.
"""
email_prefix = email_prefix or ''
return ModelSearchCommand(MainUser.query_email_starts_with(email_prefix),
page_size, cursor, cache_begin=None) | python | def find_users_by_email_starting_with(email_prefix=None, cursor=None, page_size=30):
"""
Returns a command that retrieves users by its email_prefix, ordered by email.
It returns a max number of users defined by page_size arg. Next result can be retrieved using cursor, in
a next call. It is provided in cursor attribute from command.
"""
email_prefix = email_prefix or ''
return ModelSearchCommand(MainUser.query_email_starts_with(email_prefix),
page_size, cursor, cache_begin=None) | [
"def",
"find_users_by_email_starting_with",
"(",
"email_prefix",
"=",
"None",
",",
"cursor",
"=",
"None",
",",
"page_size",
"=",
"30",
")",
":",
"email_prefix",
"=",
"email_prefix",
"or",
"''",
"return",
"ModelSearchCommand",
"(",
"MainUser",
".",
"query_email_sta... | Returns a command that retrieves users by its email_prefix, ordered by email.
It returns a max number of users defined by page_size arg. Next result can be retrieved using cursor, in
a next call. It is provided in cursor attribute from command. | [
"Returns",
"a",
"command",
"that",
"retrieves",
"users",
"by",
"its",
"email_prefix",
"ordered",
"by",
"email",
".",
"It",
"returns",
"a",
"max",
"number",
"of",
"users",
"defined",
"by",
"page_size",
"arg",
".",
"Next",
"result",
"can",
"be",
"retrieved",
... | 1a3534a7ef150ba31fa8df3bc8445557cab3d79d | https://github.com/renzon/gaepermission/blob/1a3534a7ef150ba31fa8df3bc8445557cab3d79d/gaepermission/facade.py#L108-L117 |
242,204 | renzon/gaepermission | gaepermission/facade.py | find_users_by_email_and_group | def find_users_by_email_and_group(email_prefix=None, group=None, cursor=None, page_size=30):
"""
Returns a command that retrieves users by its email_prefix, ordered by email and by Group.
If Group is None, only users without any group are going to be searched
It returns a max number of users defined by page_size arg. Next result can be retrieved using cursor, in
a next call. It is provided in cursor attribute from command.
"""
email_prefix = email_prefix or ''
return ModelSearchCommand(MainUser.query_email_and_group(email_prefix, group),
page_size, cursor, cache_begin=None) | python | def find_users_by_email_and_group(email_prefix=None, group=None, cursor=None, page_size=30):
"""
Returns a command that retrieves users by its email_prefix, ordered by email and by Group.
If Group is None, only users without any group are going to be searched
It returns a max number of users defined by page_size arg. Next result can be retrieved using cursor, in
a next call. It is provided in cursor attribute from command.
"""
email_prefix = email_prefix or ''
return ModelSearchCommand(MainUser.query_email_and_group(email_prefix, group),
page_size, cursor, cache_begin=None) | [
"def",
"find_users_by_email_and_group",
"(",
"email_prefix",
"=",
"None",
",",
"group",
"=",
"None",
",",
"cursor",
"=",
"None",
",",
"page_size",
"=",
"30",
")",
":",
"email_prefix",
"=",
"email_prefix",
"or",
"''",
"return",
"ModelSearchCommand",
"(",
"MainU... | Returns a command that retrieves users by its email_prefix, ordered by email and by Group.
If Group is None, only users without any group are going to be searched
It returns a max number of users defined by page_size arg. Next result can be retrieved using cursor, in
a next call. It is provided in cursor attribute from command. | [
"Returns",
"a",
"command",
"that",
"retrieves",
"users",
"by",
"its",
"email_prefix",
"ordered",
"by",
"email",
"and",
"by",
"Group",
".",
"If",
"Group",
"is",
"None",
"only",
"users",
"without",
"any",
"group",
"are",
"going",
"to",
"be",
"searched",
"It"... | 1a3534a7ef150ba31fa8df3bc8445557cab3d79d | https://github.com/renzon/gaepermission/blob/1a3534a7ef150ba31fa8df3bc8445557cab3d79d/gaepermission/facade.py#L120-L130 |
242,205 | rehandalal/buchner | buchner/project-template/PROJECTMODULE/errors.py | json_error | def json_error(code, message):
"""Returns a JSON-ified error object"""
# Message can be an unserializable object.
message = repr(message)
return jsonify(dict(request=request.path, message=message)), code | python | def json_error(code, message):
"""Returns a JSON-ified error object"""
# Message can be an unserializable object.
message = repr(message)
return jsonify(dict(request=request.path, message=message)), code | [
"def",
"json_error",
"(",
"code",
",",
"message",
")",
":",
"# Message can be an unserializable object.",
"message",
"=",
"repr",
"(",
"message",
")",
"return",
"jsonify",
"(",
"dict",
"(",
"request",
"=",
"request",
".",
"path",
",",
"message",
"=",
"message"... | Returns a JSON-ified error object | [
"Returns",
"a",
"JSON",
"-",
"ified",
"error",
"object"
] | dc22a61c493b9d4a74d76e8b42a319aa13e385f3 | https://github.com/rehandalal/buchner/blob/dc22a61c493b9d4a74d76e8b42a319aa13e385f3/buchner/project-template/PROJECTMODULE/errors.py#L12-L16 |
242,206 | rehandalal/buchner | buchner/project-template/PROJECTMODULE/errors.py | error | def error(code, message, template):
"""A generic error handler"""
if json_requested():
return json_error(code, message)
else:
return render_template(template, message=message), code | python | def error(code, message, template):
"""A generic error handler"""
if json_requested():
return json_error(code, message)
else:
return render_template(template, message=message), code | [
"def",
"error",
"(",
"code",
",",
"message",
",",
"template",
")",
":",
"if",
"json_requested",
"(",
")",
":",
"return",
"json_error",
"(",
"code",
",",
"message",
")",
"else",
":",
"return",
"render_template",
"(",
"template",
",",
"message",
"=",
"mess... | A generic error handler | [
"A",
"generic",
"error",
"handler"
] | dc22a61c493b9d4a74d76e8b42a319aa13e385f3 | https://github.com/rehandalal/buchner/blob/dc22a61c493b9d4a74d76e8b42a319aa13e385f3/buchner/project-template/PROJECTMODULE/errors.py#L19-L24 |
242,207 | MacHu-GWU/angora-project | angora/text/strtemplate.py | StrTemplate.straight_line_show | def straight_line_show(title, length=100, linestyle="=", pad=0):
"""Print a formatted straight line.
"""
print(StrTemplate.straight_line(
title=title, length=length, linestyle=linestyle, pad=pad)) | python | def straight_line_show(title, length=100, linestyle="=", pad=0):
"""Print a formatted straight line.
"""
print(StrTemplate.straight_line(
title=title, length=length, linestyle=linestyle, pad=pad)) | [
"def",
"straight_line_show",
"(",
"title",
",",
"length",
"=",
"100",
",",
"linestyle",
"=",
"\"=\"",
",",
"pad",
"=",
"0",
")",
":",
"print",
"(",
"StrTemplate",
".",
"straight_line",
"(",
"title",
"=",
"title",
",",
"length",
"=",
"length",
",",
"lin... | Print a formatted straight line. | [
"Print",
"a",
"formatted",
"straight",
"line",
"."
] | 689a60da51cd88680ddbe26e28dbe81e6b01d275 | https://github.com/MacHu-GWU/angora-project/blob/689a60da51cd88680ddbe26e28dbe81e6b01d275/angora/text/strtemplate.py#L42-L46 |
242,208 | MacHu-GWU/angora-project | angora/text/strtemplate.py | StrTemplate.indented_show | def indented_show(text, howmany=1):
"""Print a formatted indented text.
"""
print(StrTemplate.pad_indent(text=text, howmany=howmany)) | python | def indented_show(text, howmany=1):
"""Print a formatted indented text.
"""
print(StrTemplate.pad_indent(text=text, howmany=howmany)) | [
"def",
"indented_show",
"(",
"text",
",",
"howmany",
"=",
"1",
")",
":",
"print",
"(",
"StrTemplate",
".",
"pad_indent",
"(",
"text",
"=",
"text",
",",
"howmany",
"=",
"howmany",
")",
")"
] | Print a formatted indented text. | [
"Print",
"a",
"formatted",
"indented",
"text",
"."
] | 689a60da51cd88680ddbe26e28dbe81e6b01d275 | https://github.com/MacHu-GWU/angora-project/blob/689a60da51cd88680ddbe26e28dbe81e6b01d275/angora/text/strtemplate.py#L60-L63 |
242,209 | MacHu-GWU/angora-project | angora/text/strtemplate.py | StrTemplate.box_show | def box_show(text, width=100, height=3, corner="+", horizontal="-", vertical="|"):
"""Print a formatted ascii text box.
"""
print(StrTemplate.box(text=text, width=width, height=height,
corner=corner, horizontal=horizontal, vertical=vertical)) | python | def box_show(text, width=100, height=3, corner="+", horizontal="-", vertical="|"):
"""Print a formatted ascii text box.
"""
print(StrTemplate.box(text=text, width=width, height=height,
corner=corner, horizontal=horizontal, vertical=vertical)) | [
"def",
"box_show",
"(",
"text",
",",
"width",
"=",
"100",
",",
"height",
"=",
"3",
",",
"corner",
"=",
"\"+\"",
",",
"horizontal",
"=",
"\"-\"",
",",
"vertical",
"=",
"\"|\"",
")",
":",
"print",
"(",
"StrTemplate",
".",
"box",
"(",
"text",
"=",
"te... | Print a formatted ascii text box. | [
"Print",
"a",
"formatted",
"ascii",
"text",
"box",
"."
] | 689a60da51cd88680ddbe26e28dbe81e6b01d275 | https://github.com/MacHu-GWU/angora-project/blob/689a60da51cd88680ddbe26e28dbe81e6b01d275/angora/text/strtemplate.py#L98-L102 |
242,210 | kevinsprong23/aperture | aperture/util.py | increment | def increment(d, key, val=1):
"""
increment dict d at key by amount val
no need to return since d is mutable
"""
if key in d:
d[key] += val
else:
d[key] = val | python | def increment(d, key, val=1):
"""
increment dict d at key by amount val
no need to return since d is mutable
"""
if key in d:
d[key] += val
else:
d[key] = val | [
"def",
"increment",
"(",
"d",
",",
"key",
",",
"val",
"=",
"1",
")",
":",
"if",
"key",
"in",
"d",
":",
"d",
"[",
"key",
"]",
"+=",
"val",
"else",
":",
"d",
"[",
"key",
"]",
"=",
"val"
] | increment dict d at key by amount val
no need to return since d is mutable | [
"increment",
"dict",
"d",
"at",
"key",
"by",
"amount",
"val",
"no",
"need",
"to",
"return",
"since",
"d",
"is",
"mutable"
] | d0420fef3b25d8afc0e5ddcfb6fe5f0ff42b9799 | https://github.com/kevinsprong23/aperture/blob/d0420fef3b25d8afc0e5ddcfb6fe5f0ff42b9799/aperture/util.py#L7-L15 |
242,211 | kevinsprong23/aperture | aperture/util.py | floor_nearest | def floor_nearest(x, dx=1):
"""
floor a number to within a given rounding accuracy
"""
precision = get_sig_digits(dx)
return round(math.floor(float(x) / dx) * dx, precision) | python | def floor_nearest(x, dx=1):
"""
floor a number to within a given rounding accuracy
"""
precision = get_sig_digits(dx)
return round(math.floor(float(x) / dx) * dx, precision) | [
"def",
"floor_nearest",
"(",
"x",
",",
"dx",
"=",
"1",
")",
":",
"precision",
"=",
"get_sig_digits",
"(",
"dx",
")",
"return",
"round",
"(",
"math",
".",
"floor",
"(",
"float",
"(",
"x",
")",
"/",
"dx",
")",
"*",
"dx",
",",
"precision",
")"
] | floor a number to within a given rounding accuracy | [
"floor",
"a",
"number",
"to",
"within",
"a",
"given",
"rounding",
"accuracy"
] | d0420fef3b25d8afc0e5ddcfb6fe5f0ff42b9799 | https://github.com/kevinsprong23/aperture/blob/d0420fef3b25d8afc0e5ddcfb6fe5f0ff42b9799/aperture/util.py#L25-L30 |
242,212 | kevinsprong23/aperture | aperture/util.py | ceil_nearest | def ceil_nearest(x, dx=1):
"""
ceil a number to within a given rounding accuracy
"""
precision = get_sig_digits(dx)
return round(math.ceil(float(x) / dx) * dx, precision) | python | def ceil_nearest(x, dx=1):
"""
ceil a number to within a given rounding accuracy
"""
precision = get_sig_digits(dx)
return round(math.ceil(float(x) / dx) * dx, precision) | [
"def",
"ceil_nearest",
"(",
"x",
",",
"dx",
"=",
"1",
")",
":",
"precision",
"=",
"get_sig_digits",
"(",
"dx",
")",
"return",
"round",
"(",
"math",
".",
"ceil",
"(",
"float",
"(",
"x",
")",
"/",
"dx",
")",
"*",
"dx",
",",
"precision",
")"
] | ceil a number to within a given rounding accuracy | [
"ceil",
"a",
"number",
"to",
"within",
"a",
"given",
"rounding",
"accuracy"
] | d0420fef3b25d8afc0e5ddcfb6fe5f0ff42b9799 | https://github.com/kevinsprong23/aperture/blob/d0420fef3b25d8afc0e5ddcfb6fe5f0ff42b9799/aperture/util.py#L33-L38 |
242,213 | kevinsprong23/aperture | aperture/util.py | frange | def frange(x, y, jump=1):
"""
range for floats
"""
precision = get_sig_digits(jump)
while x < y:
yield round(x, precision)
x += jump | python | def frange(x, y, jump=1):
"""
range for floats
"""
precision = get_sig_digits(jump)
while x < y:
yield round(x, precision)
x += jump | [
"def",
"frange",
"(",
"x",
",",
"y",
",",
"jump",
"=",
"1",
")",
":",
"precision",
"=",
"get_sig_digits",
"(",
"jump",
")",
"while",
"x",
"<",
"y",
":",
"yield",
"round",
"(",
"x",
",",
"precision",
")",
"x",
"+=",
"jump"
] | range for floats | [
"range",
"for",
"floats"
] | d0420fef3b25d8afc0e5ddcfb6fe5f0ff42b9799 | https://github.com/kevinsprong23/aperture/blob/d0420fef3b25d8afc0e5ddcfb6fe5f0ff42b9799/aperture/util.py#L41-L48 |
242,214 | MacHu-GWU/angora-project | angora/dtypes/orderedset.py | OrderedSet.discard | def discard(self, key):
"""Remove a item from its member if it is a member.
Usage::
>>> s = OrderedSet([1, 2, 3])
>>> s.discard(2)
>>> s
OrderedSet([1, 3])
**中文文档**
从有序集合中删除一个元素, 同时保持集合依然有序。
"""
if key in self.map:
key, prev, next_item = self.map.pop(key)
prev[2] = next_item
next_item[1] = prev | python | def discard(self, key):
"""Remove a item from its member if it is a member.
Usage::
>>> s = OrderedSet([1, 2, 3])
>>> s.discard(2)
>>> s
OrderedSet([1, 3])
**中文文档**
从有序集合中删除一个元素, 同时保持集合依然有序。
"""
if key in self.map:
key, prev, next_item = self.map.pop(key)
prev[2] = next_item
next_item[1] = prev | [
"def",
"discard",
"(",
"self",
",",
"key",
")",
":",
"if",
"key",
"in",
"self",
".",
"map",
":",
"key",
",",
"prev",
",",
"next_item",
"=",
"self",
".",
"map",
".",
"pop",
"(",
"key",
")",
"prev",
"[",
"2",
"]",
"=",
"next_item",
"next_item",
"... | Remove a item from its member if it is a member.
Usage::
>>> s = OrderedSet([1, 2, 3])
>>> s.discard(2)
>>> s
OrderedSet([1, 3])
**中文文档**
从有序集合中删除一个元素, 同时保持集合依然有序。 | [
"Remove",
"a",
"item",
"from",
"its",
"member",
"if",
"it",
"is",
"a",
"member",
"."
] | 689a60da51cd88680ddbe26e28dbe81e6b01d275 | https://github.com/MacHu-GWU/angora-project/blob/689a60da51cd88680ddbe26e28dbe81e6b01d275/angora/dtypes/orderedset.py#L113-L130 |
242,215 | MacHu-GWU/angora-project | angora/dtypes/orderedset.py | OrderedSet.intersection | def intersection(*argv):
"""Returns the intersection of multiple sets.
Items are ordered by set1, set2, ...
**中文文档**
求多个有序集合的交集, 按照第一个集合, 第二个, ..., 这样的顺序。
"""
res = OrderedSet(argv[0])
for ods in argv:
res = ods & res
return res | python | def intersection(*argv):
"""Returns the intersection of multiple sets.
Items are ordered by set1, set2, ...
**中文文档**
求多个有序集合的交集, 按照第一个集合, 第二个, ..., 这样的顺序。
"""
res = OrderedSet(argv[0])
for ods in argv:
res = ods & res
return res | [
"def",
"intersection",
"(",
"*",
"argv",
")",
":",
"res",
"=",
"OrderedSet",
"(",
"argv",
"[",
"0",
"]",
")",
"for",
"ods",
"in",
"argv",
":",
"res",
"=",
"ods",
"&",
"res",
"return",
"res"
] | Returns the intersection of multiple sets.
Items are ordered by set1, set2, ...
**中文文档**
求多个有序集合的交集, 按照第一个集合, 第二个, ..., 这样的顺序。 | [
"Returns",
"the",
"intersection",
"of",
"multiple",
"sets",
".",
"Items",
"are",
"ordered",
"by",
"set1",
"set2",
"..."
] | 689a60da51cd88680ddbe26e28dbe81e6b01d275 | https://github.com/MacHu-GWU/angora-project/blob/689a60da51cd88680ddbe26e28dbe81e6b01d275/angora/dtypes/orderedset.py#L192-L203 |
242,216 | saintic/SpliceURL | SpliceURL.py | Splice.do | def do(self):
"run it, you can get a good stitching of the complete URL."
return urlunparse((self.scheme, self.netloc, self.path, self.params, self.query, self.fragment)) | python | def do(self):
"run it, you can get a good stitching of the complete URL."
return urlunparse((self.scheme, self.netloc, self.path, self.params, self.query, self.fragment)) | [
"def",
"do",
"(",
"self",
")",
":",
"return",
"urlunparse",
"(",
"(",
"self",
".",
"scheme",
",",
"self",
".",
"netloc",
",",
"self",
".",
"path",
",",
"self",
".",
"params",
",",
"self",
".",
"query",
",",
"self",
".",
"fragment",
")",
")"
] | run it, you can get a good stitching of the complete URL. | [
"run",
"it",
"you",
"can",
"get",
"a",
"good",
"stitching",
"of",
"the",
"complete",
"URL",
"."
] | ac2d1e854cbdfcf984bce3682b7f05ccba20938e | https://github.com/saintic/SpliceURL/blob/ac2d1e854cbdfcf984bce3682b7f05ccba20938e/SpliceURL.py#L46-L48 |
242,217 | saintic/SpliceURL | SpliceURL.py | Modify.do | def do(self):
"run it, get a new url"
scheme, netloc, path, params, query, fragment = Split(self.url).do()
if isinstance(self.query, dict):
query = query + "&" + urllib.urlencode(self.query) if query else urllib.urlencode(self.query)
path = urljoin(path, self.path).replace('\\', '/') if self.path else path
return Splice(scheme=scheme, netloc=netloc, path=path, params=params, query=query, fragment=fragment).geturl | python | def do(self):
"run it, get a new url"
scheme, netloc, path, params, query, fragment = Split(self.url).do()
if isinstance(self.query, dict):
query = query + "&" + urllib.urlencode(self.query) if query else urllib.urlencode(self.query)
path = urljoin(path, self.path).replace('\\', '/') if self.path else path
return Splice(scheme=scheme, netloc=netloc, path=path, params=params, query=query, fragment=fragment).geturl | [
"def",
"do",
"(",
"self",
")",
":",
"scheme",
",",
"netloc",
",",
"path",
",",
"params",
",",
"query",
",",
"fragment",
"=",
"Split",
"(",
"self",
".",
"url",
")",
".",
"do",
"(",
")",
"if",
"isinstance",
"(",
"self",
".",
"query",
",",
"dict",
... | run it, get a new url | [
"run",
"it",
"get",
"a",
"new",
"url"
] | ac2d1e854cbdfcf984bce3682b7f05ccba20938e | https://github.com/saintic/SpliceURL/blob/ac2d1e854cbdfcf984bce3682b7f05ccba20938e/SpliceURL.py#L89-L98 |
242,218 | luismasuelli/django-trackmodels-ritual | grimoire/django/tracked/admin.py | TrackedLiveAdmin.changelist_view | def changelist_view(self, request, extra_context=None):
"""
Updates the changelist view to include settings from this admin.
"""
return super(TrackedLiveAdmin, self).changelist_view(
request, dict(extra_context or {},
url_name='admin:%s_%s_tracking_report' % (self.model._meta.app_label, self.model._meta.model_name),
period_options=self.get_period_options(),
report_options=self.get_report_options())
) | python | def changelist_view(self, request, extra_context=None):
"""
Updates the changelist view to include settings from this admin.
"""
return super(TrackedLiveAdmin, self).changelist_view(
request, dict(extra_context or {},
url_name='admin:%s_%s_tracking_report' % (self.model._meta.app_label, self.model._meta.model_name),
period_options=self.get_period_options(),
report_options=self.get_report_options())
) | [
"def",
"changelist_view",
"(",
"self",
",",
"request",
",",
"extra_context",
"=",
"None",
")",
":",
"return",
"super",
"(",
"TrackedLiveAdmin",
",",
"self",
")",
".",
"changelist_view",
"(",
"request",
",",
"dict",
"(",
"extra_context",
"or",
"{",
"}",
","... | Updates the changelist view to include settings from this admin. | [
"Updates",
"the",
"changelist",
"view",
"to",
"include",
"settings",
"from",
"this",
"admin",
"."
] | ee0a6e07a5851ed477c9c1e3b9f8aafd9da35657 | https://github.com/luismasuelli/django-trackmodels-ritual/blob/ee0a6e07a5851ed477c9c1e3b9f8aafd9da35657/grimoire/django/tracked/admin.py#L154-L164 |
242,219 | luismasuelli/django-trackmodels-ritual | grimoire/django/tracked/admin.py | TrackedLiveAdmin.render_report_error | def render_report_error(self, request, error, status):
"""
Renders the report errors template.
"""
opts = self.model._meta
app_label = opts.app_label
request.current_app = self.admin_site.name
context = dict(
self.admin_site.each_context(request),
module_name=force_text(opts.verbose_name_plural),
title=(_('Tracking report error for %s') % force_text(opts.verbose_name)),
opts=opts, app_label=app_label, error=error
)
return TemplateResponse(request, self.report_error_template or [
"admin/{}/{}/tracking_report_error.html".format(app_label, opts.model_name),
"admin/{}/tracking_report_error.html".format(app_label),
"admin/tracking_report_error.html"
], context, status=status) | python | def render_report_error(self, request, error, status):
"""
Renders the report errors template.
"""
opts = self.model._meta
app_label = opts.app_label
request.current_app = self.admin_site.name
context = dict(
self.admin_site.each_context(request),
module_name=force_text(opts.verbose_name_plural),
title=(_('Tracking report error for %s') % force_text(opts.verbose_name)),
opts=opts, app_label=app_label, error=error
)
return TemplateResponse(request, self.report_error_template or [
"admin/{}/{}/tracking_report_error.html".format(app_label, opts.model_name),
"admin/{}/tracking_report_error.html".format(app_label),
"admin/tracking_report_error.html"
], context, status=status) | [
"def",
"render_report_error",
"(",
"self",
",",
"request",
",",
"error",
",",
"status",
")",
":",
"opts",
"=",
"self",
".",
"model",
".",
"_meta",
"app_label",
"=",
"opts",
".",
"app_label",
"request",
".",
"current_app",
"=",
"self",
".",
"admin_site",
... | Renders the report errors template. | [
"Renders",
"the",
"report",
"errors",
"template",
"."
] | ee0a6e07a5851ed477c9c1e3b9f8aafd9da35657 | https://github.com/luismasuelli/django-trackmodels-ritual/blob/ee0a6e07a5851ed477c9c1e3b9f8aafd9da35657/grimoire/django/tracked/admin.py#L166-L185 |
242,220 | luismasuelli/django-trackmodels-ritual | grimoire/django/tracked/admin.py | TrackedLiveAdmin.report_view | def report_view(self, request, key, period):
"""
Processes the reporting action.
"""
if not self.has_change_permission(request, None):
raise PermissionDenied
reporters = self.get_reporters()
try:
reporter = reporters[key]
except KeyError:
return self.render_report_error(request, _('Report not found'), 404)
allowed_periods = [k for (k, v) in self.get_period_options()]
if period == 'A':
period = ''
if period and period not in allowed_periods:
return self.render_report_error(request, _('Invalid report type'), 400)
try:
return reporter.process(request, self.get_period_queryset(request, period), period)
except:
logger.exception('Tracking Reports could not generate the report due to an internal error')
return self.render_report_error(request, _('An unexpected error has occurred'), 500) | python | def report_view(self, request, key, period):
"""
Processes the reporting action.
"""
if not self.has_change_permission(request, None):
raise PermissionDenied
reporters = self.get_reporters()
try:
reporter = reporters[key]
except KeyError:
return self.render_report_error(request, _('Report not found'), 404)
allowed_periods = [k for (k, v) in self.get_period_options()]
if period == 'A':
period = ''
if period and period not in allowed_periods:
return self.render_report_error(request, _('Invalid report type'), 400)
try:
return reporter.process(request, self.get_period_queryset(request, period), period)
except:
logger.exception('Tracking Reports could not generate the report due to an internal error')
return self.render_report_error(request, _('An unexpected error has occurred'), 500) | [
"def",
"report_view",
"(",
"self",
",",
"request",
",",
"key",
",",
"period",
")",
":",
"if",
"not",
"self",
".",
"has_change_permission",
"(",
"request",
",",
"None",
")",
":",
"raise",
"PermissionDenied",
"reporters",
"=",
"self",
".",
"get_reporters",
"... | Processes the reporting action. | [
"Processes",
"the",
"reporting",
"action",
"."
] | ee0a6e07a5851ed477c9c1e3b9f8aafd9da35657 | https://github.com/luismasuelli/django-trackmodels-ritual/blob/ee0a6e07a5851ed477c9c1e3b9f8aafd9da35657/grimoire/django/tracked/admin.py#L207-L232 |
242,221 | almcc/cinder-data | cinder_data/store.py | Store.find_record | def find_record(self, model_class, record_id, reload=False):
"""Return a instance of model_class from the API or the local cache.
Args:
model_class (:class:`cinder_data.model.CinderModel`): A subclass of
:class:`cinder_data.model.CinderModel` of your chosen model.
record_id (int): The id of the record requested.
reload (bool, optional): Don't return the cached version if reload==True.
Returns:
:class:`cinder_data.model.CinderModel`: An instance of model_class or None.
"""
cached_model = self.peek_record(model_class, record_id)
if cached_model is not None and reload is False:
return cached_model
else:
return self._get_record(model_class, record_id) | python | def find_record(self, model_class, record_id, reload=False):
"""Return a instance of model_class from the API or the local cache.
Args:
model_class (:class:`cinder_data.model.CinderModel`): A subclass of
:class:`cinder_data.model.CinderModel` of your chosen model.
record_id (int): The id of the record requested.
reload (bool, optional): Don't return the cached version if reload==True.
Returns:
:class:`cinder_data.model.CinderModel`: An instance of model_class or None.
"""
cached_model = self.peek_record(model_class, record_id)
if cached_model is not None and reload is False:
return cached_model
else:
return self._get_record(model_class, record_id) | [
"def",
"find_record",
"(",
"self",
",",
"model_class",
",",
"record_id",
",",
"reload",
"=",
"False",
")",
":",
"cached_model",
"=",
"self",
".",
"peek_record",
"(",
"model_class",
",",
"record_id",
")",
"if",
"cached_model",
"is",
"not",
"None",
"and",
"r... | Return a instance of model_class from the API or the local cache.
Args:
model_class (:class:`cinder_data.model.CinderModel`): A subclass of
:class:`cinder_data.model.CinderModel` of your chosen model.
record_id (int): The id of the record requested.
reload (bool, optional): Don't return the cached version if reload==True.
Returns:
:class:`cinder_data.model.CinderModel`: An instance of model_class or None. | [
"Return",
"a",
"instance",
"of",
"model_class",
"from",
"the",
"API",
"or",
"the",
"local",
"cache",
"."
] | 4159a5186c4b4fc32354749892e86130530f6ec5 | https://github.com/almcc/cinder-data/blob/4159a5186c4b4fc32354749892e86130530f6ec5/cinder_data/store.py#L23-L39 |
242,222 | almcc/cinder-data | cinder_data/store.py | Store.peek_record | def peek_record(self, model_class, record_id):
"""Return an instance of the model_class from the cache if it is present.
Args:
model_class (:class:`cinder_data.model.CinderModel`): A subclass of
:class:`cinder_data.model.CinderModel` of your chosen model.
record_id (int): The id of the record requested.
Returns:
:class:`cinder_data.model.CinderModel`: An instance of model_class or None.
"""
if self._cache:
return self._cache.get_record(model_class.__name__, record_id)
else:
return None | python | def peek_record(self, model_class, record_id):
"""Return an instance of the model_class from the cache if it is present.
Args:
model_class (:class:`cinder_data.model.CinderModel`): A subclass of
:class:`cinder_data.model.CinderModel` of your chosen model.
record_id (int): The id of the record requested.
Returns:
:class:`cinder_data.model.CinderModel`: An instance of model_class or None.
"""
if self._cache:
return self._cache.get_record(model_class.__name__, record_id)
else:
return None | [
"def",
"peek_record",
"(",
"self",
",",
"model_class",
",",
"record_id",
")",
":",
"if",
"self",
".",
"_cache",
":",
"return",
"self",
".",
"_cache",
".",
"get_record",
"(",
"model_class",
".",
"__name__",
",",
"record_id",
")",
"else",
":",
"return",
"N... | Return an instance of the model_class from the cache if it is present.
Args:
model_class (:class:`cinder_data.model.CinderModel`): A subclass of
:class:`cinder_data.model.CinderModel` of your chosen model.
record_id (int): The id of the record requested.
Returns:
:class:`cinder_data.model.CinderModel`: An instance of model_class or None. | [
"Return",
"an",
"instance",
"of",
"the",
"model_class",
"from",
"the",
"cache",
"if",
"it",
"is",
"present",
"."
] | 4159a5186c4b4fc32354749892e86130530f6ec5 | https://github.com/almcc/cinder-data/blob/4159a5186c4b4fc32354749892e86130530f6ec5/cinder_data/store.py#L41-L55 |
242,223 | almcc/cinder-data | cinder_data/store.py | Store.find_all | def find_all(self, model_class, params={}):
"""Return an list of models from the API and caches the result.
Args:
model_class (:class:`cinder_data.model.CinderModel`): A subclass of
:class:`cinder_data.model.CinderModel` of your chosen model.
params (dict, optional): Description
Returns:
list: A list of instances of you model_class or and empty list.
"""
url = '{host}/{namespace}/{model}{params}'.format(
host=self._host,
namespace=self._namespace,
model=self._translate_name(model_class.__name__),
params=self._build_param_string(params)
)
data = self._get_json(url)['data']
fresh_models = []
for item in data:
fresh_model = model_class(item['attributes'])
fresh_model.id = item['id']
fresh_model.validate()
fresh_models.append(fresh_model)
if self._cache is not None:
self._cache.set_record(model_class.__name__, fresh_model.id, fresh_model)
return fresh_models | python | def find_all(self, model_class, params={}):
"""Return an list of models from the API and caches the result.
Args:
model_class (:class:`cinder_data.model.CinderModel`): A subclass of
:class:`cinder_data.model.CinderModel` of your chosen model.
params (dict, optional): Description
Returns:
list: A list of instances of you model_class or and empty list.
"""
url = '{host}/{namespace}/{model}{params}'.format(
host=self._host,
namespace=self._namespace,
model=self._translate_name(model_class.__name__),
params=self._build_param_string(params)
)
data = self._get_json(url)['data']
fresh_models = []
for item in data:
fresh_model = model_class(item['attributes'])
fresh_model.id = item['id']
fresh_model.validate()
fresh_models.append(fresh_model)
if self._cache is not None:
self._cache.set_record(model_class.__name__, fresh_model.id, fresh_model)
return fresh_models | [
"def",
"find_all",
"(",
"self",
",",
"model_class",
",",
"params",
"=",
"{",
"}",
")",
":",
"url",
"=",
"'{host}/{namespace}/{model}{params}'",
".",
"format",
"(",
"host",
"=",
"self",
".",
"_host",
",",
"namespace",
"=",
"self",
".",
"_namespace",
",",
... | Return an list of models from the API and caches the result.
Args:
model_class (:class:`cinder_data.model.CinderModel`): A subclass of
:class:`cinder_data.model.CinderModel` of your chosen model.
params (dict, optional): Description
Returns:
list: A list of instances of you model_class or and empty list. | [
"Return",
"an",
"list",
"of",
"models",
"from",
"the",
"API",
"and",
"caches",
"the",
"result",
"."
] | 4159a5186c4b4fc32354749892e86130530f6ec5 | https://github.com/almcc/cinder-data/blob/4159a5186c4b4fc32354749892e86130530f6ec5/cinder_data/store.py#L57-L83 |
242,224 | almcc/cinder-data | cinder_data/store.py | Store.peek_all | def peek_all(self, model_class):
"""Return a list of models from the local cache.
Args:
model_class (:class:`cinder_data.model.CinderModel`): A subclass of
:class:`cinder_data.model.CinderModel` of your chosen model.
Returns:
list: A list of instances of you model_class or and empty list.
"""
if self._cache:
return self._cache.get_records(model_class.__name__)
else:
return [] | python | def peek_all(self, model_class):
"""Return a list of models from the local cache.
Args:
model_class (:class:`cinder_data.model.CinderModel`): A subclass of
:class:`cinder_data.model.CinderModel` of your chosen model.
Returns:
list: A list of instances of you model_class or and empty list.
"""
if self._cache:
return self._cache.get_records(model_class.__name__)
else:
return [] | [
"def",
"peek_all",
"(",
"self",
",",
"model_class",
")",
":",
"if",
"self",
".",
"_cache",
":",
"return",
"self",
".",
"_cache",
".",
"get_records",
"(",
"model_class",
".",
"__name__",
")",
"else",
":",
"return",
"[",
"]"
] | Return a list of models from the local cache.
Args:
model_class (:class:`cinder_data.model.CinderModel`): A subclass of
:class:`cinder_data.model.CinderModel` of your chosen model.
Returns:
list: A list of instances of you model_class or and empty list. | [
"Return",
"a",
"list",
"of",
"models",
"from",
"the",
"local",
"cache",
"."
] | 4159a5186c4b4fc32354749892e86130530f6ec5 | https://github.com/almcc/cinder-data/blob/4159a5186c4b4fc32354749892e86130530f6ec5/cinder_data/store.py#L85-L98 |
242,225 | almcc/cinder-data | cinder_data/store.py | Store._get_record | def _get_record(self, model_class, record_id):
"""Get a single record from the API.
Args:
model_class (:class:`cinder_data.model.CinderModel`): A subclass of
:class:`cinder_data.model.CinderModel` of your chosen model.
record_id (int): The id of the record requested.
Returns:
:class:`cinder_data.model.CinderModel`: An instance of model_class or None.
"""
url = '{host}/{namespace}/{model}/{id}'.format(
host=self._host,
namespace=self._namespace,
model=self._translate_name(model_class.__name__),
id=record_id
)
data = self._get_json(url)['data']
fresh_model = model_class(data['attributes'])
fresh_model.id = data['id']
fresh_model.validate()
if self._cache is not None:
self._cache.set_record(model_class.__name__, fresh_model.id, fresh_model)
return fresh_model | python | def _get_record(self, model_class, record_id):
"""Get a single record from the API.
Args:
model_class (:class:`cinder_data.model.CinderModel`): A subclass of
:class:`cinder_data.model.CinderModel` of your chosen model.
record_id (int): The id of the record requested.
Returns:
:class:`cinder_data.model.CinderModel`: An instance of model_class or None.
"""
url = '{host}/{namespace}/{model}/{id}'.format(
host=self._host,
namespace=self._namespace,
model=self._translate_name(model_class.__name__),
id=record_id
)
data = self._get_json(url)['data']
fresh_model = model_class(data['attributes'])
fresh_model.id = data['id']
fresh_model.validate()
if self._cache is not None:
self._cache.set_record(model_class.__name__, fresh_model.id, fresh_model)
return fresh_model | [
"def",
"_get_record",
"(",
"self",
",",
"model_class",
",",
"record_id",
")",
":",
"url",
"=",
"'{host}/{namespace}/{model}/{id}'",
".",
"format",
"(",
"host",
"=",
"self",
".",
"_host",
",",
"namespace",
"=",
"self",
".",
"_namespace",
",",
"model",
"=",
... | Get a single record from the API.
Args:
model_class (:class:`cinder_data.model.CinderModel`): A subclass of
:class:`cinder_data.model.CinderModel` of your chosen model.
record_id (int): The id of the record requested.
Returns:
:class:`cinder_data.model.CinderModel`: An instance of model_class or None. | [
"Get",
"a",
"single",
"record",
"from",
"the",
"API",
"."
] | 4159a5186c4b4fc32354749892e86130530f6ec5 | https://github.com/almcc/cinder-data/blob/4159a5186c4b4fc32354749892e86130530f6ec5/cinder_data/store.py#L100-L123 |
242,226 | almcc/cinder-data | cinder_data/store.py | Store._translate_name | def _translate_name(name):
"""Translate the class name to the API endpoint.
For example, Car would become cars, FastCar would become fast-cars.
Args:
name (string): Camel case name (singular)
Returns:
string: A pluraised, dasherized string.
"""
underscored = inflection.underscore(name)
dasherized = inflection.dasherize(underscored)
words = dasherized.split('-')
last_word = words.pop()
words.append(inflection.pluralize(last_word))
return '-'.join(words) | python | def _translate_name(name):
"""Translate the class name to the API endpoint.
For example, Car would become cars, FastCar would become fast-cars.
Args:
name (string): Camel case name (singular)
Returns:
string: A pluraised, dasherized string.
"""
underscored = inflection.underscore(name)
dasherized = inflection.dasherize(underscored)
words = dasherized.split('-')
last_word = words.pop()
words.append(inflection.pluralize(last_word))
return '-'.join(words) | [
"def",
"_translate_name",
"(",
"name",
")",
":",
"underscored",
"=",
"inflection",
".",
"underscore",
"(",
"name",
")",
"dasherized",
"=",
"inflection",
".",
"dasherize",
"(",
"underscored",
")",
"words",
"=",
"dasherized",
".",
"split",
"(",
"'-'",
")",
"... | Translate the class name to the API endpoint.
For example, Car would become cars, FastCar would become fast-cars.
Args:
name (string): Camel case name (singular)
Returns:
string: A pluraised, dasherized string. | [
"Translate",
"the",
"class",
"name",
"to",
"the",
"API",
"endpoint",
"."
] | 4159a5186c4b4fc32354749892e86130530f6ec5 | https://github.com/almcc/cinder-data/blob/4159a5186c4b4fc32354749892e86130530f6ec5/cinder_data/store.py#L139-L155 |
242,227 | almcc/cinder-data | cinder_data/store.py | Store._build_param_string | def _build_param_string(params):
"""Build query params string from a dictionary.
Args:
params (dict): A dictionary of params
Returns:
string: A valid url query params string.
"""
pairs = []
for key, value in params.iteritems():
if value is None:
value = ''
pairs.append('{0}={1}'.format(key, value))
if len(pairs) > 0:
return '?{0}'.format('&'.join(pairs))
return '' | python | def _build_param_string(params):
"""Build query params string from a dictionary.
Args:
params (dict): A dictionary of params
Returns:
string: A valid url query params string.
"""
pairs = []
for key, value in params.iteritems():
if value is None:
value = ''
pairs.append('{0}={1}'.format(key, value))
if len(pairs) > 0:
return '?{0}'.format('&'.join(pairs))
return '' | [
"def",
"_build_param_string",
"(",
"params",
")",
":",
"pairs",
"=",
"[",
"]",
"for",
"key",
",",
"value",
"in",
"params",
".",
"iteritems",
"(",
")",
":",
"if",
"value",
"is",
"None",
":",
"value",
"=",
"''",
"pairs",
".",
"append",
"(",
"'{0}={1}'"... | Build query params string from a dictionary.
Args:
params (dict): A dictionary of params
Returns:
string: A valid url query params string. | [
"Build",
"query",
"params",
"string",
"from",
"a",
"dictionary",
"."
] | 4159a5186c4b4fc32354749892e86130530f6ec5 | https://github.com/almcc/cinder-data/blob/4159a5186c4b4fc32354749892e86130530f6ec5/cinder_data/store.py#L158-L174 |
242,228 | selenol/selenol-python | selenol_python/connections.py | SelenolWSConnection.send | def send(self, message):
"""Send a the defined message to the backend.
:param message: Message to be send, usually a Python dictionary.
"""
try:
self.ws.send(json.dumps(message))
except websocket._exceptions.WebSocketConnectionClosedException:
raise SelenolWebSocketClosedException() | python | def send(self, message):
"""Send a the defined message to the backend.
:param message: Message to be send, usually a Python dictionary.
"""
try:
self.ws.send(json.dumps(message))
except websocket._exceptions.WebSocketConnectionClosedException:
raise SelenolWebSocketClosedException() | [
"def",
"send",
"(",
"self",
",",
"message",
")",
":",
"try",
":",
"self",
".",
"ws",
".",
"send",
"(",
"json",
".",
"dumps",
"(",
"message",
")",
")",
"except",
"websocket",
".",
"_exceptions",
".",
"WebSocketConnectionClosedException",
":",
"raise",
"Se... | Send a the defined message to the backend.
:param message: Message to be send, usually a Python dictionary. | [
"Send",
"a",
"the",
"defined",
"message",
"to",
"the",
"backend",
"."
] | 53775fdfc95161f4aca350305cb3459e6f2f808d | https://github.com/selenol/selenol-python/blob/53775fdfc95161f4aca350305cb3459e6f2f808d/selenol_python/connections.py#L37-L45 |
242,229 | selenol/selenol-python | selenol_python/connections.py | SelenolWSConnection.recv | def recv(self):
"""Receive message from the backend or wait unilt next message."""
try:
message = self.ws.recv()
return json.loads(message)
except websocket._exceptions.WebSocketConnectionClosedException as ex:
raise SelenolWebSocketClosedException() from ex | python | def recv(self):
"""Receive message from the backend or wait unilt next message."""
try:
message = self.ws.recv()
return json.loads(message)
except websocket._exceptions.WebSocketConnectionClosedException as ex:
raise SelenolWebSocketClosedException() from ex | [
"def",
"recv",
"(",
"self",
")",
":",
"try",
":",
"message",
"=",
"self",
".",
"ws",
".",
"recv",
"(",
")",
"return",
"json",
".",
"loads",
"(",
"message",
")",
"except",
"websocket",
".",
"_exceptions",
".",
"WebSocketConnectionClosedException",
"as",
"... | Receive message from the backend or wait unilt next message. | [
"Receive",
"message",
"from",
"the",
"backend",
"or",
"wait",
"unilt",
"next",
"message",
"."
] | 53775fdfc95161f4aca350305cb3459e6f2f808d | https://github.com/selenol/selenol-python/blob/53775fdfc95161f4aca350305cb3459e6f2f808d/selenol_python/connections.py#L47-L53 |
242,230 | krukas/Trionyx | trionyx/trionyx/management/commands/create_app.py | Command.handle | def handle(self, *args, **options):
"""Create new app"""
quickstart = Quickstart()
try:
quickstart.create_app(os.path.join(settings.BASE_DIR, 'apps'), options.get('name'))
self.stdout.write(
self.style.SUCCESS("Successfully created app ({name}), don't forget to add 'apps.{name}' to INSTALLED_APPS".format(
name=options.get('name')
))
)
except FileExistsError as e:
print(e)
raise CommandError("App with same name already exists") | python | def handle(self, *args, **options):
"""Create new app"""
quickstart = Quickstart()
try:
quickstart.create_app(os.path.join(settings.BASE_DIR, 'apps'), options.get('name'))
self.stdout.write(
self.style.SUCCESS("Successfully created app ({name}), don't forget to add 'apps.{name}' to INSTALLED_APPS".format(
name=options.get('name')
))
)
except FileExistsError as e:
print(e)
raise CommandError("App with same name already exists") | [
"def",
"handle",
"(",
"self",
",",
"*",
"args",
",",
"*",
"*",
"options",
")",
":",
"quickstart",
"=",
"Quickstart",
"(",
")",
"try",
":",
"quickstart",
".",
"create_app",
"(",
"os",
".",
"path",
".",
"join",
"(",
"settings",
".",
"BASE_DIR",
",",
... | Create new app | [
"Create",
"new",
"app"
] | edac132cc0797190153f2e60bc7e88cb50e80da6 | https://github.com/krukas/Trionyx/blob/edac132cc0797190153f2e60bc7e88cb50e80da6/trionyx/trionyx/management/commands/create_app.py#L25-L38 |
242,231 | ANCIR/granoloader | granoloader/command.py | make_client | def make_client(host, project_name, api_key, create_project):
""" Instantiate the grano client based on environment variables or
command line settings. """
if host is None:
raise click.BadParameter('No grano server host is set', param=host)
if project_name is None:
raise click.BadParameter('No grano project slug is set',
param=project_name)
if api_key is None:
raise click.BadParameter('No grano API key is set', param=api_key)
client = Grano(api_host=host,
api_key=api_key)
try:
return client.get(project_name)
except NotFound:
if not create_project:
sys.exit(-1)
data = {'slug': project_name, 'label': project_name}
return client.projects.create(data) | python | def make_client(host, project_name, api_key, create_project):
""" Instantiate the grano client based on environment variables or
command line settings. """
if host is None:
raise click.BadParameter('No grano server host is set', param=host)
if project_name is None:
raise click.BadParameter('No grano project slug is set',
param=project_name)
if api_key is None:
raise click.BadParameter('No grano API key is set', param=api_key)
client = Grano(api_host=host,
api_key=api_key)
try:
return client.get(project_name)
except NotFound:
if not create_project:
sys.exit(-1)
data = {'slug': project_name, 'label': project_name}
return client.projects.create(data) | [
"def",
"make_client",
"(",
"host",
",",
"project_name",
",",
"api_key",
",",
"create_project",
")",
":",
"if",
"host",
"is",
"None",
":",
"raise",
"click",
".",
"BadParameter",
"(",
"'No grano server host is set'",
",",
"param",
"=",
"host",
")",
"if",
"proj... | Instantiate the grano client based on environment variables or
command line settings. | [
"Instantiate",
"the",
"grano",
"client",
"based",
"on",
"environment",
"variables",
"or",
"command",
"line",
"settings",
"."
] | c48b1bd50403dd611340c5f51637f7c5ca54059c | https://github.com/ANCIR/granoloader/blob/c48b1bd50403dd611340c5f51637f7c5ca54059c/granoloader/command.py#L21-L40 |
242,232 | ANCIR/granoloader | granoloader/command.py | csv | def csv(ctx, force, threads, mapping, data):
""" Load CSV data into a grano instance using a mapping specification. """
# Find out how many lines there are (for the progress bar).
lines = 0
for line in DictReader(data):
lines += 1
data.seek(0)
# set up objects
mapping = yaml.load(mapping)
mapping_loader = MappingLoader(ctx.obj['grano'], mapping)
def process_row(row):
try:
mapping_loader.load(row)
except GranoException, ge:
msg = '\nServer error: %s' % ge.message
click.secho(msg, fg='red', bold=True)
if not force:
os._exit(1)
except RowException, re:
if not force:
msg = '\nRow %s: %s' % (row['__row_id__'], re.message)
click.secho(msg, fg='red', bold=True)
os._exit(1)
def generate():
with click.progressbar(DictReader(data),
label=data.name,
length=lines) as bar:
for i, row in enumerate(bar):
row['__row_id__'] = i
yield row
threaded(generate(), process_row, num_threads=threads,
max_queue=1) | python | def csv(ctx, force, threads, mapping, data):
""" Load CSV data into a grano instance using a mapping specification. """
# Find out how many lines there are (for the progress bar).
lines = 0
for line in DictReader(data):
lines += 1
data.seek(0)
# set up objects
mapping = yaml.load(mapping)
mapping_loader = MappingLoader(ctx.obj['grano'], mapping)
def process_row(row):
try:
mapping_loader.load(row)
except GranoException, ge:
msg = '\nServer error: %s' % ge.message
click.secho(msg, fg='red', bold=True)
if not force:
os._exit(1)
except RowException, re:
if not force:
msg = '\nRow %s: %s' % (row['__row_id__'], re.message)
click.secho(msg, fg='red', bold=True)
os._exit(1)
def generate():
with click.progressbar(DictReader(data),
label=data.name,
length=lines) as bar:
for i, row in enumerate(bar):
row['__row_id__'] = i
yield row
threaded(generate(), process_row, num_threads=threads,
max_queue=1) | [
"def",
"csv",
"(",
"ctx",
",",
"force",
",",
"threads",
",",
"mapping",
",",
"data",
")",
":",
"# Find out how many lines there are (for the progress bar).",
"lines",
"=",
"0",
"for",
"line",
"in",
"DictReader",
"(",
"data",
")",
":",
"lines",
"+=",
"1",
"da... | Load CSV data into a grano instance using a mapping specification. | [
"Load",
"CSV",
"data",
"into",
"a",
"grano",
"instance",
"using",
"a",
"mapping",
"specification",
"."
] | c48b1bd50403dd611340c5f51637f7c5ca54059c | https://github.com/ANCIR/granoloader/blob/c48b1bd50403dd611340c5f51637f7c5ca54059c/granoloader/command.py#L69-L105 |
242,233 | ANCIR/granoloader | granoloader/command.py | schema | def schema(ctx, schema):
""" Load schema definitions from a YAML file. """
data = yaml.load(schema)
if not isinstance(data, (list, tuple)):
data = [data]
with click.progressbar(data, label=schema.name) as bar:
for schema in bar:
ctx.obj['grano'].schemata.upsert(schema) | python | def schema(ctx, schema):
""" Load schema definitions from a YAML file. """
data = yaml.load(schema)
if not isinstance(data, (list, tuple)):
data = [data]
with click.progressbar(data, label=schema.name) as bar:
for schema in bar:
ctx.obj['grano'].schemata.upsert(schema) | [
"def",
"schema",
"(",
"ctx",
",",
"schema",
")",
":",
"data",
"=",
"yaml",
".",
"load",
"(",
"schema",
")",
"if",
"not",
"isinstance",
"(",
"data",
",",
"(",
"list",
",",
"tuple",
")",
")",
":",
"data",
"=",
"[",
"data",
"]",
"with",
"click",
"... | Load schema definitions from a YAML file. | [
"Load",
"schema",
"definitions",
"from",
"a",
"YAML",
"file",
"."
] | c48b1bd50403dd611340c5f51637f7c5ca54059c | https://github.com/ANCIR/granoloader/blob/c48b1bd50403dd611340c5f51637f7c5ca54059c/granoloader/command.py#L111-L118 |
242,234 | JNRowe/jnrbase | jnrbase/json_datetime.py | json_using_iso8601 | def json_using_iso8601(__obj: Dict) -> Dict:
"""Parse ISO-8601 values from JSON databases.
See :class:`json.JSONDecoder`
Args:
__obj: Object to decode
"""
for key, value in __obj.items():
with suppress(TypeError, ValueError):
__obj[key] = parse_datetime(value)
with suppress(TypeError, ValueError):
__obj[key] = parse_delta(value)
return __obj | python | def json_using_iso8601(__obj: Dict) -> Dict:
"""Parse ISO-8601 values from JSON databases.
See :class:`json.JSONDecoder`
Args:
__obj: Object to decode
"""
for key, value in __obj.items():
with suppress(TypeError, ValueError):
__obj[key] = parse_datetime(value)
with suppress(TypeError, ValueError):
__obj[key] = parse_delta(value)
return __obj | [
"def",
"json_using_iso8601",
"(",
"__obj",
":",
"Dict",
")",
"->",
"Dict",
":",
"for",
"key",
",",
"value",
"in",
"__obj",
".",
"items",
"(",
")",
":",
"with",
"suppress",
"(",
"TypeError",
",",
"ValueError",
")",
":",
"__obj",
"[",
"key",
"]",
"=",
... | Parse ISO-8601 values from JSON databases.
See :class:`json.JSONDecoder`
Args:
__obj: Object to decode | [
"Parse",
"ISO",
"-",
"8601",
"values",
"from",
"JSON",
"databases",
"."
] | ae505ef69a9feb739b5f4e62c5a8e6533104d3ea | https://github.com/JNRowe/jnrbase/blob/ae505ef69a9feb739b5f4e62c5a8e6533104d3ea/jnrbase/json_datetime.py#L61-L74 |
242,235 | lddubeau/glerbl | glerbl/__init__.py | get_tmpdir | def get_tmpdir():
"""
On first invocation, creates a temporary directory and returns its
path. Subsequent invocations uses the same directory.
:returns: A temporary directory created for this run of glerbl.
:rtype: :class:`str`
"""
global __tmpdir
if __tmpdir is not None:
return __tmpdir
__tmpdir = tempfile.mkdtemp(prefix='.tmp.glerbl.', dir=".")
atexit.register(__clean_tmpdir)
return __tmpdir | python | def get_tmpdir():
"""
On first invocation, creates a temporary directory and returns its
path. Subsequent invocations uses the same directory.
:returns: A temporary directory created for this run of glerbl.
:rtype: :class:`str`
"""
global __tmpdir
if __tmpdir is not None:
return __tmpdir
__tmpdir = tempfile.mkdtemp(prefix='.tmp.glerbl.', dir=".")
atexit.register(__clean_tmpdir)
return __tmpdir | [
"def",
"get_tmpdir",
"(",
")",
":",
"global",
"__tmpdir",
"if",
"__tmpdir",
"is",
"not",
"None",
":",
"return",
"__tmpdir",
"__tmpdir",
"=",
"tempfile",
".",
"mkdtemp",
"(",
"prefix",
"=",
"'.tmp.glerbl.'",
",",
"dir",
"=",
"\".\"",
")",
"atexit",
".",
"... | On first invocation, creates a temporary directory and returns its
path. Subsequent invocations uses the same directory.
:returns: A temporary directory created for this run of glerbl.
:rtype: :class:`str` | [
"On",
"first",
"invocation",
"creates",
"a",
"temporary",
"directory",
"and",
"returns",
"its",
"path",
".",
"Subsequent",
"invocations",
"uses",
"the",
"same",
"directory",
"."
] | b43adf851f8ce4c5b1dba22ee3556a169681bce1 | https://github.com/lddubeau/glerbl/blob/b43adf851f8ce4c5b1dba22ee3556a169681bce1/glerbl/__init__.py#L17-L32 |
242,236 | lddubeau/glerbl | glerbl/__init__.py | get_against | def get_against():
"""
Determines the revision against which the staged data ought to be checked.
:returns: The revision.
:rtype: :class:`str`
"""
global __cached_against
if __cached_against is not None:
return __cached_against
status = subprocess.call(["git", "rev-parse", "--verify", "HEAD"],
stdout=open(os.devnull, 'w'),
stderr=subprocess.STDOUT)
if not status:
against = 'HEAD'
else:
# Initial commit: diff against an empty tree object
against = '4b825dc642cb6eb9a060e54bf8d69288fbee4904'
__cached_against = against
return against | python | def get_against():
"""
Determines the revision against which the staged data ought to be checked.
:returns: The revision.
:rtype: :class:`str`
"""
global __cached_against
if __cached_against is not None:
return __cached_against
status = subprocess.call(["git", "rev-parse", "--verify", "HEAD"],
stdout=open(os.devnull, 'w'),
stderr=subprocess.STDOUT)
if not status:
against = 'HEAD'
else:
# Initial commit: diff against an empty tree object
against = '4b825dc642cb6eb9a060e54bf8d69288fbee4904'
__cached_against = against
return against | [
"def",
"get_against",
"(",
")",
":",
"global",
"__cached_against",
"if",
"__cached_against",
"is",
"not",
"None",
":",
"return",
"__cached_against",
"status",
"=",
"subprocess",
".",
"call",
"(",
"[",
"\"git\"",
",",
"\"rev-parse\"",
",",
"\"--verify\"",
",",
... | Determines the revision against which the staged data ought to be checked.
:returns: The revision.
:rtype: :class:`str` | [
"Determines",
"the",
"revision",
"against",
"which",
"the",
"staged",
"data",
"ought",
"to",
"be",
"checked",
"."
] | b43adf851f8ce4c5b1dba22ee3556a169681bce1 | https://github.com/lddubeau/glerbl/blob/b43adf851f8ce4c5b1dba22ee3556a169681bce1/glerbl/__init__.py#L36-L57 |
242,237 | JNRowe/jnrbase | jnrbase/entry.py | entry_point | def entry_point(__func: Callable) -> Callable:
"""Execute function when module is run directly.
Note:
This allows fall through for importing modules that use it.
Args:
__func: Function to run
"""
if __func.__module__ == '__main__':
import sys
sys.exit(__func())
else:
return __func | python | def entry_point(__func: Callable) -> Callable:
"""Execute function when module is run directly.
Note:
This allows fall through for importing modules that use it.
Args:
__func: Function to run
"""
if __func.__module__ == '__main__':
import sys
sys.exit(__func())
else:
return __func | [
"def",
"entry_point",
"(",
"__func",
":",
"Callable",
")",
"->",
"Callable",
":",
"if",
"__func",
".",
"__module__",
"==",
"'__main__'",
":",
"import",
"sys",
"sys",
".",
"exit",
"(",
"__func",
"(",
")",
")",
"else",
":",
"return",
"__func"
] | Execute function when module is run directly.
Note:
This allows fall through for importing modules that use it.
Args:
__func: Function to run | [
"Execute",
"function",
"when",
"module",
"is",
"run",
"directly",
"."
] | ae505ef69a9feb739b5f4e62c5a8e6533104d3ea | https://github.com/JNRowe/jnrbase/blob/ae505ef69a9feb739b5f4e62c5a8e6533104d3ea/jnrbase/entry.py#L24-L37 |
242,238 | DallasMorningNews/django-datafreezer | datafreezer/models.py | create_col_nums | def create_col_nums():
"""Return column numbers and letters that repeat up to NUM_REPEATS.
I.e., NUM_REPEATS = 2 would return a list of 26 * 26 = 676 2-tuples.
"""
NUM_REPEATS = 2
column_letters = list(
string.ascii_uppercase
) + map(
''.join,
itertools.product(
string.ascii_uppercase,
repeat=NUM_REPEATS
)
)
letter_numbers = []
count = 1
for letter in column_letters:
letter_numbers.append((count, str(count) + ' (' + letter + ')'))
count += 1
return tuple(letter_numbers) | python | def create_col_nums():
"""Return column numbers and letters that repeat up to NUM_REPEATS.
I.e., NUM_REPEATS = 2 would return a list of 26 * 26 = 676 2-tuples.
"""
NUM_REPEATS = 2
column_letters = list(
string.ascii_uppercase
) + map(
''.join,
itertools.product(
string.ascii_uppercase,
repeat=NUM_REPEATS
)
)
letter_numbers = []
count = 1
for letter in column_letters:
letter_numbers.append((count, str(count) + ' (' + letter + ')'))
count += 1
return tuple(letter_numbers) | [
"def",
"create_col_nums",
"(",
")",
":",
"NUM_REPEATS",
"=",
"2",
"column_letters",
"=",
"list",
"(",
"string",
".",
"ascii_uppercase",
")",
"+",
"map",
"(",
"''",
".",
"join",
",",
"itertools",
".",
"product",
"(",
"string",
".",
"ascii_uppercase",
",",
... | Return column numbers and letters that repeat up to NUM_REPEATS.
I.e., NUM_REPEATS = 2 would return a list of 26 * 26 = 676 2-tuples. | [
"Return",
"column",
"numbers",
"and",
"letters",
"that",
"repeat",
"up",
"to",
"NUM_REPEATS",
"."
] | 982dcf2015c80a280f1a093e32977cb71d4ea7aa | https://github.com/DallasMorningNews/django-datafreezer/blob/982dcf2015c80a280f1a093e32977cb71d4ea7aa/datafreezer/models.py#L18-L41 |
242,239 | emin63/eyap | eyap/core/redis_comments.py | RedisCommentThread.add_comment | def add_comment(self, body, allow_create=False, allow_hashes=False,
summary=None):
"""Add comment as required by comments.CommentThread parent class.
"""
thread_id = self.lookup_thread_id()
if not allow_create and not self.redis.exists(thread_id):
raise ValueError('Tried to add comment to non-exist thread %s' % (
thread_id))
comment = comments.SingleComment(
self.user, datetime.datetime.now(datetime.timezone.utc), body,
summary=summary)
lpush = self.redis.lpush(thread_id, comment.to_json())
logging.debug('Pushing comment to redis returned %s', str(lpush))
if self.ltrim:
ltrim = self.redis.ltrim(thread_id, 0, self.ltrim)
logging.debug('Redis ltrim returend %s', str(ltrim))
else:
ltrim = None
return {'status': 'OK', 'lpush': lpush, 'ltrim': ltrim} | python | def add_comment(self, body, allow_create=False, allow_hashes=False,
summary=None):
"""Add comment as required by comments.CommentThread parent class.
"""
thread_id = self.lookup_thread_id()
if not allow_create and not self.redis.exists(thread_id):
raise ValueError('Tried to add comment to non-exist thread %s' % (
thread_id))
comment = comments.SingleComment(
self.user, datetime.datetime.now(datetime.timezone.utc), body,
summary=summary)
lpush = self.redis.lpush(thread_id, comment.to_json())
logging.debug('Pushing comment to redis returned %s', str(lpush))
if self.ltrim:
ltrim = self.redis.ltrim(thread_id, 0, self.ltrim)
logging.debug('Redis ltrim returend %s', str(ltrim))
else:
ltrim = None
return {'status': 'OK', 'lpush': lpush, 'ltrim': ltrim} | [
"def",
"add_comment",
"(",
"self",
",",
"body",
",",
"allow_create",
"=",
"False",
",",
"allow_hashes",
"=",
"False",
",",
"summary",
"=",
"None",
")",
":",
"thread_id",
"=",
"self",
".",
"lookup_thread_id",
"(",
")",
"if",
"not",
"allow_create",
"and",
... | Add comment as required by comments.CommentThread parent class. | [
"Add",
"comment",
"as",
"required",
"by",
"comments",
".",
"CommentThread",
"parent",
"class",
"."
] | a610761973b478ca0e864e970be05ce29d5994a5 | https://github.com/emin63/eyap/blob/a610761973b478ca0e864e970be05ce29d5994a5/eyap/core/redis_comments.py#L25-L45 |
242,240 | CivicSpleen/ckcache | ckcache/async.py | submit_task | def submit_task(rel_path, cache_string, buffer):
"""Put an upload job on the queue, and start the thread if required"""
global upload_queue
global upload_thread
upload_queue.put((rel_path, cache_string, buffer))
if upload_thread is None or not upload_thread.is_alive():
upload_thread = UploaderThread()
upload_thread.start() | python | def submit_task(rel_path, cache_string, buffer):
"""Put an upload job on the queue, and start the thread if required"""
global upload_queue
global upload_thread
upload_queue.put((rel_path, cache_string, buffer))
if upload_thread is None or not upload_thread.is_alive():
upload_thread = UploaderThread()
upload_thread.start() | [
"def",
"submit_task",
"(",
"rel_path",
",",
"cache_string",
",",
"buffer",
")",
":",
"global",
"upload_queue",
"global",
"upload_thread",
"upload_queue",
".",
"put",
"(",
"(",
"rel_path",
",",
"cache_string",
",",
"buffer",
")",
")",
"if",
"upload_thread",
"is... | Put an upload job on the queue, and start the thread if required | [
"Put",
"an",
"upload",
"job",
"on",
"the",
"queue",
"and",
"start",
"the",
"thread",
"if",
"required"
] | 0c699b6ba97ff164e9702504f0e1643dd4cd39e1 | https://github.com/CivicSpleen/ckcache/blob/0c699b6ba97ff164e9702504f0e1643dd4cd39e1/ckcache/async.py#L53-L61 |
242,241 | herereadthis/anyone | anyone/anyone.py | Poem.get_verse | def get_verse(self, v=1):
"""Get a specific verse."""
verse_count = len(self.verses)
if v - 1 < verse_count:
return self.verses[v - 1] | python | def get_verse(self, v=1):
"""Get a specific verse."""
verse_count = len(self.verses)
if v - 1 < verse_count:
return self.verses[v - 1] | [
"def",
"get_verse",
"(",
"self",
",",
"v",
"=",
"1",
")",
":",
"verse_count",
"=",
"len",
"(",
"self",
".",
"verses",
")",
"if",
"v",
"-",
"1",
"<",
"verse_count",
":",
"return",
"self",
".",
"verses",
"[",
"v",
"-",
"1",
"]"
] | Get a specific verse. | [
"Get",
"a",
"specific",
"verse",
"."
] | ac9917014ec5d4882df89b1757416e9ceca86edb | https://github.com/herereadthis/anyone/blob/ac9917014ec5d4882df89b1757416e9ceca86edb/anyone/anyone.py#L35-L39 |
242,242 | herereadthis/anyone | anyone/anyone.py | Poem.get_line | def get_line(self, line=1):
"""Return a specific line."""
verse_size = len(self.get_verse()) + 1
if line > 1:
verse = math.floor((line - 1) / verse_size)
line_in_verse = (line - 1) % verse_size
try:
return self.verses[verse][line_in_verse]
except IndexError:
return ''
else:
return self.verses[0][0] | python | def get_line(self, line=1):
"""Return a specific line."""
verse_size = len(self.get_verse()) + 1
if line > 1:
verse = math.floor((line - 1) / verse_size)
line_in_verse = (line - 1) % verse_size
try:
return self.verses[verse][line_in_verse]
except IndexError:
return ''
else:
return self.verses[0][0] | [
"def",
"get_line",
"(",
"self",
",",
"line",
"=",
"1",
")",
":",
"verse_size",
"=",
"len",
"(",
"self",
".",
"get_verse",
"(",
")",
")",
"+",
"1",
"if",
"line",
">",
"1",
":",
"verse",
"=",
"math",
".",
"floor",
"(",
"(",
"line",
"-",
"1",
")... | Return a specific line. | [
"Return",
"a",
"specific",
"line",
"."
] | ac9917014ec5d4882df89b1757416e9ceca86edb | https://github.com/herereadthis/anyone/blob/ac9917014ec5d4882df89b1757416e9ceca86edb/anyone/anyone.py#L41-L52 |
242,243 | herereadthis/anyone | anyone/anyone.py | Poem.print_poem | def print_poem(self):
"""Print all the verses."""
for index, verse in enumerate(self.verses):
for line in verse:
print(line)
if index != len(self.verses) - 1:
print('') | python | def print_poem(self):
"""Print all the verses."""
for index, verse in enumerate(self.verses):
for line in verse:
print(line)
if index != len(self.verses) - 1:
print('') | [
"def",
"print_poem",
"(",
"self",
")",
":",
"for",
"index",
",",
"verse",
"in",
"enumerate",
"(",
"self",
".",
"verses",
")",
":",
"for",
"line",
"in",
"verse",
":",
"print",
"(",
"line",
")",
"if",
"index",
"!=",
"len",
"(",
"self",
".",
"verses",... | Print all the verses. | [
"Print",
"all",
"the",
"verses",
"."
] | ac9917014ec5d4882df89b1757416e9ceca86edb | https://github.com/herereadthis/anyone/blob/ac9917014ec5d4882df89b1757416e9ceca86edb/anyone/anyone.py#L54-L60 |
242,244 | Othernet-Project/squery-pg | squery_pg/migrations.py | get_mods | def get_mods(package):
""" List all loadable python modules in a directory
This function looks inside the specified directory for all files that look
like Python modules with a numeric prefix and returns them. It will omit
any duplicates and return file names without extension.
:param package: package object
:returns: list of tuples containing filename without extension,
major_version and minor_version
"""
pkgdir = package.__path__[0]
matches = filter(None, [PYMOD_RE.match(f) for f in os.listdir(pkgdir)])
parse_match = lambda groups: (groups[0], int(groups[1]), int(groups[2]))
return sorted(list(set([parse_match(m.groups()) for m in matches])),
key=lambda x: (x[1], x[2])) | python | def get_mods(package):
""" List all loadable python modules in a directory
This function looks inside the specified directory for all files that look
like Python modules with a numeric prefix and returns them. It will omit
any duplicates and return file names without extension.
:param package: package object
:returns: list of tuples containing filename without extension,
major_version and minor_version
"""
pkgdir = package.__path__[0]
matches = filter(None, [PYMOD_RE.match(f) for f in os.listdir(pkgdir)])
parse_match = lambda groups: (groups[0], int(groups[1]), int(groups[2]))
return sorted(list(set([parse_match(m.groups()) for m in matches])),
key=lambda x: (x[1], x[2])) | [
"def",
"get_mods",
"(",
"package",
")",
":",
"pkgdir",
"=",
"package",
".",
"__path__",
"[",
"0",
"]",
"matches",
"=",
"filter",
"(",
"None",
",",
"[",
"PYMOD_RE",
".",
"match",
"(",
"f",
")",
"for",
"f",
"in",
"os",
".",
"listdir",
"(",
"pkgdir",
... | List all loadable python modules in a directory
This function looks inside the specified directory for all files that look
like Python modules with a numeric prefix and returns them. It will omit
any duplicates and return file names without extension.
:param package: package object
:returns: list of tuples containing filename without extension,
major_version and minor_version | [
"List",
"all",
"loadable",
"python",
"modules",
"in",
"a",
"directory"
] | eaa695c3719e2d2b7e1b049bb58c987c132b6b34 | https://github.com/Othernet-Project/squery-pg/blob/eaa695c3719e2d2b7e1b049bb58c987c132b6b34/squery_pg/migrations.py#L45-L60 |
242,245 | Othernet-Project/squery-pg | squery_pg/migrations.py | get_new | def get_new(modules, min_major_version, min_minor_version):
""" Get list of migrations that haven't been run yet
:param modules: iterable containing module names
:param min_major_version: minimum major version
:param min_minor_version: minimum minor version
:returns: return an iterator that yields only items which
versions are >= min_ver
"""
for mod_data in modules:
(modname, mod_major_version, mod_minor_version) = mod_data
if (mod_major_version > min_major_version or
(mod_major_version == min_major_version and
mod_minor_version >= min_minor_version)):
yield mod_data | python | def get_new(modules, min_major_version, min_minor_version):
""" Get list of migrations that haven't been run yet
:param modules: iterable containing module names
:param min_major_version: minimum major version
:param min_minor_version: minimum minor version
:returns: return an iterator that yields only items which
versions are >= min_ver
"""
for mod_data in modules:
(modname, mod_major_version, mod_minor_version) = mod_data
if (mod_major_version > min_major_version or
(mod_major_version == min_major_version and
mod_minor_version >= min_minor_version)):
yield mod_data | [
"def",
"get_new",
"(",
"modules",
",",
"min_major_version",
",",
"min_minor_version",
")",
":",
"for",
"mod_data",
"in",
"modules",
":",
"(",
"modname",
",",
"mod_major_version",
",",
"mod_minor_version",
")",
"=",
"mod_data",
"if",
"(",
"mod_major_version",
">"... | Get list of migrations that haven't been run yet
:param modules: iterable containing module names
:param min_major_version: minimum major version
:param min_minor_version: minimum minor version
:returns: return an iterator that yields only items which
versions are >= min_ver | [
"Get",
"list",
"of",
"migrations",
"that",
"haven",
"t",
"been",
"run",
"yet"
] | eaa695c3719e2d2b7e1b049bb58c987c132b6b34 | https://github.com/Othernet-Project/squery-pg/blob/eaa695c3719e2d2b7e1b049bb58c987c132b6b34/squery_pg/migrations.py#L63-L77 |
242,246 | Othernet-Project/squery-pg | squery_pg/migrations.py | load_mod | def load_mod(module, package):
""" Load a module named ``module`` from given search``path``
The module path prefix is set according to the ``prefix`` argument.
By defualt the module is loaded as if it comes from a global
'db_migrations' package. As such, it may conflict with any 'db_migration'
package. The module can be looked up in ``sys.modules`` as
``db_migration.MODNAME`` where ``MODNAME`` is the name supplied as
``module`` argument. Keep in mind that relative imports from within the
module depend on this prefix.
This function raises an ``ImportError`` exception if module is not found.
:param module: name of the module to load
:param package: package object
:returns: module object
"""
name = '%s.%s' % (package.__name__, module)
if name in sys.modules:
return sys.modules[name]
return importlib.import_module(name, package=package.__name__) | python | def load_mod(module, package):
""" Load a module named ``module`` from given search``path``
The module path prefix is set according to the ``prefix`` argument.
By defualt the module is loaded as if it comes from a global
'db_migrations' package. As such, it may conflict with any 'db_migration'
package. The module can be looked up in ``sys.modules`` as
``db_migration.MODNAME`` where ``MODNAME`` is the name supplied as
``module`` argument. Keep in mind that relative imports from within the
module depend on this prefix.
This function raises an ``ImportError`` exception if module is not found.
:param module: name of the module to load
:param package: package object
:returns: module object
"""
name = '%s.%s' % (package.__name__, module)
if name in sys.modules:
return sys.modules[name]
return importlib.import_module(name, package=package.__name__) | [
"def",
"load_mod",
"(",
"module",
",",
"package",
")",
":",
"name",
"=",
"'%s.%s'",
"%",
"(",
"package",
".",
"__name__",
",",
"module",
")",
"if",
"name",
"in",
"sys",
".",
"modules",
":",
"return",
"sys",
".",
"modules",
"[",
"name",
"]",
"return",... | Load a module named ``module`` from given search``path``
The module path prefix is set according to the ``prefix`` argument.
By defualt the module is loaded as if it comes from a global
'db_migrations' package. As such, it may conflict with any 'db_migration'
package. The module can be looked up in ``sys.modules`` as
``db_migration.MODNAME`` where ``MODNAME`` is the name supplied as
``module`` argument. Keep in mind that relative imports from within the
module depend on this prefix.
This function raises an ``ImportError`` exception if module is not found.
:param module: name of the module to load
:param package: package object
:returns: module object | [
"Load",
"a",
"module",
"named",
"module",
"from",
"given",
"search",
"path"
] | eaa695c3719e2d2b7e1b049bb58c987c132b6b34 | https://github.com/Othernet-Project/squery-pg/blob/eaa695c3719e2d2b7e1b049bb58c987c132b6b34/squery_pg/migrations.py#L80-L100 |
242,247 | Othernet-Project/squery-pg | squery_pg/migrations.py | unpack_version | def unpack_version(version):
"""Unpack a single version integer into the two major and minor
components."""
minor_version = version % VERSION_MULTIPLIER
major_version = (version - minor_version) / VERSION_MULTIPLIER
return (major_version, minor_version) | python | def unpack_version(version):
"""Unpack a single version integer into the two major and minor
components."""
minor_version = version % VERSION_MULTIPLIER
major_version = (version - minor_version) / VERSION_MULTIPLIER
return (major_version, minor_version) | [
"def",
"unpack_version",
"(",
"version",
")",
":",
"minor_version",
"=",
"version",
"%",
"VERSION_MULTIPLIER",
"major_version",
"=",
"(",
"version",
"-",
"minor_version",
")",
"/",
"VERSION_MULTIPLIER",
"return",
"(",
"major_version",
",",
"minor_version",
")"
] | Unpack a single version integer into the two major and minor
components. | [
"Unpack",
"a",
"single",
"version",
"integer",
"into",
"the",
"two",
"major",
"and",
"minor",
"components",
"."
] | eaa695c3719e2d2b7e1b049bb58c987c132b6b34 | https://github.com/Othernet-Project/squery-pg/blob/eaa695c3719e2d2b7e1b049bb58c987c132b6b34/squery_pg/migrations.py#L108-L113 |
242,248 | Othernet-Project/squery-pg | squery_pg/migrations.py | set_version | def set_version(db, name, major_version, minor_version):
""" Set database migration version
:param db: connetion object
:param name: associated name
:param major_version: integer major version of migration
:param minor_version: integer minor version of migration
"""
version = pack_version(major_version, minor_version)
db.execute(SET_VERSION_SQL, dict(name=name, version=version)) | python | def set_version(db, name, major_version, minor_version):
""" Set database migration version
:param db: connetion object
:param name: associated name
:param major_version: integer major version of migration
:param minor_version: integer minor version of migration
"""
version = pack_version(major_version, minor_version)
db.execute(SET_VERSION_SQL, dict(name=name, version=version)) | [
"def",
"set_version",
"(",
"db",
",",
"name",
",",
"major_version",
",",
"minor_version",
")",
":",
"version",
"=",
"pack_version",
"(",
"major_version",
",",
"minor_version",
")",
"db",
".",
"execute",
"(",
"SET_VERSION_SQL",
",",
"dict",
"(",
"name",
"=",
... | Set database migration version
:param db: connetion object
:param name: associated name
:param major_version: integer major version of migration
:param minor_version: integer minor version of migration | [
"Set",
"database",
"migration",
"version"
] | eaa695c3719e2d2b7e1b049bb58c987c132b6b34 | https://github.com/Othernet-Project/squery-pg/blob/eaa695c3719e2d2b7e1b049bb58c987c132b6b34/squery_pg/migrations.py#L146-L155 |
242,249 | Othernet-Project/squery-pg | squery_pg/migrations.py | run_migration | def run_migration(name, major_version, minor_version, db, mod, conf={}):
""" Run migration script
:param major_version: major version number of the migration
:param minor_version: minor version number of the migration
:param db: database connection object
:param path: path of the migration script
:param conf: application configuration (if any)
"""
with db.transaction():
mod.up(db, conf)
set_version(db, name, major_version, minor_version) | python | def run_migration(name, major_version, minor_version, db, mod, conf={}):
""" Run migration script
:param major_version: major version number of the migration
:param minor_version: minor version number of the migration
:param db: database connection object
:param path: path of the migration script
:param conf: application configuration (if any)
"""
with db.transaction():
mod.up(db, conf)
set_version(db, name, major_version, minor_version) | [
"def",
"run_migration",
"(",
"name",
",",
"major_version",
",",
"minor_version",
",",
"db",
",",
"mod",
",",
"conf",
"=",
"{",
"}",
")",
":",
"with",
"db",
".",
"transaction",
"(",
")",
":",
"mod",
".",
"up",
"(",
"db",
",",
"conf",
")",
"set_versi... | Run migration script
:param major_version: major version number of the migration
:param minor_version: minor version number of the migration
:param db: database connection object
:param path: path of the migration script
:param conf: application configuration (if any) | [
"Run",
"migration",
"script"
] | eaa695c3719e2d2b7e1b049bb58c987c132b6b34 | https://github.com/Othernet-Project/squery-pg/blob/eaa695c3719e2d2b7e1b049bb58c987c132b6b34/squery_pg/migrations.py#L158-L169 |
242,250 | Othernet-Project/squery-pg | squery_pg/migrations.py | migrate | def migrate(db, name, package, conf={}):
""" Run all migrations that have not been run
Migrations will be run inside a transaction.
:param db: database connection object
:param name: name associated with the migrations
:param package: package that contains the migrations
:param conf: application configuration object
"""
(current_major_version, current_minor_version) = get_version(db, name)
package = importlib.import_module(package)
logging.debug('Migration version for %s is %s.%s',
package.__name__,
current_major_version,
current_minor_version)
mods = get_mods(package)
migrations = get_new(mods,
current_major_version,
current_minor_version + 1)
for (modname, major_version, minor_version) in migrations:
mod = load_mod(modname, package)
run_migration(name, major_version, minor_version, db, mod, conf)
logging.debug("Finished migrating to %s", modname) | python | def migrate(db, name, package, conf={}):
""" Run all migrations that have not been run
Migrations will be run inside a transaction.
:param db: database connection object
:param name: name associated with the migrations
:param package: package that contains the migrations
:param conf: application configuration object
"""
(current_major_version, current_minor_version) = get_version(db, name)
package = importlib.import_module(package)
logging.debug('Migration version for %s is %s.%s',
package.__name__,
current_major_version,
current_minor_version)
mods = get_mods(package)
migrations = get_new(mods,
current_major_version,
current_minor_version + 1)
for (modname, major_version, minor_version) in migrations:
mod = load_mod(modname, package)
run_migration(name, major_version, minor_version, db, mod, conf)
logging.debug("Finished migrating to %s", modname) | [
"def",
"migrate",
"(",
"db",
",",
"name",
",",
"package",
",",
"conf",
"=",
"{",
"}",
")",
":",
"(",
"current_major_version",
",",
"current_minor_version",
")",
"=",
"get_version",
"(",
"db",
",",
"name",
")",
"package",
"=",
"importlib",
".",
"import_mo... | Run all migrations that have not been run
Migrations will be run inside a transaction.
:param db: database connection object
:param name: name associated with the migrations
:param package: package that contains the migrations
:param conf: application configuration object | [
"Run",
"all",
"migrations",
"that",
"have",
"not",
"been",
"run"
] | eaa695c3719e2d2b7e1b049bb58c987c132b6b34 | https://github.com/Othernet-Project/squery-pg/blob/eaa695c3719e2d2b7e1b049bb58c987c132b6b34/squery_pg/migrations.py#L172-L195 |
242,251 | mayfield/shellish | shellish/rendering/html.py | html2vtml | def html2vtml(vtmarkup):
""" Convert hypertext markup into vt markup.
The output can be given to `vtmlrender` for converstion to VT100
sequences. """
try:
htmlconv.feed(vtmarkup)
htmlconv.close()
return htmlconv.getvalue()
finally:
htmlconv.reset() | python | def html2vtml(vtmarkup):
""" Convert hypertext markup into vt markup.
The output can be given to `vtmlrender` for converstion to VT100
sequences. """
try:
htmlconv.feed(vtmarkup)
htmlconv.close()
return htmlconv.getvalue()
finally:
htmlconv.reset() | [
"def",
"html2vtml",
"(",
"vtmarkup",
")",
":",
"try",
":",
"htmlconv",
".",
"feed",
"(",
"vtmarkup",
")",
"htmlconv",
".",
"close",
"(",
")",
"return",
"htmlconv",
".",
"getvalue",
"(",
")",
"finally",
":",
"htmlconv",
".",
"reset",
"(",
")"
] | Convert hypertext markup into vt markup.
The output can be given to `vtmlrender` for converstion to VT100
sequences. | [
"Convert",
"hypertext",
"markup",
"into",
"vt",
"markup",
".",
"The",
"output",
"can",
"be",
"given",
"to",
"vtmlrender",
"for",
"converstion",
"to",
"VT100",
"sequences",
"."
] | df0f0e4612d138c34d8cb99b66ab5b8e47f1414a | https://github.com/mayfield/shellish/blob/df0f0e4612d138c34d8cb99b66ab5b8e47f1414a/shellish/rendering/html.py#L206-L215 |
242,252 | OiNutter/lean | lean/template.py | Template.render | def render(self,scope=None,local_vars=None,block=None):
''' Render the template in the given scope with the locals specified. If a
block is given, it is typically available within the template via
+yield+.
'''
if not scope:
class Scope(object):
pass
scope = Scope()
return self.evaluate(scope,local_vars or {}, block) | python | def render(self,scope=None,local_vars=None,block=None):
''' Render the template in the given scope with the locals specified. If a
block is given, it is typically available within the template via
+yield+.
'''
if not scope:
class Scope(object):
pass
scope = Scope()
return self.evaluate(scope,local_vars or {}, block) | [
"def",
"render",
"(",
"self",
",",
"scope",
"=",
"None",
",",
"local_vars",
"=",
"None",
",",
"block",
"=",
"None",
")",
":",
"if",
"not",
"scope",
":",
"class",
"Scope",
"(",
"object",
")",
":",
"pass",
"scope",
"=",
"Scope",
"(",
")",
"return",
... | Render the template in the given scope with the locals specified. If a
block is given, it is typically available within the template via
+yield+. | [
"Render",
"the",
"template",
"in",
"the",
"given",
"scope",
"with",
"the",
"locals",
"specified",
".",
"If",
"a",
"block",
"is",
"given",
"it",
"is",
"typically",
"available",
"within",
"the",
"template",
"via",
"+",
"yield",
"+",
"."
] | 5d251f923acd44265ed401de14a9ead6752c543f | https://github.com/OiNutter/lean/blob/5d251f923acd44265ed401de14a9ead6752c543f/lean/template.py#L66-L78 |
242,253 | OiNutter/lean | lean/template.py | Template.basename | def basename(self, suffix=''):
''' The basename of the template file.'''
return os.path.basename(self._file, suffix) if self._file else None | python | def basename(self, suffix=''):
''' The basename of the template file.'''
return os.path.basename(self._file, suffix) if self._file else None | [
"def",
"basename",
"(",
"self",
",",
"suffix",
"=",
"''",
")",
":",
"return",
"os",
".",
"path",
".",
"basename",
"(",
"self",
".",
"_file",
",",
"suffix",
")",
"if",
"self",
".",
"_file",
"else",
"None"
] | The basename of the template file. | [
"The",
"basename",
"of",
"the",
"template",
"file",
"."
] | 5d251f923acd44265ed401de14a9ead6752c543f | https://github.com/OiNutter/lean/blob/5d251f923acd44265ed401de14a9ead6752c543f/lean/template.py#L80-L82 |
242,254 | firstprayer/monsql | monsql/queryset.py | QuerySet.distinct | def distinct(self):
"""
Only return distinct row.
Return a new query set with distinct mark
"""
new_query_set = self.clone()
new_query_set.query.distinct = True
return new_query_set | python | def distinct(self):
"""
Only return distinct row.
Return a new query set with distinct mark
"""
new_query_set = self.clone()
new_query_set.query.distinct = True
return new_query_set | [
"def",
"distinct",
"(",
"self",
")",
":",
"new_query_set",
"=",
"self",
".",
"clone",
"(",
")",
"new_query_set",
".",
"query",
".",
"distinct",
"=",
"True",
"return",
"new_query_set"
] | Only return distinct row.
Return a new query set with distinct mark | [
"Only",
"return",
"distinct",
"row",
".",
"Return",
"a",
"new",
"query",
"set",
"with",
"distinct",
"mark"
] | 6285c15b574c8664046eae2edfeb548c7b173efd | https://github.com/firstprayer/monsql/blob/6285c15b574c8664046eae2edfeb548c7b173efd/monsql/queryset.py#L110-L117 |
242,255 | uw-it-aca/uw-restclients-grad | uw_grad/degree.py | _process_json | def _process_json(json_data):
"""
return a list of GradDegree objects.
"""
requests = []
for item in json_data:
degree = GradDegree()
degree.degree_title = item["degreeTitle"]
degree.exam_place = item["examPlace"]
degree.exam_date = parse_datetime(item.get("examDate"))
degree.req_type = item["requestType"]
degree.major_full_name = item["majorFullName"]
degree.submit_date = parse_datetime(item.get("requestSubmitDate"))
degree.decision_date = parse_datetime(item.get('decisionDate'))
degree.status = item["status"]
degree.target_award_year = item["targetAwardYear"]
if item.get("targetAwardQuarter")and\
len(item.get("targetAwardQuarter")):
degree.target_award_quarter = item["targetAwardQuarter"].lower()
requests.append(degree)
return requests | python | def _process_json(json_data):
"""
return a list of GradDegree objects.
"""
requests = []
for item in json_data:
degree = GradDegree()
degree.degree_title = item["degreeTitle"]
degree.exam_place = item["examPlace"]
degree.exam_date = parse_datetime(item.get("examDate"))
degree.req_type = item["requestType"]
degree.major_full_name = item["majorFullName"]
degree.submit_date = parse_datetime(item.get("requestSubmitDate"))
degree.decision_date = parse_datetime(item.get('decisionDate'))
degree.status = item["status"]
degree.target_award_year = item["targetAwardYear"]
if item.get("targetAwardQuarter")and\
len(item.get("targetAwardQuarter")):
degree.target_award_quarter = item["targetAwardQuarter"].lower()
requests.append(degree)
return requests | [
"def",
"_process_json",
"(",
"json_data",
")",
":",
"requests",
"=",
"[",
"]",
"for",
"item",
"in",
"json_data",
":",
"degree",
"=",
"GradDegree",
"(",
")",
"degree",
".",
"degree_title",
"=",
"item",
"[",
"\"degreeTitle\"",
"]",
"degree",
".",
"exam_place... | return a list of GradDegree objects. | [
"return",
"a",
"list",
"of",
"GradDegree",
"objects",
"."
] | ca06ed2f24f3683314a5690f6078e97d37fc8e52 | https://github.com/uw-it-aca/uw-restclients-grad/blob/ca06ed2f24f3683314a5690f6078e97d37fc8e52/uw_grad/degree.py#L23-L44 |
242,256 | ryanjdillon/pylleo | pylleo/lleoio.py | read_meta | def read_meta(path_dir, tag_model, tag_id):
'''Read meta data from Little Leonardo data header rows
Args
----
path_dir: str
Parent directory containing lleo data files
tag_model: str
Little Leonardo tag model name
tag_id: str, int
Little Leonardo tag ID number
Returns
-------
meta: dict
dictionary with meta data from header lines of lleo data files
'''
from collections import OrderedDict
import os
import yamlord
from . import utils
def _parse_meta_line(line):
'''Return key, value pair parsed from data header line'''
# Parse the key and its value from the line
key, val = line.replace(':', '').replace('"', '').split(',')
return key.strip(), val.strip()
def _read_meta_all(f, meta, n_header):
'''Read all meta data from header rows of data file'''
# Skip 'File name' line
f.seek(0)
_ = f.readline()
# Create child dictionary for channel / file
line = f.readline()
key_ch, val_ch = _parse_meta_line(line)
val_ch = utils.posix_string(val_ch)
meta['parameters'][val_ch] = OrderedDict()
# Write header values to channel dict
for _ in range(n_header-2):
line = f.readline()
key, val = _parse_meta_line(line)
meta['parameters'][val_ch][key] = val.strip()
return meta
def _create_meta(path_dir, tag_model, tag_id):
'''Create meta data dictionary'''
import datetime
from . import utils
param_strs = utils.get_tag_params(tag_model)
# Create dictionary of meta data
meta = OrderedDict()
# Create fields for the parameters in data directory name
exp_name = os.path.split(path_dir)[1]
params_tag = utils.parse_experiment_params(exp_name)
for key, value in params_tag.items():
meta[key] = value
fmt = "%Y-%m-%d %H:%M:%S"
meta['date_modified'] = datetime.datetime.now().strftime(fmt)
meta['parameters'] = OrderedDict()
for param_str in param_strs:
print('Create meta entry for {}'.format(param_str))
path_file = utils.find_file(path_dir, param_str, '.TXT')
# Get number of header rows
enc = utils.predict_encoding(path_file, n_lines=20)
with open(path_file, 'r', encoding=enc) as f:
n_header = utils.get_n_header(f)
f.seek(0)
meta = _read_meta_all(f, meta, n_header=n_header)
return meta
# Load meta data from YAML file if it already exists
meta_yaml_path = os.path.join(path_dir, 'meta.yml')
# Load file if exists else create
if os.path.isfile(meta_yaml_path):
meta = yamlord.read_yaml(meta_yaml_path)
# Else create meta dictionary and save to YAML
else:
meta = _create_meta(path_dir, tag_model, tag_id)
yamlord.write_yaml(meta, meta_yaml_path)
return meta | python | def read_meta(path_dir, tag_model, tag_id):
'''Read meta data from Little Leonardo data header rows
Args
----
path_dir: str
Parent directory containing lleo data files
tag_model: str
Little Leonardo tag model name
tag_id: str, int
Little Leonardo tag ID number
Returns
-------
meta: dict
dictionary with meta data from header lines of lleo data files
'''
from collections import OrderedDict
import os
import yamlord
from . import utils
def _parse_meta_line(line):
'''Return key, value pair parsed from data header line'''
# Parse the key and its value from the line
key, val = line.replace(':', '').replace('"', '').split(',')
return key.strip(), val.strip()
def _read_meta_all(f, meta, n_header):
'''Read all meta data from header rows of data file'''
# Skip 'File name' line
f.seek(0)
_ = f.readline()
# Create child dictionary for channel / file
line = f.readline()
key_ch, val_ch = _parse_meta_line(line)
val_ch = utils.posix_string(val_ch)
meta['parameters'][val_ch] = OrderedDict()
# Write header values to channel dict
for _ in range(n_header-2):
line = f.readline()
key, val = _parse_meta_line(line)
meta['parameters'][val_ch][key] = val.strip()
return meta
def _create_meta(path_dir, tag_model, tag_id):
'''Create meta data dictionary'''
import datetime
from . import utils
param_strs = utils.get_tag_params(tag_model)
# Create dictionary of meta data
meta = OrderedDict()
# Create fields for the parameters in data directory name
exp_name = os.path.split(path_dir)[1]
params_tag = utils.parse_experiment_params(exp_name)
for key, value in params_tag.items():
meta[key] = value
fmt = "%Y-%m-%d %H:%M:%S"
meta['date_modified'] = datetime.datetime.now().strftime(fmt)
meta['parameters'] = OrderedDict()
for param_str in param_strs:
print('Create meta entry for {}'.format(param_str))
path_file = utils.find_file(path_dir, param_str, '.TXT')
# Get number of header rows
enc = utils.predict_encoding(path_file, n_lines=20)
with open(path_file, 'r', encoding=enc) as f:
n_header = utils.get_n_header(f)
f.seek(0)
meta = _read_meta_all(f, meta, n_header=n_header)
return meta
# Load meta data from YAML file if it already exists
meta_yaml_path = os.path.join(path_dir, 'meta.yml')
# Load file if exists else create
if os.path.isfile(meta_yaml_path):
meta = yamlord.read_yaml(meta_yaml_path)
# Else create meta dictionary and save to YAML
else:
meta = _create_meta(path_dir, tag_model, tag_id)
yamlord.write_yaml(meta, meta_yaml_path)
return meta | [
"def",
"read_meta",
"(",
"path_dir",
",",
"tag_model",
",",
"tag_id",
")",
":",
"from",
"collections",
"import",
"OrderedDict",
"import",
"os",
"import",
"yamlord",
"from",
".",
"import",
"utils",
"def",
"_parse_meta_line",
"(",
"line",
")",
":",
"'''Return ke... | Read meta data from Little Leonardo data header rows
Args
----
path_dir: str
Parent directory containing lleo data files
tag_model: str
Little Leonardo tag model name
tag_id: str, int
Little Leonardo tag ID number
Returns
-------
meta: dict
dictionary with meta data from header lines of lleo data files | [
"Read",
"meta",
"data",
"from",
"Little",
"Leonardo",
"data",
"header",
"rows"
] | b9b999fef19eaeccce4f207ab1b6198287c1bfec | https://github.com/ryanjdillon/pylleo/blob/b9b999fef19eaeccce4f207ab1b6198287c1bfec/pylleo/lleoio.py#L2-L103 |
242,257 | ikumen/flask-cfg | flask_cfg/core.py | AbstractConfig._merge_values | def _merge_values(self, to_values, from_values):
"""Merges two dictionaries of values recursively. This is a very naive
implementation that expects the two dictionaries to be fairly similar
in structure.
@param to_values destination dictionary
@param from_values dictionary with values to copy
"""
if from_values is not None:
for k, v in from_values.items():
if k in to_values and isinstance(to_values[k], dict):
self._merge_values(to_values[k], v) # merge
else:
to_values[k] = v # replaces instead of merge
return to_values | python | def _merge_values(self, to_values, from_values):
"""Merges two dictionaries of values recursively. This is a very naive
implementation that expects the two dictionaries to be fairly similar
in structure.
@param to_values destination dictionary
@param from_values dictionary with values to copy
"""
if from_values is not None:
for k, v in from_values.items():
if k in to_values and isinstance(to_values[k], dict):
self._merge_values(to_values[k], v) # merge
else:
to_values[k] = v # replaces instead of merge
return to_values | [
"def",
"_merge_values",
"(",
"self",
",",
"to_values",
",",
"from_values",
")",
":",
"if",
"from_values",
"is",
"not",
"None",
":",
"for",
"k",
",",
"v",
"in",
"from_values",
".",
"items",
"(",
")",
":",
"if",
"k",
"in",
"to_values",
"and",
"isinstance... | Merges two dictionaries of values recursively. This is a very naive
implementation that expects the two dictionaries to be fairly similar
in structure.
@param to_values destination dictionary
@param from_values dictionary with values to copy | [
"Merges",
"two",
"dictionaries",
"of",
"values",
"recursively",
".",
"This",
"is",
"a",
"very",
"naive",
"implementation",
"that",
"expects",
"the",
"two",
"dictionaries",
"to",
"be",
"fairly",
"similar",
"in",
"structure",
"."
] | 28f3370121419d22b6a5a3713ab3cb8bb2da6e43 | https://github.com/ikumen/flask-cfg/blob/28f3370121419d22b6a5a3713ab3cb8bb2da6e43/flask_cfg/core.py#L160-L174 |
242,258 | ikumen/flask-cfg | flask_cfg/core.py | AbstractConfig._load_config | def _load_config(self, path):
"""Return YAML values from given config file.
@param path file to load
"""
try:
with open(path) as f:
values = yaml.safe_load(f)
if isinstance(values, dict):
return values
else:
raise yaml.YAMLError('Unable to parse/load {}'.format(path))
except(IOError, yaml.YAMLError) as e:
if self.ignore_errors:
return None
else:
raise e | python | def _load_config(self, path):
"""Return YAML values from given config file.
@param path file to load
"""
try:
with open(path) as f:
values = yaml.safe_load(f)
if isinstance(values, dict):
return values
else:
raise yaml.YAMLError('Unable to parse/load {}'.format(path))
except(IOError, yaml.YAMLError) as e:
if self.ignore_errors:
return None
else:
raise e | [
"def",
"_load_config",
"(",
"self",
",",
"path",
")",
":",
"try",
":",
"with",
"open",
"(",
"path",
")",
"as",
"f",
":",
"values",
"=",
"yaml",
".",
"safe_load",
"(",
"f",
")",
"if",
"isinstance",
"(",
"values",
",",
"dict",
")",
":",
"return",
"... | Return YAML values from given config file.
@param path file to load | [
"Return",
"YAML",
"values",
"from",
"given",
"config",
"file",
"."
] | 28f3370121419d22b6a5a3713ab3cb8bb2da6e43 | https://github.com/ikumen/flask-cfg/blob/28f3370121419d22b6a5a3713ab3cb8bb2da6e43/flask_cfg/core.py#L177-L193 |
242,259 | ikumen/flask-cfg | flask_cfg/core.py | AbstractConfig._normalize_file_paths | def _normalize_file_paths(self, *args):
"""Returns all given configuration file paths as one list."""
paths = []
for arg in args:
if arg is None:
continue
elif self._is_valid_file(arg):
paths.append(arg)
elif isinstance(arg, list) and all(self._is_valid_file(_) for _ in arg):
paths = paths + arg
elif not self.ignore_errors:
raise TypeError('Config file paths must be string path or list of paths!')
return paths | python | def _normalize_file_paths(self, *args):
"""Returns all given configuration file paths as one list."""
paths = []
for arg in args:
if arg is None:
continue
elif self._is_valid_file(arg):
paths.append(arg)
elif isinstance(arg, list) and all(self._is_valid_file(_) for _ in arg):
paths = paths + arg
elif not self.ignore_errors:
raise TypeError('Config file paths must be string path or list of paths!')
return paths | [
"def",
"_normalize_file_paths",
"(",
"self",
",",
"*",
"args",
")",
":",
"paths",
"=",
"[",
"]",
"for",
"arg",
"in",
"args",
":",
"if",
"arg",
"is",
"None",
":",
"continue",
"elif",
"self",
".",
"_is_valid_file",
"(",
"arg",
")",
":",
"paths",
".",
... | Returns all given configuration file paths as one list. | [
"Returns",
"all",
"given",
"configuration",
"file",
"paths",
"as",
"one",
"list",
"."
] | 28f3370121419d22b6a5a3713ab3cb8bb2da6e43 | https://github.com/ikumen/flask-cfg/blob/28f3370121419d22b6a5a3713ab3cb8bb2da6e43/flask_cfg/core.py#L196-L208 |
242,260 | ikumen/flask-cfg | flask_cfg/core.py | AbstractConfig._is_valid_file | def _is_valid_file(self, path):
"""Simple check to see if file path exists. Does not check for valid YAML format."""
return isinstance(path, basestring) and os.path.isfile(path) | python | def _is_valid_file(self, path):
"""Simple check to see if file path exists. Does not check for valid YAML format."""
return isinstance(path, basestring) and os.path.isfile(path) | [
"def",
"_is_valid_file",
"(",
"self",
",",
"path",
")",
":",
"return",
"isinstance",
"(",
"path",
",",
"basestring",
")",
"and",
"os",
".",
"path",
".",
"isfile",
"(",
"path",
")"
] | Simple check to see if file path exists. Does not check for valid YAML format. | [
"Simple",
"check",
"to",
"see",
"if",
"file",
"path",
"exists",
".",
"Does",
"not",
"check",
"for",
"valid",
"YAML",
"format",
"."
] | 28f3370121419d22b6a5a3713ab3cb8bb2da6e43 | https://github.com/ikumen/flask-cfg/blob/28f3370121419d22b6a5a3713ab3cb8bb2da6e43/flask_cfg/core.py#L211-L213 |
242,261 | almcc/cinder-data | docs/conf.py | run_apidoc | def run_apidoc(_):
"""Heler function for run apidoc as part of the build."""
current_directory = os.path.abspath(os.path.dirname(__file__))
output_path = os.path.join(current_directory, 'source')
cmd_path = 'sphinx-apidoc'
if hasattr(sys, 'real_prefix'): # Check to see if we are in a virtualenv
# If we are, assemble the path manually
cmd_path = os.path.abspath(os.path.join(sys.prefix, 'bin', 'sphinx-apidoc'))
main([cmd_path, '-e', '-o', output_path, '../cinder_data', '--force']) | python | def run_apidoc(_):
"""Heler function for run apidoc as part of the build."""
current_directory = os.path.abspath(os.path.dirname(__file__))
output_path = os.path.join(current_directory, 'source')
cmd_path = 'sphinx-apidoc'
if hasattr(sys, 'real_prefix'): # Check to see if we are in a virtualenv
# If we are, assemble the path manually
cmd_path = os.path.abspath(os.path.join(sys.prefix, 'bin', 'sphinx-apidoc'))
main([cmd_path, '-e', '-o', output_path, '../cinder_data', '--force']) | [
"def",
"run_apidoc",
"(",
"_",
")",
":",
"current_directory",
"=",
"os",
".",
"path",
".",
"abspath",
"(",
"os",
".",
"path",
".",
"dirname",
"(",
"__file__",
")",
")",
"output_path",
"=",
"os",
".",
"path",
".",
"join",
"(",
"current_directory",
",",
... | Heler function for run apidoc as part of the build. | [
"Heler",
"function",
"for",
"run",
"apidoc",
"as",
"part",
"of",
"the",
"build",
"."
] | 4159a5186c4b4fc32354749892e86130530f6ec5 | https://github.com/almcc/cinder-data/blob/4159a5186c4b4fc32354749892e86130530f6ec5/docs/conf.py#L23-L31 |
242,262 | mayfield/shellish | shellish/layout/column.py | columnize | def columnize(items, width=None, file=sys.stdout):
""" Smart display width handling when showing a list of stuff. """
if not items:
return
if width is None:
width = shutil.get_terminal_size()[0] if file is sys.stdout else 80
items = [rendering.vtmlrender(x) for x in items]
maxcol = max(items, key=len)
colsize = len(maxcol) + 2
cols = width // colsize
if cols < 2:
for x in items:
print(x, file=file)
return
lines = math.ceil(len(items) / cols)
for i in range(lines):
row = items[i:None:lines]
print(*[x.ljust(colsize) for x in row], sep='', file=file) | python | def columnize(items, width=None, file=sys.stdout):
""" Smart display width handling when showing a list of stuff. """
if not items:
return
if width is None:
width = shutil.get_terminal_size()[0] if file is sys.stdout else 80
items = [rendering.vtmlrender(x) for x in items]
maxcol = max(items, key=len)
colsize = len(maxcol) + 2
cols = width // colsize
if cols < 2:
for x in items:
print(x, file=file)
return
lines = math.ceil(len(items) / cols)
for i in range(lines):
row = items[i:None:lines]
print(*[x.ljust(colsize) for x in row], sep='', file=file) | [
"def",
"columnize",
"(",
"items",
",",
"width",
"=",
"None",
",",
"file",
"=",
"sys",
".",
"stdout",
")",
":",
"if",
"not",
"items",
":",
"return",
"if",
"width",
"is",
"None",
":",
"width",
"=",
"shutil",
".",
"get_terminal_size",
"(",
")",
"[",
"... | Smart display width handling when showing a list of stuff. | [
"Smart",
"display",
"width",
"handling",
"when",
"showing",
"a",
"list",
"of",
"stuff",
"."
] | df0f0e4612d138c34d8cb99b66ab5b8e47f1414a | https://github.com/mayfield/shellish/blob/df0f0e4612d138c34d8cb99b66ab5b8e47f1414a/shellish/layout/column.py#L11-L28 |
242,263 | drongo-framework/drongo | drongo/managers/url.py | UrlManager.find_call | def find_call(self, path, method):
"""Find callable for the specified URL path and HTTP method.
Args:
path (:obj:`str`): URL path to match
method (:obj:`str`): HTTP method
Note:
A trailing '/' is always assumed in the path.
"""
if not path.endswith('/'):
path += '/'
path = path.split('/')[1:]
return self._recursive_route_match(self._routes, path, method, []) | python | def find_call(self, path, method):
"""Find callable for the specified URL path and HTTP method.
Args:
path (:obj:`str`): URL path to match
method (:obj:`str`): HTTP method
Note:
A trailing '/' is always assumed in the path.
"""
if not path.endswith('/'):
path += '/'
path = path.split('/')[1:]
return self._recursive_route_match(self._routes, path, method, []) | [
"def",
"find_call",
"(",
"self",
",",
"path",
",",
"method",
")",
":",
"if",
"not",
"path",
".",
"endswith",
"(",
"'/'",
")",
":",
"path",
"+=",
"'/'",
"path",
"=",
"path",
".",
"split",
"(",
"'/'",
")",
"[",
"1",
":",
"]",
"return",
"self",
".... | Find callable for the specified URL path and HTTP method.
Args:
path (:obj:`str`): URL path to match
method (:obj:`str`): HTTP method
Note:
A trailing '/' is always assumed in the path. | [
"Find",
"callable",
"for",
"the",
"specified",
"URL",
"path",
"and",
"HTTP",
"method",
"."
] | 487edb370ae329f370bcf3b433ed3f28ba4c1d8c | https://github.com/drongo-framework/drongo/blob/487edb370ae329f370bcf3b433ed3f28ba4c1d8c/drongo/managers/url.py#L50-L63 |
242,264 | edwards-lab/libGWAS | libgwas/standardizer.py | StandardizedVariable.get_variables | def get_variables(self, missing_in_geno=None):
"""Extract the complete set of data based on missingness over all
for the current locus.
:param missing_in_geno: mask associated with missingness in genotype
:return: (phenotypes, covariates, nonmissing used for this set of vars)
"""
count = 0
mismatch = 0
if missing_in_geno is None:
nonmissing = numpy.invert(self.missing[self.idx])
else:
nonmissing = numpy.invert(self.missing[self.idx] | missing_in_geno)
nmcount = sum(nonmissing)
covars = numpy.zeros((self.covar_count, nmcount))
for idx in range(0, self.covar_count):
covars[idx] = self.covariates[idx][nonmissing]
min = covars[idx][covars[idx] != pheno_covar.PhenoCovar.missing_encoding].min()
max = covars[idx][covars[idx] != pheno_covar.PhenoCovar.missing_encoding].max()
if min == max:
raise InvariantVar("Covar %s doesn't have enough variation to continue" % (self.datasource.covariate_labels[idx]))
min = self.phenotypes[self.idx][nonmissing].min()
max = self.phenotypes[self.idx][nonmissing].max()
if min == max:
raise InvariantVar("Phenotype %s doesn't have enough variation to continue" % (self.datasource.phenotype_names[self.idx]))
return (self.phenotypes[self.idx][nonmissing], covars, nonmissing) | python | def get_variables(self, missing_in_geno=None):
"""Extract the complete set of data based on missingness over all
for the current locus.
:param missing_in_geno: mask associated with missingness in genotype
:return: (phenotypes, covariates, nonmissing used for this set of vars)
"""
count = 0
mismatch = 0
if missing_in_geno is None:
nonmissing = numpy.invert(self.missing[self.idx])
else:
nonmissing = numpy.invert(self.missing[self.idx] | missing_in_geno)
nmcount = sum(nonmissing)
covars = numpy.zeros((self.covar_count, nmcount))
for idx in range(0, self.covar_count):
covars[idx] = self.covariates[idx][nonmissing]
min = covars[idx][covars[idx] != pheno_covar.PhenoCovar.missing_encoding].min()
max = covars[idx][covars[idx] != pheno_covar.PhenoCovar.missing_encoding].max()
if min == max:
raise InvariantVar("Covar %s doesn't have enough variation to continue" % (self.datasource.covariate_labels[idx]))
min = self.phenotypes[self.idx][nonmissing].min()
max = self.phenotypes[self.idx][nonmissing].max()
if min == max:
raise InvariantVar("Phenotype %s doesn't have enough variation to continue" % (self.datasource.phenotype_names[self.idx]))
return (self.phenotypes[self.idx][nonmissing], covars, nonmissing) | [
"def",
"get_variables",
"(",
"self",
",",
"missing_in_geno",
"=",
"None",
")",
":",
"count",
"=",
"0",
"mismatch",
"=",
"0",
"if",
"missing_in_geno",
"is",
"None",
":",
"nonmissing",
"=",
"numpy",
".",
"invert",
"(",
"self",
".",
"missing",
"[",
"self",
... | Extract the complete set of data based on missingness over all
for the current locus.
:param missing_in_geno: mask associated with missingness in genotype
:return: (phenotypes, covariates, nonmissing used for this set of vars) | [
"Extract",
"the",
"complete",
"set",
"of",
"data",
"based",
"on",
"missingness",
"over",
"all",
"for",
"the",
"current",
"locus",
"."
] | d68c9a083d443dfa5d7c5112de29010909cfe23f | https://github.com/edwards-lab/libGWAS/blob/d68c9a083d443dfa5d7c5112de29010909cfe23f/libgwas/standardizer.py#L65-L92 |
242,265 | gambogi/CSHLDAP | CSHLDAP.py | CSHLDAP.member | def member(self, user, objects=False):
""" Returns a user as a dict of attributes
"""
try:
member = self.search(uid=user, objects=objects)[0]
except IndexError:
return None
if objects:
return member
return member[1] | python | def member(self, user, objects=False):
""" Returns a user as a dict of attributes
"""
try:
member = self.search(uid=user, objects=objects)[0]
except IndexError:
return None
if objects:
return member
return member[1] | [
"def",
"member",
"(",
"self",
",",
"user",
",",
"objects",
"=",
"False",
")",
":",
"try",
":",
"member",
"=",
"self",
".",
"search",
"(",
"uid",
"=",
"user",
",",
"objects",
"=",
"objects",
")",
"[",
"0",
"]",
"except",
"IndexError",
":",
"return",... | Returns a user as a dict of attributes | [
"Returns",
"a",
"user",
"as",
"a",
"dict",
"of",
"attributes"
] | 09cb754b1e72437834e0d8cb4c7ac1830cfa6829 | https://github.com/gambogi/CSHLDAP/blob/09cb754b1e72437834e0d8cb4c7ac1830cfa6829/CSHLDAP.py#L49-L58 |
242,266 | gambogi/CSHLDAP | CSHLDAP.py | CSHLDAP.drinkAdmins | def drinkAdmins(self, objects=False):
""" Returns a list of drink admins uids
"""
admins = self.group('drink', objects=objects)
return admins | python | def drinkAdmins(self, objects=False):
""" Returns a list of drink admins uids
"""
admins = self.group('drink', objects=objects)
return admins | [
"def",
"drinkAdmins",
"(",
"self",
",",
"objects",
"=",
"False",
")",
":",
"admins",
"=",
"self",
".",
"group",
"(",
"'drink'",
",",
"objects",
"=",
"objects",
")",
"return",
"admins"
] | Returns a list of drink admins uids | [
"Returns",
"a",
"list",
"of",
"drink",
"admins",
"uids"
] | 09cb754b1e72437834e0d8cb4c7ac1830cfa6829 | https://github.com/gambogi/CSHLDAP/blob/09cb754b1e72437834e0d8cb4c7ac1830cfa6829/CSHLDAP.py#L99-L103 |
242,267 | gambogi/CSHLDAP | CSHLDAP.py | Member.isBirthday | def isBirthday(self):
""" Is it the user's birthday today?
"""
if not self.birthday:
return False
birthday = self.birthdate()
today = date.today()
return (birthday.month == today.month and
birthday.day == today.day) | python | def isBirthday(self):
""" Is it the user's birthday today?
"""
if not self.birthday:
return False
birthday = self.birthdate()
today = date.today()
return (birthday.month == today.month and
birthday.day == today.day) | [
"def",
"isBirthday",
"(",
"self",
")",
":",
"if",
"not",
"self",
".",
"birthday",
":",
"return",
"False",
"birthday",
"=",
"self",
".",
"birthdate",
"(",
")",
"today",
"=",
"date",
".",
"today",
"(",
")",
"return",
"(",
"birthday",
".",
"month",
"=="... | Is it the user's birthday today? | [
"Is",
"it",
"the",
"user",
"s",
"birthday",
"today?"
] | 09cb754b1e72437834e0d8cb4c7ac1830cfa6829 | https://github.com/gambogi/CSHLDAP/blob/09cb754b1e72437834e0d8cb4c7ac1830cfa6829/CSHLDAP.py#L263-L271 |
242,268 | gambogi/CSHLDAP | CSHLDAP.py | Member.reload | def reload(self):
""" If there is an LDAP connection, query it for another
instance of this member and set its internal dictionary
to that result.
"""
if not self.ldap:
return
self.memberDict = self.ldap.member(self.uid) | python | def reload(self):
""" If there is an LDAP connection, query it for another
instance of this member and set its internal dictionary
to that result.
"""
if not self.ldap:
return
self.memberDict = self.ldap.member(self.uid) | [
"def",
"reload",
"(",
"self",
")",
":",
"if",
"not",
"self",
".",
"ldap",
":",
"return",
"self",
".",
"memberDict",
"=",
"self",
".",
"ldap",
".",
"member",
"(",
"self",
".",
"uid",
")"
] | If there is an LDAP connection, query it for another
instance of this member and set its internal dictionary
to that result. | [
"If",
"there",
"is",
"an",
"LDAP",
"connection",
"query",
"it",
"for",
"another",
"instance",
"of",
"this",
"member",
"and",
"set",
"its",
"internal",
"dictionary",
"to",
"that",
"result",
"."
] | 09cb754b1e72437834e0d8cb4c7ac1830cfa6829 | https://github.com/gambogi/CSHLDAP/blob/09cb754b1e72437834e0d8cb4c7ac1830cfa6829/CSHLDAP.py#L305-L312 |
242,269 | skitazaki/python-clitool | clitool/processor.py | Streamer.consume | def consume(self, stream, source=None, chunksize=1):
""" Consuming given strem object and returns processing stats.
:param stream: streaming object to consume
:type stream: iterable
:param source: source of stream to consume
:type source: string
:param chunksize: chunk size for multiprocessing
:type chunksize: integer
:rtype: dict
"""
stats = {
PROCESSING_TOTAL: 0,
PROCESSING_SKIPPED: 0,
PROCESSING_SUCCESS: 0,
PROCESSING_ERROR: 0
}
if source:
stats['source'] = source
def skip_unless(r):
if r:
return r
stats[PROCESSING_SKIPPED] += 1
stats[PROCESSING_TOTAL] += 1
rs = ifilter(skip_unless, stream)
if self.processes:
pool = multiprocessing.Pool(processes=self.processes)
for f in self.procedures:
rs = pool.imap_unordered(f, ifilter(skip_unless, rs),
chunksize=chunksize)
else:
for f in self.procedures:
rs = imap(f, ifilter(skip_unless, rs))
start = time.time()
i = 0
try:
while 1:
processed = next(rs)
if processed is None:
stats[PROCESSING_SKIPPED] += 1
elif processed is False:
stats[PROCESSING_ERROR] += 1
else:
stats[PROCESSING_SUCCESS] += 1
self.collect(processed)
i += 1
stats[PROCESSING_TOTAL] += 1
if i % self.reporting_interval == 0:
logging.info(" ===> Processed %dth item <=== ", i)
except StopIteration:
pass
except KeyboardInterrupt:
logging.info("Stopped by user interruption at %dth item.", i)
raise
except:
e = sys.exc_info()[1]
logging.error(e)
raise
finally:
if self.processes:
pool.close()
pool.join()
stats[PROCESSING_TIME] = time.time() - start
logging.info(
'STATS: total=%d, skipped=%d, success=%d, error=%d on %f[sec]'
' from "%s"',
stats[PROCESSING_TOTAL], stats[PROCESSING_SKIPPED],
stats[PROCESSING_SUCCESS], stats[PROCESSING_ERROR],
stats[PROCESSING_TIME], stats.get('source', 'unknown'))
return stats | python | def consume(self, stream, source=None, chunksize=1):
""" Consuming given strem object and returns processing stats.
:param stream: streaming object to consume
:type stream: iterable
:param source: source of stream to consume
:type source: string
:param chunksize: chunk size for multiprocessing
:type chunksize: integer
:rtype: dict
"""
stats = {
PROCESSING_TOTAL: 0,
PROCESSING_SKIPPED: 0,
PROCESSING_SUCCESS: 0,
PROCESSING_ERROR: 0
}
if source:
stats['source'] = source
def skip_unless(r):
if r:
return r
stats[PROCESSING_SKIPPED] += 1
stats[PROCESSING_TOTAL] += 1
rs = ifilter(skip_unless, stream)
if self.processes:
pool = multiprocessing.Pool(processes=self.processes)
for f in self.procedures:
rs = pool.imap_unordered(f, ifilter(skip_unless, rs),
chunksize=chunksize)
else:
for f in self.procedures:
rs = imap(f, ifilter(skip_unless, rs))
start = time.time()
i = 0
try:
while 1:
processed = next(rs)
if processed is None:
stats[PROCESSING_SKIPPED] += 1
elif processed is False:
stats[PROCESSING_ERROR] += 1
else:
stats[PROCESSING_SUCCESS] += 1
self.collect(processed)
i += 1
stats[PROCESSING_TOTAL] += 1
if i % self.reporting_interval == 0:
logging.info(" ===> Processed %dth item <=== ", i)
except StopIteration:
pass
except KeyboardInterrupt:
logging.info("Stopped by user interruption at %dth item.", i)
raise
except:
e = sys.exc_info()[1]
logging.error(e)
raise
finally:
if self.processes:
pool.close()
pool.join()
stats[PROCESSING_TIME] = time.time() - start
logging.info(
'STATS: total=%d, skipped=%d, success=%d, error=%d on %f[sec]'
' from "%s"',
stats[PROCESSING_TOTAL], stats[PROCESSING_SKIPPED],
stats[PROCESSING_SUCCESS], stats[PROCESSING_ERROR],
stats[PROCESSING_TIME], stats.get('source', 'unknown'))
return stats | [
"def",
"consume",
"(",
"self",
",",
"stream",
",",
"source",
"=",
"None",
",",
"chunksize",
"=",
"1",
")",
":",
"stats",
"=",
"{",
"PROCESSING_TOTAL",
":",
"0",
",",
"PROCESSING_SKIPPED",
":",
"0",
",",
"PROCESSING_SUCCESS",
":",
"0",
",",
"PROCESSING_ER... | Consuming given strem object and returns processing stats.
:param stream: streaming object to consume
:type stream: iterable
:param source: source of stream to consume
:type source: string
:param chunksize: chunk size for multiprocessing
:type chunksize: integer
:rtype: dict | [
"Consuming",
"given",
"strem",
"object",
"and",
"returns",
"processing",
"stats",
"."
] | 4971f8d093d51c6fd0e6cc536bbb597f78b570ab | https://github.com/skitazaki/python-clitool/blob/4971f8d093d51c6fd0e6cc536bbb597f78b570ab/clitool/processor.py#L169-L240 |
242,270 | skitazaki/python-clitool | clitool/processor.py | CliHandler.reader | def reader(self, fp, encoding):
""" Simple `open` wrapper for several file types.
This supports ``.gz`` and ``.json``.
:param fp: opened file
:type fp: file pointer
:param encoding: encoding of opened file
:type encoding: string
:rtype: file pointer
"""
_, suffix = os.path.splitext(fp.name)
if suffix == '.gz':
fp.close()
return gzip.open(fp.name)
elif suffix == '.json':
return json.load(fp)
elif suffix == '.csv' or self.delimiter:
return csvreader(fp, encoding, delimiter=self.delimiter or ',')
elif suffix == '.tsv':
return csvreader(fp, encoding, delimiter='\t')
return fp | python | def reader(self, fp, encoding):
""" Simple `open` wrapper for several file types.
This supports ``.gz`` and ``.json``.
:param fp: opened file
:type fp: file pointer
:param encoding: encoding of opened file
:type encoding: string
:rtype: file pointer
"""
_, suffix = os.path.splitext(fp.name)
if suffix == '.gz':
fp.close()
return gzip.open(fp.name)
elif suffix == '.json':
return json.load(fp)
elif suffix == '.csv' or self.delimiter:
return csvreader(fp, encoding, delimiter=self.delimiter or ',')
elif suffix == '.tsv':
return csvreader(fp, encoding, delimiter='\t')
return fp | [
"def",
"reader",
"(",
"self",
",",
"fp",
",",
"encoding",
")",
":",
"_",
",",
"suffix",
"=",
"os",
".",
"path",
".",
"splitext",
"(",
"fp",
".",
"name",
")",
"if",
"suffix",
"==",
"'.gz'",
":",
"fp",
".",
"close",
"(",
")",
"return",
"gzip",
".... | Simple `open` wrapper for several file types.
This supports ``.gz`` and ``.json``.
:param fp: opened file
:type fp: file pointer
:param encoding: encoding of opened file
:type encoding: string
:rtype: file pointer | [
"Simple",
"open",
"wrapper",
"for",
"several",
"file",
"types",
".",
"This",
"supports",
".",
"gz",
"and",
".",
"json",
"."
] | 4971f8d093d51c6fd0e6cc536bbb597f78b570ab | https://github.com/skitazaki/python-clitool/blob/4971f8d093d51c6fd0e6cc536bbb597f78b570ab/clitool/processor.py#L257-L277 |
242,271 | skitazaki/python-clitool | clitool/processor.py | CliHandler.handle | def handle(self, files, encoding, chunksize=1):
""" Handle given files with given encoding.
:param files: opened files.
:type files: list
:param encoding: encoding of opened file
:type encoding: string
:param chunksize: a number of chunk
:type chunksize: int
:rtype: list
"""
stats = []
if files:
logging.info("Input file count: %d", len(files))
for fp in files:
stream = self.reader(fp, encoding)
parsed = self.streamer.consume(stream,
source=fp.name, chunksize=chunksize)
stats.append(parsed)
if not fp.closed:
fp.close()
else:
stream = sys.stdin
if self.delimiter:
stream = csvreader(stream, encoding, delimiter=self.delimiter)
parsed = self.streamer.consume(stream, chunksize=chunksize)
stats.append(parsed)
return stats | python | def handle(self, files, encoding, chunksize=1):
""" Handle given files with given encoding.
:param files: opened files.
:type files: list
:param encoding: encoding of opened file
:type encoding: string
:param chunksize: a number of chunk
:type chunksize: int
:rtype: list
"""
stats = []
if files:
logging.info("Input file count: %d", len(files))
for fp in files:
stream = self.reader(fp, encoding)
parsed = self.streamer.consume(stream,
source=fp.name, chunksize=chunksize)
stats.append(parsed)
if not fp.closed:
fp.close()
else:
stream = sys.stdin
if self.delimiter:
stream = csvreader(stream, encoding, delimiter=self.delimiter)
parsed = self.streamer.consume(stream, chunksize=chunksize)
stats.append(parsed)
return stats | [
"def",
"handle",
"(",
"self",
",",
"files",
",",
"encoding",
",",
"chunksize",
"=",
"1",
")",
":",
"stats",
"=",
"[",
"]",
"if",
"files",
":",
"logging",
".",
"info",
"(",
"\"Input file count: %d\"",
",",
"len",
"(",
"files",
")",
")",
"for",
"fp",
... | Handle given files with given encoding.
:param files: opened files.
:type files: list
:param encoding: encoding of opened file
:type encoding: string
:param chunksize: a number of chunk
:type chunksize: int
:rtype: list | [
"Handle",
"given",
"files",
"with",
"given",
"encoding",
"."
] | 4971f8d093d51c6fd0e6cc536bbb597f78b570ab | https://github.com/skitazaki/python-clitool/blob/4971f8d093d51c6fd0e6cc536bbb597f78b570ab/clitool/processor.py#L279-L306 |
242,272 | vivainio/argp | argp.py | init | def init(parser = None):
""" module needs to be initialized by 'init'.
Can be called with parser to use a pre-built parser, otherwise
a simple default parser is created
"""
global p,subparsers
if parser is None:
p = argparse.ArgumentParser()
else:
p = parser
arg = p.add_argument
subparsers = p.add_subparsers() | python | def init(parser = None):
""" module needs to be initialized by 'init'.
Can be called with parser to use a pre-built parser, otherwise
a simple default parser is created
"""
global p,subparsers
if parser is None:
p = argparse.ArgumentParser()
else:
p = parser
arg = p.add_argument
subparsers = p.add_subparsers() | [
"def",
"init",
"(",
"parser",
"=",
"None",
")",
":",
"global",
"p",
",",
"subparsers",
"if",
"parser",
"is",
"None",
":",
"p",
"=",
"argparse",
".",
"ArgumentParser",
"(",
")",
"else",
":",
"p",
"=",
"parser",
"arg",
"=",
"p",
".",
"add_argument",
... | module needs to be initialized by 'init'.
Can be called with parser to use a pre-built parser, otherwise
a simple default parser is created | [
"module",
"needs",
"to",
"be",
"initialized",
"by",
"init",
"."
] | ad90edbecd5f84d7dce09a2bd98e1e9d98140f1e | https://github.com/vivainio/argp/blob/ad90edbecd5f84d7dce09a2bd98e1e9d98140f1e/argp.py#L21-L36 |
242,273 | ncc-tools/python-pa-api | paapi/paapi.py | PaAuth.authenticate | def authenticate(self):
"""
Authenticates with the PA Oauth system
"""
if self._auth_token is None or self._token_expiry < time.time():
self._perform_auth()
yield self._auth_token | python | def authenticate(self):
"""
Authenticates with the PA Oauth system
"""
if self._auth_token is None or self._token_expiry < time.time():
self._perform_auth()
yield self._auth_token | [
"def",
"authenticate",
"(",
"self",
")",
":",
"if",
"self",
".",
"_auth_token",
"is",
"None",
"or",
"self",
".",
"_token_expiry",
"<",
"time",
".",
"time",
"(",
")",
":",
"self",
".",
"_perform_auth",
"(",
")",
"yield",
"self",
".",
"_auth_token"
] | Authenticates with the PA Oauth system | [
"Authenticates",
"with",
"the",
"PA",
"Oauth",
"system"
] | a27481dd323d282d0f4457586198d9faec896f11 | https://github.com/ncc-tools/python-pa-api/blob/a27481dd323d282d0f4457586198d9faec896f11/paapi/paapi.py#L65-L72 |
242,274 | ncc-tools/python-pa-api | paapi/paapi.py | PaApi._query_api | def _query_api(self, method, url, fields=None, extra_headers=None, req_body=None):
"""
Abstracts http queries to the API
"""
with self.auth.authenticate() as token:
logging.debug('PA Authentication returned token %s', token)
headers = {
'Authorization': 'Bearer %s' % (token,),
'Realm': self.auth_realm
}
if extra_headers is not None:
headers.update(extra_headers)
logging.info('[%s] %s', method, url)
if req_body is not None:
response = self.http.request(method, url, fields, headers, body=req_body)
else:
response = self.http.request(method, url, fields, headers)
if response.status != 200:
print(response.data)
logging.warning('Got non-200 HTTP status from API: %d', response.status)
raise ApiQueryError("Failed to get API data", response.status)
return json.loads(response.data.decode()) | python | def _query_api(self, method, url, fields=None, extra_headers=None, req_body=None):
"""
Abstracts http queries to the API
"""
with self.auth.authenticate() as token:
logging.debug('PA Authentication returned token %s', token)
headers = {
'Authorization': 'Bearer %s' % (token,),
'Realm': self.auth_realm
}
if extra_headers is not None:
headers.update(extra_headers)
logging.info('[%s] %s', method, url)
if req_body is not None:
response = self.http.request(method, url, fields, headers, body=req_body)
else:
response = self.http.request(method, url, fields, headers)
if response.status != 200:
print(response.data)
logging.warning('Got non-200 HTTP status from API: %d', response.status)
raise ApiQueryError("Failed to get API data", response.status)
return json.loads(response.data.decode()) | [
"def",
"_query_api",
"(",
"self",
",",
"method",
",",
"url",
",",
"fields",
"=",
"None",
",",
"extra_headers",
"=",
"None",
",",
"req_body",
"=",
"None",
")",
":",
"with",
"self",
".",
"auth",
".",
"authenticate",
"(",
")",
"as",
"token",
":",
"loggi... | Abstracts http queries to the API | [
"Abstracts",
"http",
"queries",
"to",
"the",
"API"
] | a27481dd323d282d0f4457586198d9faec896f11 | https://github.com/ncc-tools/python-pa-api/blob/a27481dd323d282d0f4457586198d9faec896f11/paapi/paapi.py#L121-L143 |
242,275 | ncc-tools/python-pa-api | paapi/paapi.py | PaApi.get_all_jobtemplates | def get_all_jobtemplates(self):
"""
Retrieves the list of jobTemplates for the current realm.
"""
endpoint = self._build_url('jobTemplates', {
'paginationPageSize': self.PAGE_SIZE
})
data = self._query_api('GET', endpoint)
return data['results'] | python | def get_all_jobtemplates(self):
"""
Retrieves the list of jobTemplates for the current realm.
"""
endpoint = self._build_url('jobTemplates', {
'paginationPageSize': self.PAGE_SIZE
})
data = self._query_api('GET', endpoint)
return data['results'] | [
"def",
"get_all_jobtemplates",
"(",
"self",
")",
":",
"endpoint",
"=",
"self",
".",
"_build_url",
"(",
"'jobTemplates'",
",",
"{",
"'paginationPageSize'",
":",
"self",
".",
"PAGE_SIZE",
"}",
")",
"data",
"=",
"self",
".",
"_query_api",
"(",
"'GET'",
",",
"... | Retrieves the list of jobTemplates for the current realm. | [
"Retrieves",
"the",
"list",
"of",
"jobTemplates",
"for",
"the",
"current",
"realm",
"."
] | a27481dd323d282d0f4457586198d9faec896f11 | https://github.com/ncc-tools/python-pa-api/blob/a27481dd323d282d0f4457586198d9faec896f11/paapi/paapi.py#L151-L159 |
242,276 | ncc-tools/python-pa-api | paapi/paapi.py | PaApi.create_job_template | def create_job_template(self, template):
"""
Creates a job template
"""
endpoint = self._build_url('jobTemplates')
data = self._query_api('POST',
endpoint,
None,
{'Content-Type': 'application/json'},
json.dumps(template))
return data['results'] | python | def create_job_template(self, template):
"""
Creates a job template
"""
endpoint = self._build_url('jobTemplates')
data = self._query_api('POST',
endpoint,
None,
{'Content-Type': 'application/json'},
json.dumps(template))
return data['results'] | [
"def",
"create_job_template",
"(",
"self",
",",
"template",
")",
":",
"endpoint",
"=",
"self",
".",
"_build_url",
"(",
"'jobTemplates'",
")",
"data",
"=",
"self",
".",
"_query_api",
"(",
"'POST'",
",",
"endpoint",
",",
"None",
",",
"{",
"'Content-Type'",
"... | Creates a job template | [
"Creates",
"a",
"job",
"template"
] | a27481dd323d282d0f4457586198d9faec896f11 | https://github.com/ncc-tools/python-pa-api/blob/a27481dd323d282d0f4457586198d9faec896f11/paapi/paapi.py#L182-L192 |
242,277 | ncc-tools/python-pa-api | paapi/paapi.py | PaApi.create_job | def create_job(self, job_template_uri):
"""
Creates a job
"""
endpoint = self._build_url('jobs')
data = self._query_api('POST',
endpoint,
None,
{'Content-Type': 'application/json'},
json.dumps({'jobTemplateUri': job_template_uri}))
return data['results'] | python | def create_job(self, job_template_uri):
"""
Creates a job
"""
endpoint = self._build_url('jobs')
data = self._query_api('POST',
endpoint,
None,
{'Content-Type': 'application/json'},
json.dumps({'jobTemplateUri': job_template_uri}))
return data['results'] | [
"def",
"create_job",
"(",
"self",
",",
"job_template_uri",
")",
":",
"endpoint",
"=",
"self",
".",
"_build_url",
"(",
"'jobs'",
")",
"data",
"=",
"self",
".",
"_query_api",
"(",
"'POST'",
",",
"endpoint",
",",
"None",
",",
"{",
"'Content-Type'",
":",
"'a... | Creates a job | [
"Creates",
"a",
"job"
] | a27481dd323d282d0f4457586198d9faec896f11 | https://github.com/ncc-tools/python-pa-api/blob/a27481dd323d282d0f4457586198d9faec896f11/paapi/paapi.py#L193-L203 |
242,278 | wdbm/datavision | datavision.py | normalize_to_range | def normalize_to_range(
values,
minimum = 0.0,
maximum = 1.0
):
"""
This function normalizes values of a list to a specified range and returns
the original object if the values are not of the types integer or float.
"""
normalized_values = []
minimum_value = min(values)
maximum_value = max(values)
for value in values:
numerator = value - minimum_value
denominator = maximum_value - minimum_value
value_normalized = (maximum - minimum) * numerator/denominator + minimum
normalized_values.append(value_normalized)
return normalized_values | python | def normalize_to_range(
values,
minimum = 0.0,
maximum = 1.0
):
"""
This function normalizes values of a list to a specified range and returns
the original object if the values are not of the types integer or float.
"""
normalized_values = []
minimum_value = min(values)
maximum_value = max(values)
for value in values:
numerator = value - minimum_value
denominator = maximum_value - minimum_value
value_normalized = (maximum - minimum) * numerator/denominator + minimum
normalized_values.append(value_normalized)
return normalized_values | [
"def",
"normalize_to_range",
"(",
"values",
",",
"minimum",
"=",
"0.0",
",",
"maximum",
"=",
"1.0",
")",
":",
"normalized_values",
"=",
"[",
"]",
"minimum_value",
"=",
"min",
"(",
"values",
")",
"maximum_value",
"=",
"max",
"(",
"values",
")",
"for",
"va... | This function normalizes values of a list to a specified range and returns
the original object if the values are not of the types integer or float. | [
"This",
"function",
"normalizes",
"values",
"of",
"a",
"list",
"to",
"a",
"specified",
"range",
"and",
"returns",
"the",
"original",
"object",
"if",
"the",
"values",
"are",
"not",
"of",
"the",
"types",
"integer",
"or",
"float",
"."
] | b6f26287264632d6f8c9f8911aaf3a8e4fc4dcf5 | https://github.com/wdbm/datavision/blob/b6f26287264632d6f8c9f8911aaf3a8e4fc4dcf5/datavision.py#L1206-L1225 |
242,279 | wdbm/datavision | datavision.py | list_element_combinations_variadic | def list_element_combinations_variadic(
elements_specification
):
"""
This function accepts a specification of lists of elements for each place in
lists in the form of a list, the elements of which are lists of possible
elements and returns a list of lists corresponding to the combinations of
elements of the specification with varying numbers of elements.
For example, the list elements specification [[10, 20], [30, 40], [50, 60]]
yields the following lists:
[10]
[20]
[10, 30]
[10, 40]
[20, 30]
[20, 40]
[10, 30, 50]
[10, 30, 60]
[10, 40, 50]
[10, 40, 60]
[20, 30, 50]
[20, 30, 60]
[20, 40, 50]
[20, 40, 60]
"""
lists = [list(list_generated) for index, element_specification in enumerate(elements_specification) for list_generated in itertools.product(*elements_specification[:index + 1])]
return lists | python | def list_element_combinations_variadic(
elements_specification
):
"""
This function accepts a specification of lists of elements for each place in
lists in the form of a list, the elements of which are lists of possible
elements and returns a list of lists corresponding to the combinations of
elements of the specification with varying numbers of elements.
For example, the list elements specification [[10, 20], [30, 40], [50, 60]]
yields the following lists:
[10]
[20]
[10, 30]
[10, 40]
[20, 30]
[20, 40]
[10, 30, 50]
[10, 30, 60]
[10, 40, 50]
[10, 40, 60]
[20, 30, 50]
[20, 30, 60]
[20, 40, 50]
[20, 40, 60]
"""
lists = [list(list_generated) for index, element_specification in enumerate(elements_specification) for list_generated in itertools.product(*elements_specification[:index + 1])]
return lists | [
"def",
"list_element_combinations_variadic",
"(",
"elements_specification",
")",
":",
"lists",
"=",
"[",
"list",
"(",
"list_generated",
")",
"for",
"index",
",",
"element_specification",
"in",
"enumerate",
"(",
"elements_specification",
")",
"for",
"list_generated",
"... | This function accepts a specification of lists of elements for each place in
lists in the form of a list, the elements of which are lists of possible
elements and returns a list of lists corresponding to the combinations of
elements of the specification with varying numbers of elements.
For example, the list elements specification [[10, 20], [30, 40], [50, 60]]
yields the following lists:
[10]
[20]
[10, 30]
[10, 40]
[20, 30]
[20, 40]
[10, 30, 50]
[10, 30, 60]
[10, 40, 50]
[10, 40, 60]
[20, 30, 50]
[20, 30, 60]
[20, 40, 50]
[20, 40, 60] | [
"This",
"function",
"accepts",
"a",
"specification",
"of",
"lists",
"of",
"elements",
"for",
"each",
"place",
"in",
"lists",
"in",
"the",
"form",
"of",
"a",
"list",
"the",
"elements",
"of",
"which",
"are",
"lists",
"of",
"possible",
"elements",
"and",
"ret... | b6f26287264632d6f8c9f8911aaf3a8e4fc4dcf5 | https://github.com/wdbm/datavision/blob/b6f26287264632d6f8c9f8911aaf3a8e4fc4dcf5/datavision.py#L1263-L1293 |
242,280 | wdbm/datavision | datavision.py | correlation_linear | def correlation_linear(
values_1,
values_2,
printout = None
):
"""
This function calculates the Pearson product-moment correlation coefficient.
This is a measure of the linear collelation of two variables. The value can
be between +1 and -1 inclusive, where 1 is total positive correlation, 0 is
no correlation and -1 is total negative correlation. It is a measure of the
linear dependence between two variables.
This function also calculates the significance (2-tailed p-value) of the
correlation coefficient given the sample size.
"""
r, p_value = scipy.stats.pearsonr(values_1, values_2)
if printout is not True:
return r, p_value
else:
text = (
"Pearson linear correlation coefficient: {r}\n"
"2-tailed p-value: {p_value}"
).format(
r = r,
p_value = p_value
)
return text | python | def correlation_linear(
values_1,
values_2,
printout = None
):
"""
This function calculates the Pearson product-moment correlation coefficient.
This is a measure of the linear collelation of two variables. The value can
be between +1 and -1 inclusive, where 1 is total positive correlation, 0 is
no correlation and -1 is total negative correlation. It is a measure of the
linear dependence between two variables.
This function also calculates the significance (2-tailed p-value) of the
correlation coefficient given the sample size.
"""
r, p_value = scipy.stats.pearsonr(values_1, values_2)
if printout is not True:
return r, p_value
else:
text = (
"Pearson linear correlation coefficient: {r}\n"
"2-tailed p-value: {p_value}"
).format(
r = r,
p_value = p_value
)
return text | [
"def",
"correlation_linear",
"(",
"values_1",
",",
"values_2",
",",
"printout",
"=",
"None",
")",
":",
"r",
",",
"p_value",
"=",
"scipy",
".",
"stats",
".",
"pearsonr",
"(",
"values_1",
",",
"values_2",
")",
"if",
"printout",
"is",
"not",
"True",
":",
... | This function calculates the Pearson product-moment correlation coefficient.
This is a measure of the linear collelation of two variables. The value can
be between +1 and -1 inclusive, where 1 is total positive correlation, 0 is
no correlation and -1 is total negative correlation. It is a measure of the
linear dependence between two variables.
This function also calculates the significance (2-tailed p-value) of the
correlation coefficient given the sample size. | [
"This",
"function",
"calculates",
"the",
"Pearson",
"product",
"-",
"moment",
"correlation",
"coefficient",
".",
"This",
"is",
"a",
"measure",
"of",
"the",
"linear",
"collelation",
"of",
"two",
"variables",
".",
"The",
"value",
"can",
"be",
"between",
"+",
"... | b6f26287264632d6f8c9f8911aaf3a8e4fc4dcf5 | https://github.com/wdbm/datavision/blob/b6f26287264632d6f8c9f8911aaf3a8e4fc4dcf5/datavision.py#L1295-L1323 |
242,281 | wdbm/datavision | datavision.py | propose_number_of_bins | def propose_number_of_bins(
values,
binning_logic_system = None,
):
"""
This function returns a proposal for binning for a histogram of a specified
list using an optional specified binning logic system.
Freedman-Diaconis: bin width is proportional to the interquartile range of
the data divided by the cube root of the size of the data
Scott: bin width is proportional to the standard deviation of
the values divided by the cube root of the size of the
data
"""
# Set the default binning logic system.
if binning_logic_system is None:
binning_logic_system = "Scott"
# Engage the requested logic system.
if binning_logic_system == "Freedman-Diaconis":
#log.debug("engage Freedman-Diaconis binning logic")
bin_size =\
2 * interquartile_range(values) * \
len(values) ** (-1/3)
elif binning_logic_system == "Scott":
#log.debug("engage Scott binning logic")
bin_size =\
3.5 * standard_deviation(values) * \
len(values) ** (-1/3)
else:
log.error("undefined binning logic system requested")
raise(ValueError)
number_of_bins = (max(values) - min(values)) / bin_size
if numpy.isinf(number_of_bins) or numpy.isnan(number_of_bins):
number_of_bins = len(set(values)) # number of unique values
#log.debug(
# "binning algorithms ineffective -- " +
# "propose binning by unique values"
#)
return int(round(number_of_bins)) | python | def propose_number_of_bins(
values,
binning_logic_system = None,
):
"""
This function returns a proposal for binning for a histogram of a specified
list using an optional specified binning logic system.
Freedman-Diaconis: bin width is proportional to the interquartile range of
the data divided by the cube root of the size of the data
Scott: bin width is proportional to the standard deviation of
the values divided by the cube root of the size of the
data
"""
# Set the default binning logic system.
if binning_logic_system is None:
binning_logic_system = "Scott"
# Engage the requested logic system.
if binning_logic_system == "Freedman-Diaconis":
#log.debug("engage Freedman-Diaconis binning logic")
bin_size =\
2 * interquartile_range(values) * \
len(values) ** (-1/3)
elif binning_logic_system == "Scott":
#log.debug("engage Scott binning logic")
bin_size =\
3.5 * standard_deviation(values) * \
len(values) ** (-1/3)
else:
log.error("undefined binning logic system requested")
raise(ValueError)
number_of_bins = (max(values) - min(values)) / bin_size
if numpy.isinf(number_of_bins) or numpy.isnan(number_of_bins):
number_of_bins = len(set(values)) # number of unique values
#log.debug(
# "binning algorithms ineffective -- " +
# "propose binning by unique values"
#)
return int(round(number_of_bins)) | [
"def",
"propose_number_of_bins",
"(",
"values",
",",
"binning_logic_system",
"=",
"None",
",",
")",
":",
"# Set the default binning logic system.",
"if",
"binning_logic_system",
"is",
"None",
":",
"binning_logic_system",
"=",
"\"Scott\"",
"# Engage the requested logic system.... | This function returns a proposal for binning for a histogram of a specified
list using an optional specified binning logic system.
Freedman-Diaconis: bin width is proportional to the interquartile range of
the data divided by the cube root of the size of the data
Scott: bin width is proportional to the standard deviation of
the values divided by the cube root of the size of the
data | [
"This",
"function",
"returns",
"a",
"proposal",
"for",
"binning",
"for",
"a",
"histogram",
"of",
"a",
"specified",
"list",
"using",
"an",
"optional",
"specified",
"binning",
"logic",
"system",
"."
] | b6f26287264632d6f8c9f8911aaf3a8e4fc4dcf5 | https://github.com/wdbm/datavision/blob/b6f26287264632d6f8c9f8911aaf3a8e4fc4dcf5/datavision.py#L1401-L1442 |
242,282 | wdbm/datavision | datavision.py | TTYFigureData.extent | def extent(self):
"""
return range of 2D data
"""
return [min(self.x), max(self.x), min(self.y), max(self.y)] | python | def extent(self):
"""
return range of 2D data
"""
return [min(self.x), max(self.x), min(self.y), max(self.y)] | [
"def",
"extent",
"(",
"self",
")",
":",
"return",
"[",
"min",
"(",
"self",
".",
"x",
")",
",",
"max",
"(",
"self",
".",
"x",
")",
",",
"min",
"(",
"self",
".",
"y",
")",
",",
"max",
"(",
"self",
".",
"y",
")",
"]"
] | return range of 2D data | [
"return",
"range",
"of",
"2D",
"data"
] | b6f26287264632d6f8c9f8911aaf3a8e4fc4dcf5 | https://github.com/wdbm/datavision/blob/b6f26287264632d6f8c9f8911aaf3a8e4fc4dcf5/datavision.py#L2030-L2036 |
242,283 | wdbm/datavision | datavision.py | TTYFigure._get_symbol_by_slope | def _get_symbol_by_slope(
self,
slope,
default_symbol
):
"""
return line oriented approximatively along the slope value
"""
if slope > math.tan(3 * math.pi / 8):
draw_symbol = "|"
elif math.tan(math.pi / 8) < slope < math.tan(3 * math.pi / 8):
draw_symbol = u"\u27cb" # "/"
elif abs(slope) < math.tan(math.pi / 8):
draw_symbol = "-"
elif slope < math.tan(-math.pi / 8) and\
slope > math.tan(-3 * math.pi / 8):
draw_symbol = u"\u27CD" # "\\"
elif slope < math.tan(-3 * math.pi / 8):
draw_symbol = "|"
else:
draw_symbol = default_symbol
return draw_symbol | python | def _get_symbol_by_slope(
self,
slope,
default_symbol
):
"""
return line oriented approximatively along the slope value
"""
if slope > math.tan(3 * math.pi / 8):
draw_symbol = "|"
elif math.tan(math.pi / 8) < slope < math.tan(3 * math.pi / 8):
draw_symbol = u"\u27cb" # "/"
elif abs(slope) < math.tan(math.pi / 8):
draw_symbol = "-"
elif slope < math.tan(-math.pi / 8) and\
slope > math.tan(-3 * math.pi / 8):
draw_symbol = u"\u27CD" # "\\"
elif slope < math.tan(-3 * math.pi / 8):
draw_symbol = "|"
else:
draw_symbol = default_symbol
return draw_symbol | [
"def",
"_get_symbol_by_slope",
"(",
"self",
",",
"slope",
",",
"default_symbol",
")",
":",
"if",
"slope",
">",
"math",
".",
"tan",
"(",
"3",
"*",
"math",
".",
"pi",
"/",
"8",
")",
":",
"draw_symbol",
"=",
"\"|\"",
"elif",
"math",
".",
"tan",
"(",
"... | return line oriented approximatively along the slope value | [
"return",
"line",
"oriented",
"approximatively",
"along",
"the",
"slope",
"value"
] | b6f26287264632d6f8c9f8911aaf3a8e4fc4dcf5 | https://github.com/wdbm/datavision/blob/b6f26287264632d6f8c9f8911aaf3a8e4fc4dcf5/datavision.py#L2124-L2147 |
242,284 | wdbm/datavision | datavision.py | TTYCanvas.limit_x | def limit_x(
self,
limit_lower = None, # float
limit_upper = None # float
):
"""
get or set x limits of the current axes
x_min, x_max = limit_x() # return the current limit_x
limit_x(x_min, x_max) # set the limit_x to x_min, x_max
"""
if limit_lower is None and limit_upper is None:
return self._limit_x
elif hasattr(limit_lower, "__iter__"):
self._limit_x = limit_lower[:2]
else:
self._limit_x = [limit_lower, limit_upper]
if self._limit_x[0] == self._limit_x[1]:
self._limit_x[1] += 1
self._limit_x[0] -= self.mod_x
self._limit_x[1] += self.mod_x | python | def limit_x(
self,
limit_lower = None, # float
limit_upper = None # float
):
"""
get or set x limits of the current axes
x_min, x_max = limit_x() # return the current limit_x
limit_x(x_min, x_max) # set the limit_x to x_min, x_max
"""
if limit_lower is None and limit_upper is None:
return self._limit_x
elif hasattr(limit_lower, "__iter__"):
self._limit_x = limit_lower[:2]
else:
self._limit_x = [limit_lower, limit_upper]
if self._limit_x[0] == self._limit_x[1]:
self._limit_x[1] += 1
self._limit_x[0] -= self.mod_x
self._limit_x[1] += self.mod_x | [
"def",
"limit_x",
"(",
"self",
",",
"limit_lower",
"=",
"None",
",",
"# float",
"limit_upper",
"=",
"None",
"# float",
")",
":",
"if",
"limit_lower",
"is",
"None",
"and",
"limit_upper",
"is",
"None",
":",
"return",
"self",
".",
"_limit_x",
"elif",
"hasattr... | get or set x limits of the current axes
x_min, x_max = limit_x() # return the current limit_x
limit_x(x_min, x_max) # set the limit_x to x_min, x_max | [
"get",
"or",
"set",
"x",
"limits",
"of",
"the",
"current",
"axes"
] | b6f26287264632d6f8c9f8911aaf3a8e4fc4dcf5 | https://github.com/wdbm/datavision/blob/b6f26287264632d6f8c9f8911aaf3a8e4fc4dcf5/datavision.py#L2538-L2560 |
242,285 | wdbm/datavision | datavision.py | TTYCanvas.limit_y | def limit_y(
self,
limit_lower = None,
limit_upper = None
):
"""
get or set y limits of the current axes
y_min, y_max = limit_x() # return the current limit_y
limit_y(y_min, y_max) # set the limit_y to y_min, y_max
"""
if limit_lower is None and limit_upper is None:
return self._limit_y
elif hasattr(limit_lower, "__iter__"):
self._limit_y = limit_lower[:2]
else:
self._limit_y = [limit_lower, limit_upper]
if self._limit_y[0] == self._limit_y[1]:
self._limit_y[1] += 1
self._limit_y[0] -= self.mod_y
self._limit_y[1] += self.mod_y | python | def limit_y(
self,
limit_lower = None,
limit_upper = None
):
"""
get or set y limits of the current axes
y_min, y_max = limit_x() # return the current limit_y
limit_y(y_min, y_max) # set the limit_y to y_min, y_max
"""
if limit_lower is None and limit_upper is None:
return self._limit_y
elif hasattr(limit_lower, "__iter__"):
self._limit_y = limit_lower[:2]
else:
self._limit_y = [limit_lower, limit_upper]
if self._limit_y[0] == self._limit_y[1]:
self._limit_y[1] += 1
self._limit_y[0] -= self.mod_y
self._limit_y[1] += self.mod_y | [
"def",
"limit_y",
"(",
"self",
",",
"limit_lower",
"=",
"None",
",",
"limit_upper",
"=",
"None",
")",
":",
"if",
"limit_lower",
"is",
"None",
"and",
"limit_upper",
"is",
"None",
":",
"return",
"self",
".",
"_limit_y",
"elif",
"hasattr",
"(",
"limit_lower",... | get or set y limits of the current axes
y_min, y_max = limit_x() # return the current limit_y
limit_y(y_min, y_max) # set the limit_y to y_min, y_max | [
"get",
"or",
"set",
"y",
"limits",
"of",
"the",
"current",
"axes"
] | b6f26287264632d6f8c9f8911aaf3a8e4fc4dcf5 | https://github.com/wdbm/datavision/blob/b6f26287264632d6f8c9f8911aaf3a8e4fc4dcf5/datavision.py#L2562-L2584 |
242,286 | wdbm/datavision | datavision.py | TTYCanvas._clip_line | def _clip_line(
self,
line_pt_1,
line_pt_2
):
"""
clip line to canvas
"""
x_min = min(line_pt_1[0], line_pt_2[0])
x_max = max(line_pt_1[0], line_pt_2[0])
y_min = min(line_pt_1[1], line_pt_2[1])
y_max = max(line_pt_1[1], line_pt_2[1])
extent = self.extent()
if line_pt_1[0] == line_pt_2[0]:
return (
(line_pt_1[0], max(y_min, extent[1])),
(line_pt_1[0], min(y_max, extent[3]))
)
if line_pt_1[1] == line_pt_2[1]:
return (
(max(x_min, extent[0]), line_pt_1[1]),
(min(x_max, extent[2]), line_pt_1[1])
)
if ((extent[0] <= line_pt_1[0] < extent[2]) and
(extent[1] <= line_pt_1[1] < extent[3]) and
(extent[0] <= line_pt_2[0] < extent[2]) and
(extent[1] <= line_pt_2[1] < extent[3])):
return line_pt_1, line_pt_2
ts = [0.0,
1.0,
float(extent[0] - line_pt_1[0]) / (line_pt_2[0] - line_pt_1[0]),
float(extent[2] - line_pt_1[0]) / (line_pt_2[0] - line_pt_1[0]),
float(extent[1] - line_pt_1[1]) / (line_pt_2[1] - line_pt_1[1]),
float(extent[3] - line_pt_1[1]) / (line_pt_2[1] - line_pt_1[1])
]
ts.sort()
if (ts[2] < 0) or (ts[2] >= 1) or (ts[3] < 0) or (ts[2] >= 1):
return None
result =\
[(pt_1 + t * (pt_2 - pt_1))\
for t in (ts[2], ts[3])\
for (pt_1, pt_2) in zip(line_pt_1, line_pt_2)]
return (result[:2], result[2:]) | python | def _clip_line(
self,
line_pt_1,
line_pt_2
):
"""
clip line to canvas
"""
x_min = min(line_pt_1[0], line_pt_2[0])
x_max = max(line_pt_1[0], line_pt_2[0])
y_min = min(line_pt_1[1], line_pt_2[1])
y_max = max(line_pt_1[1], line_pt_2[1])
extent = self.extent()
if line_pt_1[0] == line_pt_2[0]:
return (
(line_pt_1[0], max(y_min, extent[1])),
(line_pt_1[0], min(y_max, extent[3]))
)
if line_pt_1[1] == line_pt_2[1]:
return (
(max(x_min, extent[0]), line_pt_1[1]),
(min(x_max, extent[2]), line_pt_1[1])
)
if ((extent[0] <= line_pt_1[0] < extent[2]) and
(extent[1] <= line_pt_1[1] < extent[3]) and
(extent[0] <= line_pt_2[0] < extent[2]) and
(extent[1] <= line_pt_2[1] < extent[3])):
return line_pt_1, line_pt_2
ts = [0.0,
1.0,
float(extent[0] - line_pt_1[0]) / (line_pt_2[0] - line_pt_1[0]),
float(extent[2] - line_pt_1[0]) / (line_pt_2[0] - line_pt_1[0]),
float(extent[1] - line_pt_1[1]) / (line_pt_2[1] - line_pt_1[1]),
float(extent[3] - line_pt_1[1]) / (line_pt_2[1] - line_pt_1[1])
]
ts.sort()
if (ts[2] < 0) or (ts[2] >= 1) or (ts[3] < 0) or (ts[2] >= 1):
return None
result =\
[(pt_1 + t * (pt_2 - pt_1))\
for t in (ts[2], ts[3])\
for (pt_1, pt_2) in zip(line_pt_1, line_pt_2)]
return (result[:2], result[2:]) | [
"def",
"_clip_line",
"(",
"self",
",",
"line_pt_1",
",",
"line_pt_2",
")",
":",
"x_min",
"=",
"min",
"(",
"line_pt_1",
"[",
"0",
"]",
",",
"line_pt_2",
"[",
"0",
"]",
")",
"x_max",
"=",
"max",
"(",
"line_pt_1",
"[",
"0",
"]",
",",
"line_pt_2",
"[",... | clip line to canvas | [
"clip",
"line",
"to",
"canvas"
] | b6f26287264632d6f8c9f8911aaf3a8e4fc4dcf5 | https://github.com/wdbm/datavision/blob/b6f26287264632d6f8c9f8911aaf3a8e4fc4dcf5/datavision.py#L2710-L2762 |
242,287 | atdt/afraid | afraid/__init__.py | get_dyndns_records | def get_dyndns_records(login, password):
"""Gets the set of dynamic DNS records associated with this account"""
params = dict(action='getdyndns', sha=get_auth_key(login, password))
response = requests.get('http://freedns.afraid.org/api/', params=params, timeout=timeout)
raw_records = (line.split('|') for line in response.content.split())
try:
records = frozenset(DnsRecord(*record) for record in raw_records)
except TypeError:
raise ApiError("Couldn't parse the server's response",
response.content)
return records | python | def get_dyndns_records(login, password):
"""Gets the set of dynamic DNS records associated with this account"""
params = dict(action='getdyndns', sha=get_auth_key(login, password))
response = requests.get('http://freedns.afraid.org/api/', params=params, timeout=timeout)
raw_records = (line.split('|') for line in response.content.split())
try:
records = frozenset(DnsRecord(*record) for record in raw_records)
except TypeError:
raise ApiError("Couldn't parse the server's response",
response.content)
return records | [
"def",
"get_dyndns_records",
"(",
"login",
",",
"password",
")",
":",
"params",
"=",
"dict",
"(",
"action",
"=",
"'getdyndns'",
",",
"sha",
"=",
"get_auth_key",
"(",
"login",
",",
"password",
")",
")",
"response",
"=",
"requests",
".",
"get",
"(",
"'http... | Gets the set of dynamic DNS records associated with this account | [
"Gets",
"the",
"set",
"of",
"dynamic",
"DNS",
"records",
"associated",
"with",
"this",
"account"
] | d74b2d4e41ed14e420da2793a89bef5d9b26ea26 | https://github.com/atdt/afraid/blob/d74b2d4e41ed14e420da2793a89bef5d9b26ea26/afraid/__init__.py#L91-L103 |
242,288 | atdt/afraid | afraid/__init__.py | update_continuously | def update_continuously(records, update_interval=600):
"""Update `records` every `update_interval` seconds"""
while True:
for record in records:
try:
record.update()
except (ApiError, RequestException):
pass
time.sleep(update_interval) | python | def update_continuously(records, update_interval=600):
"""Update `records` every `update_interval` seconds"""
while True:
for record in records:
try:
record.update()
except (ApiError, RequestException):
pass
time.sleep(update_interval) | [
"def",
"update_continuously",
"(",
"records",
",",
"update_interval",
"=",
"600",
")",
":",
"while",
"True",
":",
"for",
"record",
"in",
"records",
":",
"try",
":",
"record",
".",
"update",
"(",
")",
"except",
"(",
"ApiError",
",",
"RequestException",
")",... | Update `records` every `update_interval` seconds | [
"Update",
"records",
"every",
"update_interval",
"seconds"
] | d74b2d4e41ed14e420da2793a89bef5d9b26ea26 | https://github.com/atdt/afraid/blob/d74b2d4e41ed14e420da2793a89bef5d9b26ea26/afraid/__init__.py#L106-L114 |
242,289 | atdt/afraid | afraid/__init__.py | DnsRecord.update | def update(self):
"""Updates remote DNS record by requesting its special endpoint URL"""
response = requests.get(self.update_url, timeout=timeout)
match = ip_pattern.search(response.content)
# response must contain an ip address, or else we can't parse it
if not match:
raise ApiError("Couldn't parse the server's response",
response.content)
self.ip = match.group(0) | python | def update(self):
"""Updates remote DNS record by requesting its special endpoint URL"""
response = requests.get(self.update_url, timeout=timeout)
match = ip_pattern.search(response.content)
# response must contain an ip address, or else we can't parse it
if not match:
raise ApiError("Couldn't parse the server's response",
response.content)
self.ip = match.group(0) | [
"def",
"update",
"(",
"self",
")",
":",
"response",
"=",
"requests",
".",
"get",
"(",
"self",
".",
"update_url",
",",
"timeout",
"=",
"timeout",
")",
"match",
"=",
"ip_pattern",
".",
"search",
"(",
"response",
".",
"content",
")",
"# response must contain ... | Updates remote DNS record by requesting its special endpoint URL | [
"Updates",
"remote",
"DNS",
"record",
"by",
"requesting",
"its",
"special",
"endpoint",
"URL"
] | d74b2d4e41ed14e420da2793a89bef5d9b26ea26 | https://github.com/atdt/afraid/blob/d74b2d4e41ed14e420da2793a89bef5d9b26ea26/afraid/__init__.py#L72-L82 |
242,290 | inveniosoftware-attic/invenio-knowledge | invenio_knowledge/restful.py | setup_app | def setup_app(app, api):
"""setup the resources urls."""
api.add_resource(
KnwKBAllResource,
'/api/knowledge'
)
api.add_resource(
KnwKBResource,
'/api/knowledge/<string:slug>'
)
api.add_resource(
KnwKBMappingsResource,
'/api/knowledge/<string:slug>/mappings'
)
api.add_resource(
KnwKBMappingsToResource,
'/api/knowledge/<string:slug>/mappings/to'
)
api.add_resource(
KnwKBMappingsFromResource,
'/api/knowledge/<string:slug>/mappings/from'
)
# for other urls, return "Method Not Allowed"
api.add_resource(
NotImplementedKnowledegeResource,
'/api/knowledge/<string:slug>/<path:foo>'
) | python | def setup_app(app, api):
"""setup the resources urls."""
api.add_resource(
KnwKBAllResource,
'/api/knowledge'
)
api.add_resource(
KnwKBResource,
'/api/knowledge/<string:slug>'
)
api.add_resource(
KnwKBMappingsResource,
'/api/knowledge/<string:slug>/mappings'
)
api.add_resource(
KnwKBMappingsToResource,
'/api/knowledge/<string:slug>/mappings/to'
)
api.add_resource(
KnwKBMappingsFromResource,
'/api/knowledge/<string:slug>/mappings/from'
)
# for other urls, return "Method Not Allowed"
api.add_resource(
NotImplementedKnowledegeResource,
'/api/knowledge/<string:slug>/<path:foo>'
) | [
"def",
"setup_app",
"(",
"app",
",",
"api",
")",
":",
"api",
".",
"add_resource",
"(",
"KnwKBAllResource",
",",
"'/api/knowledge'",
")",
"api",
".",
"add_resource",
"(",
"KnwKBResource",
",",
"'/api/knowledge/<string:slug>'",
")",
"api",
".",
"add_resource",
"("... | setup the resources urls. | [
"setup",
"the",
"resources",
"urls",
"."
] | b31722dc14243ca8f626f8b3bce9718d0119de55 | https://github.com/inveniosoftware-attic/invenio-knowledge/blob/b31722dc14243ca8f626f8b3bce9718d0119de55/invenio_knowledge/restful.py#L487-L514 |
242,291 | inveniosoftware-attic/invenio-knowledge | invenio_knowledge/restful.py | KnwKBResource.get | def get(self, slug):
"""Get KnwKB.
Url parameters:
- from: filter "mappings from"
- to: filter "mappings to"
- page
- per_page
- match_type: s=substring, e=exact, sw=startswith
- sortby: 'from' or 'to'
"""
kb = api.get_kb_by_slug(slug)
# check if is accessible from api
check_knowledge_access(kb)
parser = reqparse.RequestParser()
parser.add_argument(
'from', type=str,
help="Return only entries where key matches this.")
parser.add_argument(
'to', type=str,
help="Return only entries where value matches this.")
parser.add_argument('page', type=int,
help="Require a specific page")
parser.add_argument('per_page', type=int,
help="Set how much result per page")
parser.add_argument('match_type', type=str,
help="s=substring, e=exact, sw=startswith")
parser.add_argument('sortby', type=str,
help="the sorting criteria ('from' or 'to')")
args = parser.parse_args()
kb_dict = kb.to_dict()
kb_dict['mappings'] = KnwKBMappingsResource \
.search_mappings(kb=kb, key=args['from'], value=args['to'],
match_type=args['match_type'],
sortby=args['sortby'], page=args['page'],
per_page=args['per_page'])
return kb_dict | python | def get(self, slug):
"""Get KnwKB.
Url parameters:
- from: filter "mappings from"
- to: filter "mappings to"
- page
- per_page
- match_type: s=substring, e=exact, sw=startswith
- sortby: 'from' or 'to'
"""
kb = api.get_kb_by_slug(slug)
# check if is accessible from api
check_knowledge_access(kb)
parser = reqparse.RequestParser()
parser.add_argument(
'from', type=str,
help="Return only entries where key matches this.")
parser.add_argument(
'to', type=str,
help="Return only entries where value matches this.")
parser.add_argument('page', type=int,
help="Require a specific page")
parser.add_argument('per_page', type=int,
help="Set how much result per page")
parser.add_argument('match_type', type=str,
help="s=substring, e=exact, sw=startswith")
parser.add_argument('sortby', type=str,
help="the sorting criteria ('from' or 'to')")
args = parser.parse_args()
kb_dict = kb.to_dict()
kb_dict['mappings'] = KnwKBMappingsResource \
.search_mappings(kb=kb, key=args['from'], value=args['to'],
match_type=args['match_type'],
sortby=args['sortby'], page=args['page'],
per_page=args['per_page'])
return kb_dict | [
"def",
"get",
"(",
"self",
",",
"slug",
")",
":",
"kb",
"=",
"api",
".",
"get_kb_by_slug",
"(",
"slug",
")",
"# check if is accessible from api",
"check_knowledge_access",
"(",
"kb",
")",
"parser",
"=",
"reqparse",
".",
"RequestParser",
"(",
")",
"parser",
"... | Get KnwKB.
Url parameters:
- from: filter "mappings from"
- to: filter "mappings to"
- page
- per_page
- match_type: s=substring, e=exact, sw=startswith
- sortby: 'from' or 'to' | [
"Get",
"KnwKB",
"."
] | b31722dc14243ca8f626f8b3bce9718d0119de55 | https://github.com/inveniosoftware-attic/invenio-knowledge/blob/b31722dc14243ca8f626f8b3bce9718d0119de55/invenio_knowledge/restful.py#L106-L144 |
242,292 | inveniosoftware-attic/invenio-knowledge | invenio_knowledge/restful.py | KnwKBMappingsResource.search_mappings | def search_mappings(kb, key=None, value=None, match_type=None,
sortby=None, page=None, per_page=None):
"""Search tags for knowledge."""
if kb.kbtype == models.KnwKB.KNWKB_TYPES['written_as']:
return pagination.RestfulSQLAlchemyPagination(
api.query_kb_mappings(
kbid=kb.id,
key=key or '',
value=value or '',
match_type=match_type or 's',
sortby=sortby or 'to',
), page=page or 1, per_page=per_page or 10
).items
return [] | python | def search_mappings(kb, key=None, value=None, match_type=None,
sortby=None, page=None, per_page=None):
"""Search tags for knowledge."""
if kb.kbtype == models.KnwKB.KNWKB_TYPES['written_as']:
return pagination.RestfulSQLAlchemyPagination(
api.query_kb_mappings(
kbid=kb.id,
key=key or '',
value=value or '',
match_type=match_type or 's',
sortby=sortby or 'to',
), page=page or 1, per_page=per_page or 10
).items
return [] | [
"def",
"search_mappings",
"(",
"kb",
",",
"key",
"=",
"None",
",",
"value",
"=",
"None",
",",
"match_type",
"=",
"None",
",",
"sortby",
"=",
"None",
",",
"page",
"=",
"None",
",",
"per_page",
"=",
"None",
")",
":",
"if",
"kb",
".",
"kbtype",
"==",
... | Search tags for knowledge. | [
"Search",
"tags",
"for",
"knowledge",
"."
] | b31722dc14243ca8f626f8b3bce9718d0119de55 | https://github.com/inveniosoftware-attic/invenio-knowledge/blob/b31722dc14243ca8f626f8b3bce9718d0119de55/invenio_knowledge/restful.py#L185-L198 |
242,293 | inveniosoftware-attic/invenio-knowledge | invenio_knowledge/restful.py | KnwKBMappingsResource.get | def get(self, slug):
"""Get list of mappings.
Url parameters:
- from: filter "mappings from"
- to: filter "mappings to"
- page
- per_page
- match_type: s=substring, e=exact, sw=startswith
- sortby: 'from' or 'to'
"""
kb = api.get_kb_by_slug(slug)
# check if is accessible from api
check_knowledge_access(kb)
parser = reqparse.RequestParser()
parser.add_argument(
'from', type=str,
help="Return only entries where 'from' matches this.")
parser.add_argument(
'to', type=str,
help="Return only entries where 'to' matches this.")
parser.add_argument('page', type=int,
help="Require a specific page")
parser.add_argument('per_page', type=int,
help="Set how much result per page")
parser.add_argument('match_type', type=str,
help="s=substring, e=exact, sw=startswith")
parser.add_argument('sortby', type=str,
help="the sorting criteria ('from' or 'to')")
args = parser.parse_args()
return KnwKBMappingsResource \
.search_mappings(kb, args['from'], args['to'],
args['match_type'], args['sortby'],
args['page'], args['per_page']) | python | def get(self, slug):
"""Get list of mappings.
Url parameters:
- from: filter "mappings from"
- to: filter "mappings to"
- page
- per_page
- match_type: s=substring, e=exact, sw=startswith
- sortby: 'from' or 'to'
"""
kb = api.get_kb_by_slug(slug)
# check if is accessible from api
check_knowledge_access(kb)
parser = reqparse.RequestParser()
parser.add_argument(
'from', type=str,
help="Return only entries where 'from' matches this.")
parser.add_argument(
'to', type=str,
help="Return only entries where 'to' matches this.")
parser.add_argument('page', type=int,
help="Require a specific page")
parser.add_argument('per_page', type=int,
help="Set how much result per page")
parser.add_argument('match_type', type=str,
help="s=substring, e=exact, sw=startswith")
parser.add_argument('sortby', type=str,
help="the sorting criteria ('from' or 'to')")
args = parser.parse_args()
return KnwKBMappingsResource \
.search_mappings(kb, args['from'], args['to'],
args['match_type'], args['sortby'],
args['page'], args['per_page']) | [
"def",
"get",
"(",
"self",
",",
"slug",
")",
":",
"kb",
"=",
"api",
".",
"get_kb_by_slug",
"(",
"slug",
")",
"# check if is accessible from api",
"check_knowledge_access",
"(",
"kb",
")",
"parser",
"=",
"reqparse",
".",
"RequestParser",
"(",
")",
"parser",
"... | Get list of mappings.
Url parameters:
- from: filter "mappings from"
- to: filter "mappings to"
- page
- per_page
- match_type: s=substring, e=exact, sw=startswith
- sortby: 'from' or 'to' | [
"Get",
"list",
"of",
"mappings",
"."
] | b31722dc14243ca8f626f8b3bce9718d0119de55 | https://github.com/inveniosoftware-attic/invenio-knowledge/blob/b31722dc14243ca8f626f8b3bce9718d0119de55/invenio_knowledge/restful.py#L201-L236 |
242,294 | inveniosoftware-attic/invenio-knowledge | invenio_knowledge/restful.py | KnwKBMappingsToResource.search_list | def search_list(kb, value=None, match_type=None,
page=None, per_page=None, unique=False):
"""Search "mappings to" for knowledge."""
# init
page = page or 1
per_page = per_page or 10
if kb.kbtype == models.KnwKB.KNWKB_TYPES['written_as']:
# get the base query
query = api.query_kb_mappings(
kbid=kb.id,
value=value or '',
match_type=match_type or 's'
).with_entities(models.KnwKBRVAL.m_value)
# if you want a 'unique' list
if unique:
query = query.distinct()
# run query and paginate
return [item.m_value for item in
pagination.RestfulSQLAlchemyPagination(
query, page=page or 1,
per_page=per_page or 10
).items]
elif kb.kbtype == models.KnwKB.KNWKB_TYPES['dynamic']:
items = api.get_kbd_values(kb.name, value)
return pagination.RestfulPagination(
page=page, per_page=per_page,
total_count=len(items)
).slice(items)
return [] | python | def search_list(kb, value=None, match_type=None,
page=None, per_page=None, unique=False):
"""Search "mappings to" for knowledge."""
# init
page = page or 1
per_page = per_page or 10
if kb.kbtype == models.KnwKB.KNWKB_TYPES['written_as']:
# get the base query
query = api.query_kb_mappings(
kbid=kb.id,
value=value or '',
match_type=match_type or 's'
).with_entities(models.KnwKBRVAL.m_value)
# if you want a 'unique' list
if unique:
query = query.distinct()
# run query and paginate
return [item.m_value for item in
pagination.RestfulSQLAlchemyPagination(
query, page=page or 1,
per_page=per_page or 10
).items]
elif kb.kbtype == models.KnwKB.KNWKB_TYPES['dynamic']:
items = api.get_kbd_values(kb.name, value)
return pagination.RestfulPagination(
page=page, per_page=per_page,
total_count=len(items)
).slice(items)
return [] | [
"def",
"search_list",
"(",
"kb",
",",
"value",
"=",
"None",
",",
"match_type",
"=",
"None",
",",
"page",
"=",
"None",
",",
"per_page",
"=",
"None",
",",
"unique",
"=",
"False",
")",
":",
"# init",
"page",
"=",
"page",
"or",
"1",
"per_page",
"=",
"p... | Search "mappings to" for knowledge. | [
"Search",
"mappings",
"to",
"for",
"knowledge",
"."
] | b31722dc14243ca8f626f8b3bce9718d0119de55 | https://github.com/inveniosoftware-attic/invenio-knowledge/blob/b31722dc14243ca8f626f8b3bce9718d0119de55/invenio_knowledge/restful.py#L277-L306 |
242,295 | inveniosoftware-attic/invenio-knowledge | invenio_knowledge/restful.py | KnwKBMappingsFromResource.search_list | def search_list(kb, from_=None, match_type=None,
page=None, per_page=None, unique=False):
"""Search "mapping from" for knowledge."""
# init
page = page or 1
per_page = per_page or 10
if kb.kbtype == models.KnwKB.KNWKB_TYPES['written_as']:
# get the base query
query = api.query_kb_mappings(
kbid=kb.id,
key=from_ or '',
match_type=match_type or 's'
).with_entities(models.KnwKBRVAL.m_key)
# if you want a 'unique' list
if unique:
query = query.distinct()
# run query and paginate
return [item.m_key for item in
pagination.RestfulSQLAlchemyPagination(
query, page=page or 1,
per_page=per_page or 10
).items]
return [] | python | def search_list(kb, from_=None, match_type=None,
page=None, per_page=None, unique=False):
"""Search "mapping from" for knowledge."""
# init
page = page or 1
per_page = per_page or 10
if kb.kbtype == models.KnwKB.KNWKB_TYPES['written_as']:
# get the base query
query = api.query_kb_mappings(
kbid=kb.id,
key=from_ or '',
match_type=match_type or 's'
).with_entities(models.KnwKBRVAL.m_key)
# if you want a 'unique' list
if unique:
query = query.distinct()
# run query and paginate
return [item.m_key for item in
pagination.RestfulSQLAlchemyPagination(
query, page=page or 1,
per_page=per_page or 10
).items]
return [] | [
"def",
"search_list",
"(",
"kb",
",",
"from_",
"=",
"None",
",",
"match_type",
"=",
"None",
",",
"page",
"=",
"None",
",",
"per_page",
"=",
"None",
",",
"unique",
"=",
"False",
")",
":",
"# init",
"page",
"=",
"page",
"or",
"1",
"per_page",
"=",
"p... | Search "mapping from" for knowledge. | [
"Search",
"mapping",
"from",
"for",
"knowledge",
"."
] | b31722dc14243ca8f626f8b3bce9718d0119de55 | https://github.com/inveniosoftware-attic/invenio-knowledge/blob/b31722dc14243ca8f626f8b3bce9718d0119de55/invenio_knowledge/restful.py#L382-L405 |
242,296 | inveniosoftware-attic/invenio-knowledge | invenio_knowledge/restful.py | KnwKBMappingsFromResource.get | def get(self, slug):
"""Get list of "mappings from".
Url parameters
- unique: if set, return a unique list
- filter: filter "mappings from"
- page
- per_page
- match_type: s=substring, e=exact, sw=startswith
"""
kb = api.get_kb_by_slug(slug)
# check if is accessible from api
check_knowledge_access(kb)
parser = reqparse.RequestParser()
parser.add_argument(
'unique', type=bool,
help="The list contains unique names of 'mapping to'")
parser.add_argument(
'filter', type=str,
help="Return only entries where 'from' matches this.")
parser.add_argument('page', type=int,
help="Require a specific page")
parser.add_argument('per_page', type=int,
help="Set how much result per page")
parser.add_argument('match_type', type=str,
help="s=substring, e=exact, sw=startswith")
args = parser.parse_args()
return KnwKBMappingsFromResource \
.search_list(kb, args['filter'],
args['match_type'],
args['page'], args['per_page'], args['unique']) | python | def get(self, slug):
"""Get list of "mappings from".
Url parameters
- unique: if set, return a unique list
- filter: filter "mappings from"
- page
- per_page
- match_type: s=substring, e=exact, sw=startswith
"""
kb = api.get_kb_by_slug(slug)
# check if is accessible from api
check_knowledge_access(kb)
parser = reqparse.RequestParser()
parser.add_argument(
'unique', type=bool,
help="The list contains unique names of 'mapping to'")
parser.add_argument(
'filter', type=str,
help="Return only entries where 'from' matches this.")
parser.add_argument('page', type=int,
help="Require a specific page")
parser.add_argument('per_page', type=int,
help="Set how much result per page")
parser.add_argument('match_type', type=str,
help="s=substring, e=exact, sw=startswith")
args = parser.parse_args()
return KnwKBMappingsFromResource \
.search_list(kb, args['filter'],
args['match_type'],
args['page'], args['per_page'], args['unique']) | [
"def",
"get",
"(",
"self",
",",
"slug",
")",
":",
"kb",
"=",
"api",
".",
"get_kb_by_slug",
"(",
"slug",
")",
"# check if is accessible from api",
"check_knowledge_access",
"(",
"kb",
")",
"parser",
"=",
"reqparse",
".",
"RequestParser",
"(",
")",
"parser",
"... | Get list of "mappings from".
Url parameters
- unique: if set, return a unique list
- filter: filter "mappings from"
- page
- per_page
- match_type: s=substring, e=exact, sw=startswith | [
"Get",
"list",
"of",
"mappings",
"from",
"."
] | b31722dc14243ca8f626f8b3bce9718d0119de55 | https://github.com/inveniosoftware-attic/invenio-knowledge/blob/b31722dc14243ca8f626f8b3bce9718d0119de55/invenio_knowledge/restful.py#L407-L439 |
242,297 | political-memory/django-representatives | representatives/contrib/parltrack/import_representatives.py | ParltrackImporter.manage_mep | def manage_mep(self, mep_json):
'''
Import a mep as a representative from the json dict fetched from
parltrack
'''
# Some versions of memopol will connect to this and skip inactive meps.
responses = representative_pre_import.send(sender=self,
representative_data=mep_json)
for receiver, response in responses:
if response is False:
logger.debug(
'Skipping MEP %s', mep_json['Name']['full'])
return
changed = False
slug = slugify('%s-%s' % (
mep_json["Name"]["full"] if 'full' in mep_json["Name"]
else mep_json["Name"]["sur"] + " " + mep_json["Name"]["family"],
_parse_date(mep_json["Birth"]["date"])
))
try:
representative = Representative.objects.get(slug=slug)
except Representative.DoesNotExist:
representative = Representative(slug=slug)
changed = True
# Save representative attributes
self.import_representative_details(representative, mep_json, changed)
self.add_mandates(representative, mep_json)
self.add_contacts(representative, mep_json)
logger.debug('Imported MEP %s', unicode(representative))
return representative | python | def manage_mep(self, mep_json):
'''
Import a mep as a representative from the json dict fetched from
parltrack
'''
# Some versions of memopol will connect to this and skip inactive meps.
responses = representative_pre_import.send(sender=self,
representative_data=mep_json)
for receiver, response in responses:
if response is False:
logger.debug(
'Skipping MEP %s', mep_json['Name']['full'])
return
changed = False
slug = slugify('%s-%s' % (
mep_json["Name"]["full"] if 'full' in mep_json["Name"]
else mep_json["Name"]["sur"] + " " + mep_json["Name"]["family"],
_parse_date(mep_json["Birth"]["date"])
))
try:
representative = Representative.objects.get(slug=slug)
except Representative.DoesNotExist:
representative = Representative(slug=slug)
changed = True
# Save representative attributes
self.import_representative_details(representative, mep_json, changed)
self.add_mandates(representative, mep_json)
self.add_contacts(representative, mep_json)
logger.debug('Imported MEP %s', unicode(representative))
return representative | [
"def",
"manage_mep",
"(",
"self",
",",
"mep_json",
")",
":",
"# Some versions of memopol will connect to this and skip inactive meps.",
"responses",
"=",
"representative_pre_import",
".",
"send",
"(",
"sender",
"=",
"self",
",",
"representative_data",
"=",
"mep_json",
")"... | Import a mep as a representative from the json dict fetched from
parltrack | [
"Import",
"a",
"mep",
"as",
"a",
"representative",
"from",
"the",
"json",
"dict",
"fetched",
"from",
"parltrack"
] | 811c90d0250149e913e6196f0ab11c97d396be39 | https://github.com/political-memory/django-representatives/blob/811c90d0250149e913e6196f0ab11c97d396be39/representatives/contrib/parltrack/import_representatives.py#L77-L114 |
242,298 | edeposit/edeposit.amqp.ltp | src/edeposit/amqp/ltp/info_composer.py | _calc_dir_size | def _calc_dir_size(path):
"""
Calculate size of all files in `path`.
Args:
path (str): Path to the directory.
Returns:
int: Size of the directory in bytes.
"""
dir_size = 0
for (root, dirs, files) in os.walk(path):
for fn in files:
full_fn = os.path.join(root, fn)
dir_size += os.path.getsize(full_fn)
return dir_size | python | def _calc_dir_size(path):
"""
Calculate size of all files in `path`.
Args:
path (str): Path to the directory.
Returns:
int: Size of the directory in bytes.
"""
dir_size = 0
for (root, dirs, files) in os.walk(path):
for fn in files:
full_fn = os.path.join(root, fn)
dir_size += os.path.getsize(full_fn)
return dir_size | [
"def",
"_calc_dir_size",
"(",
"path",
")",
":",
"dir_size",
"=",
"0",
"for",
"(",
"root",
",",
"dirs",
",",
"files",
")",
"in",
"os",
".",
"walk",
"(",
"path",
")",
":",
"for",
"fn",
"in",
"files",
":",
"full_fn",
"=",
"os",
".",
"path",
".",
"... | Calculate size of all files in `path`.
Args:
path (str): Path to the directory.
Returns:
int: Size of the directory in bytes. | [
"Calculate",
"size",
"of",
"all",
"files",
"in",
"path",
"."
] | df9ac7ec6cbdbeaaeed438ca66df75ea967b6d8e | https://github.com/edeposit/edeposit.amqp.ltp/blob/df9ac7ec6cbdbeaaeed438ca66df75ea967b6d8e/src/edeposit/amqp/ltp/info_composer.py#L38-L54 |
242,299 | edeposit/edeposit.amqp.ltp | src/edeposit/amqp/ltp/info_composer.py | _get_localized_fn | def _get_localized_fn(path, root_dir):
"""
Return absolute `path` relative to `root_dir`.
When `path` == ``/home/xex/somefile.txt`` and `root_dir` == ``/home``,
returned path will be ``/xex/somefile.txt``.
Args:
path (str): Absolute path beginning in `root_dir`.
root_dir (str): Absolute path containing `path` argument.
Returns:
str: Local `path` when `root_dir` is considered as root of FS.
"""
local_fn = path
if path.startswith(root_dir):
local_fn = path.replace(root_dir, "", 1)
if not local_fn.startswith("/"):
return "/" + local_fn
return local_fn | python | def _get_localized_fn(path, root_dir):
"""
Return absolute `path` relative to `root_dir`.
When `path` == ``/home/xex/somefile.txt`` and `root_dir` == ``/home``,
returned path will be ``/xex/somefile.txt``.
Args:
path (str): Absolute path beginning in `root_dir`.
root_dir (str): Absolute path containing `path` argument.
Returns:
str: Local `path` when `root_dir` is considered as root of FS.
"""
local_fn = path
if path.startswith(root_dir):
local_fn = path.replace(root_dir, "", 1)
if not local_fn.startswith("/"):
return "/" + local_fn
return local_fn | [
"def",
"_get_localized_fn",
"(",
"path",
",",
"root_dir",
")",
":",
"local_fn",
"=",
"path",
"if",
"path",
".",
"startswith",
"(",
"root_dir",
")",
":",
"local_fn",
"=",
"path",
".",
"replace",
"(",
"root_dir",
",",
"\"\"",
",",
"1",
")",
"if",
"not",
... | Return absolute `path` relative to `root_dir`.
When `path` == ``/home/xex/somefile.txt`` and `root_dir` == ``/home``,
returned path will be ``/xex/somefile.txt``.
Args:
path (str): Absolute path beginning in `root_dir`.
root_dir (str): Absolute path containing `path` argument.
Returns:
str: Local `path` when `root_dir` is considered as root of FS. | [
"Return",
"absolute",
"path",
"relative",
"to",
"root_dir",
"."
] | df9ac7ec6cbdbeaaeed438ca66df75ea967b6d8e | https://github.com/edeposit/edeposit.amqp.ltp/blob/df9ac7ec6cbdbeaaeed438ca66df75ea967b6d8e/src/edeposit/amqp/ltp/info_composer.py#L57-L78 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.