id int32 0 252k | repo stringlengths 7 55 | path stringlengths 4 127 | func_name stringlengths 1 88 | original_string stringlengths 75 19.8k | language stringclasses 1
value | code stringlengths 75 19.8k | code_tokens list | docstring stringlengths 3 17.3k | docstring_tokens list | sha stringlengths 40 40 | url stringlengths 87 242 |
|---|---|---|---|---|---|---|---|---|---|---|---|
236,900 | closeio/tasktiger | tasktiger/worker.py | Worker._execute_task_group | def _execute_task_group(self, queue, tasks, all_task_ids, queue_lock):
"""
Executes the given tasks in the queue. Updates the heartbeat for task
IDs passed in all_task_ids. This internal method is only meant to be
called from within _process_from_queue.
"""
log = self.log... | python | def _execute_task_group(self, queue, tasks, all_task_ids, queue_lock):
"""
Executes the given tasks in the queue. Updates the heartbeat for task
IDs passed in all_task_ids. This internal method is only meant to be
called from within _process_from_queue.
"""
log = self.log... | [
"def",
"_execute_task_group",
"(",
"self",
",",
"queue",
",",
"tasks",
",",
"all_task_ids",
",",
"queue_lock",
")",
":",
"log",
"=",
"self",
".",
"log",
".",
"bind",
"(",
"queue",
"=",
"queue",
")",
"locks",
"=",
"[",
"]",
"# Keep track of the acquired loc... | Executes the given tasks in the queue. Updates the heartbeat for task
IDs passed in all_task_ids. This internal method is only meant to be
called from within _process_from_queue. | [
"Executes",
"the",
"given",
"tasks",
"in",
"the",
"queue",
".",
"Updates",
"the",
"heartbeat",
"for",
"task",
"IDs",
"passed",
"in",
"all_task_ids",
".",
"This",
"internal",
"method",
"is",
"only",
"meant",
"to",
"be",
"called",
"from",
"within",
"_process_f... | 59f893152d6eb4b7f1f62fc4b35aeeca7f26c07a | https://github.com/closeio/tasktiger/blob/59f893152d6eb4b7f1f62fc4b35aeeca7f26c07a/tasktiger/worker.py#L692-L757 |
236,901 | closeio/tasktiger | tasktiger/worker.py | Worker._finish_task_processing | def _finish_task_processing(self, queue, task, success):
"""
After a task is executed, this method is called and ensures that
the task gets properly removed from the ACTIVE queue and, in case of an
error, retried or marked as failed.
"""
log = self.log.bind(queue=queue, t... | python | def _finish_task_processing(self, queue, task, success):
"""
After a task is executed, this method is called and ensures that
the task gets properly removed from the ACTIVE queue and, in case of an
error, retried or marked as failed.
"""
log = self.log.bind(queue=queue, t... | [
"def",
"_finish_task_processing",
"(",
"self",
",",
"queue",
",",
"task",
",",
"success",
")",
":",
"log",
"=",
"self",
".",
"log",
".",
"bind",
"(",
"queue",
"=",
"queue",
",",
"task_id",
"=",
"task",
".",
"id",
")",
"def",
"_mark_done",
"(",
")",
... | After a task is executed, this method is called and ensures that
the task gets properly removed from the ACTIVE queue and, in case of an
error, retried or marked as failed. | [
"After",
"a",
"task",
"is",
"executed",
"this",
"method",
"is",
"called",
"and",
"ensures",
"that",
"the",
"task",
"gets",
"properly",
"removed",
"from",
"the",
"ACTIVE",
"queue",
"and",
"in",
"case",
"of",
"an",
"error",
"retried",
"or",
"marked",
"as",
... | 59f893152d6eb4b7f1f62fc4b35aeeca7f26c07a | https://github.com/closeio/tasktiger/blob/59f893152d6eb4b7f1f62fc4b35aeeca7f26c07a/tasktiger/worker.py#L759-L859 |
236,902 | closeio/tasktiger | tasktiger/worker.py | Worker.run | def run(self, once=False, force_once=False):
"""
Main loop of the worker.
Use once=True to execute any queued tasks and then exit.
Use force_once=True with once=True to always exit after one processing
loop even if tasks remain queued.
"""
self.log.info('ready',... | python | def run(self, once=False, force_once=False):
"""
Main loop of the worker.
Use once=True to execute any queued tasks and then exit.
Use force_once=True with once=True to always exit after one processing
loop even if tasks remain queued.
"""
self.log.info('ready',... | [
"def",
"run",
"(",
"self",
",",
"once",
"=",
"False",
",",
"force_once",
"=",
"False",
")",
":",
"self",
".",
"log",
".",
"info",
"(",
"'ready'",
",",
"id",
"=",
"self",
".",
"id",
",",
"queues",
"=",
"sorted",
"(",
"self",
".",
"only_queues",
")... | Main loop of the worker.
Use once=True to execute any queued tasks and then exit.
Use force_once=True with once=True to always exit after one processing
loop even if tasks remain queued. | [
"Main",
"loop",
"of",
"the",
"worker",
"."
] | 59f893152d6eb4b7f1f62fc4b35aeeca7f26c07a | https://github.com/closeio/tasktiger/blob/59f893152d6eb4b7f1f62fc4b35aeeca7f26c07a/tasktiger/worker.py#L938-L1004 |
236,903 | closeio/tasktiger | tasktiger/redis_scripts.py | RedisScripts.can_replicate_commands | def can_replicate_commands(self):
"""
Whether Redis supports single command replication.
"""
if not hasattr(self, '_can_replicate_commands'):
info = self.redis.info('server')
version_info = info['redis_version'].split('.')
major, minor = int(version_in... | python | def can_replicate_commands(self):
"""
Whether Redis supports single command replication.
"""
if not hasattr(self, '_can_replicate_commands'):
info = self.redis.info('server')
version_info = info['redis_version'].split('.')
major, minor = int(version_in... | [
"def",
"can_replicate_commands",
"(",
"self",
")",
":",
"if",
"not",
"hasattr",
"(",
"self",
",",
"'_can_replicate_commands'",
")",
":",
"info",
"=",
"self",
".",
"redis",
".",
"info",
"(",
"'server'",
")",
"version_info",
"=",
"info",
"[",
"'redis_version'"... | Whether Redis supports single command replication. | [
"Whether",
"Redis",
"supports",
"single",
"command",
"replication",
"."
] | 59f893152d6eb4b7f1f62fc4b35aeeca7f26c07a | https://github.com/closeio/tasktiger/blob/59f893152d6eb4b7f1f62fc4b35aeeca7f26c07a/tasktiger/redis_scripts.py#L301-L311 |
236,904 | closeio/tasktiger | tasktiger/redis_scripts.py | RedisScripts.zpoppush | def zpoppush(self, source, destination, count, score, new_score,
client=None, withscores=False, on_success=None,
if_exists=None):
"""
Pops the first ``count`` members from the ZSET ``source`` and adds them
to the ZSET ``destination`` with a score of ``new_score`... | python | def zpoppush(self, source, destination, count, score, new_score,
client=None, withscores=False, on_success=None,
if_exists=None):
"""
Pops the first ``count`` members from the ZSET ``source`` and adds them
to the ZSET ``destination`` with a score of ``new_score`... | [
"def",
"zpoppush",
"(",
"self",
",",
"source",
",",
"destination",
",",
"count",
",",
"score",
",",
"new_score",
",",
"client",
"=",
"None",
",",
"withscores",
"=",
"False",
",",
"on_success",
"=",
"None",
",",
"if_exists",
"=",
"None",
")",
":",
"if",... | Pops the first ``count`` members from the ZSET ``source`` and adds them
to the ZSET ``destination`` with a score of ``new_score``. If ``score``
is not None, only members up to a score of ``score`` are used. Returns
the members that were moved and, if ``withscores`` is True, their
origina... | [
"Pops",
"the",
"first",
"count",
"members",
"from",
"the",
"ZSET",
"source",
"and",
"adds",
"them",
"to",
"the",
"ZSET",
"destination",
"with",
"a",
"score",
"of",
"new_score",
".",
"If",
"score",
"is",
"not",
"None",
"only",
"members",
"up",
"to",
"a",
... | 59f893152d6eb4b7f1f62fc4b35aeeca7f26c07a | https://github.com/closeio/tasktiger/blob/59f893152d6eb4b7f1f62fc4b35aeeca7f26c07a/tasktiger/redis_scripts.py#L339-L423 |
236,905 | closeio/tasktiger | tasktiger/redis_scripts.py | RedisScripts.execute_pipeline | def execute_pipeline(self, pipeline, client=None):
"""
Executes the given Redis pipeline as a Lua script. When an error
occurs, the transaction stops executing, and an exception is raised.
This differs from Redis transactions, where execution continues after an
error. On success,... | python | def execute_pipeline(self, pipeline, client=None):
"""
Executes the given Redis pipeline as a Lua script. When an error
occurs, the transaction stops executing, and an exception is raised.
This differs from Redis transactions, where execution continues after an
error. On success,... | [
"def",
"execute_pipeline",
"(",
"self",
",",
"pipeline",
",",
"client",
"=",
"None",
")",
":",
"client",
"=",
"client",
"or",
"self",
".",
"redis",
"executing_pipeline",
"=",
"None",
"try",
":",
"# Prepare args",
"stack",
"=",
"pipeline",
".",
"command_stack... | Executes the given Redis pipeline as a Lua script. When an error
occurs, the transaction stops executing, and an exception is raised.
This differs from Redis transactions, where execution continues after an
error. On success, a list of results is returned. The pipeline is
cleared after e... | [
"Executes",
"the",
"given",
"Redis",
"pipeline",
"as",
"a",
"Lua",
"script",
".",
"When",
"an",
"error",
"occurs",
"the",
"transaction",
"stops",
"executing",
"and",
"an",
"exception",
"is",
"raised",
".",
"This",
"differs",
"from",
"Redis",
"transactions",
... | 59f893152d6eb4b7f1f62fc4b35aeeca7f26c07a | https://github.com/closeio/tasktiger/blob/59f893152d6eb4b7f1f62fc4b35aeeca7f26c07a/tasktiger/redis_scripts.py#L466-L534 |
236,906 | closeio/tasktiger | tasktiger/_internal.py | gen_unique_id | def gen_unique_id(serialized_name, args, kwargs):
"""
Generates and returns a hex-encoded 256-bit ID for the given task name and
args. Used to generate IDs for unique tasks or for task locks.
"""
return hashlib.sha256(json.dumps({
'func': serialized_name,
'args': args,
'kwarg... | python | def gen_unique_id(serialized_name, args, kwargs):
"""
Generates and returns a hex-encoded 256-bit ID for the given task name and
args. Used to generate IDs for unique tasks or for task locks.
"""
return hashlib.sha256(json.dumps({
'func': serialized_name,
'args': args,
'kwarg... | [
"def",
"gen_unique_id",
"(",
"serialized_name",
",",
"args",
",",
"kwargs",
")",
":",
"return",
"hashlib",
".",
"sha256",
"(",
"json",
".",
"dumps",
"(",
"{",
"'func'",
":",
"serialized_name",
",",
"'args'",
":",
"args",
",",
"'kwargs'",
":",
"kwargs",
"... | Generates and returns a hex-encoded 256-bit ID for the given task name and
args. Used to generate IDs for unique tasks or for task locks. | [
"Generates",
"and",
"returns",
"a",
"hex",
"-",
"encoded",
"256",
"-",
"bit",
"ID",
"for",
"the",
"given",
"task",
"name",
"and",
"args",
".",
"Used",
"to",
"generate",
"IDs",
"for",
"unique",
"tasks",
"or",
"for",
"task",
"locks",
"."
] | 59f893152d6eb4b7f1f62fc4b35aeeca7f26c07a | https://github.com/closeio/tasktiger/blob/59f893152d6eb4b7f1f62fc4b35aeeca7f26c07a/tasktiger/_internal.py#L56-L65 |
236,907 | closeio/tasktiger | tasktiger/_internal.py | serialize_func_name | def serialize_func_name(func):
"""
Returns the dotted serialized path to the passed function.
"""
if func.__module__ == '__main__':
raise ValueError('Functions from the __main__ module cannot be '
'processed by workers.')
try:
# This will only work on Python ... | python | def serialize_func_name(func):
"""
Returns the dotted serialized path to the passed function.
"""
if func.__module__ == '__main__':
raise ValueError('Functions from the __main__ module cannot be '
'processed by workers.')
try:
# This will only work on Python ... | [
"def",
"serialize_func_name",
"(",
"func",
")",
":",
"if",
"func",
".",
"__module__",
"==",
"'__main__'",
":",
"raise",
"ValueError",
"(",
"'Functions from the __main__ module cannot be '",
"'processed by workers.'",
")",
"try",
":",
"# This will only work on Python 3.3 or ... | Returns the dotted serialized path to the passed function. | [
"Returns",
"the",
"dotted",
"serialized",
"path",
"to",
"the",
"passed",
"function",
"."
] | 59f893152d6eb4b7f1f62fc4b35aeeca7f26c07a | https://github.com/closeio/tasktiger/blob/59f893152d6eb4b7f1f62fc4b35aeeca7f26c07a/tasktiger/_internal.py#L67-L79 |
236,908 | closeio/tasktiger | tasktiger/_internal.py | dotted_parts | def dotted_parts(s):
"""
For a string "a.b.c", yields "a", "a.b", "a.b.c".
"""
idx = -1
while s:
idx = s.find('.', idx+1)
if idx == -1:
yield s
break
yield s[:idx] | python | def dotted_parts(s):
"""
For a string "a.b.c", yields "a", "a.b", "a.b.c".
"""
idx = -1
while s:
idx = s.find('.', idx+1)
if idx == -1:
yield s
break
yield s[:idx] | [
"def",
"dotted_parts",
"(",
"s",
")",
":",
"idx",
"=",
"-",
"1",
"while",
"s",
":",
"idx",
"=",
"s",
".",
"find",
"(",
"'.'",
",",
"idx",
"+",
"1",
")",
"if",
"idx",
"==",
"-",
"1",
":",
"yield",
"s",
"break",
"yield",
"s",
"[",
":",
"idx",... | For a string "a.b.c", yields "a", "a.b", "a.b.c". | [
"For",
"a",
"string",
"a",
".",
"b",
".",
"c",
"yields",
"a",
"a",
".",
"b",
"a",
".",
"b",
".",
"c",
"."
] | 59f893152d6eb4b7f1f62fc4b35aeeca7f26c07a | https://github.com/closeio/tasktiger/blob/59f893152d6eb4b7f1f62fc4b35aeeca7f26c07a/tasktiger/_internal.py#L81-L91 |
236,909 | closeio/tasktiger | tasktiger/_internal.py | reversed_dotted_parts | def reversed_dotted_parts(s):
"""
For a string "a.b.c", yields "a.b.c", "a.b", "a".
"""
idx = -1
if s:
yield s
while s:
idx = s.rfind('.', 0, idx)
if idx == -1:
break
yield s[:idx] | python | def reversed_dotted_parts(s):
"""
For a string "a.b.c", yields "a.b.c", "a.b", "a".
"""
idx = -1
if s:
yield s
while s:
idx = s.rfind('.', 0, idx)
if idx == -1:
break
yield s[:idx] | [
"def",
"reversed_dotted_parts",
"(",
"s",
")",
":",
"idx",
"=",
"-",
"1",
"if",
"s",
":",
"yield",
"s",
"while",
"s",
":",
"idx",
"=",
"s",
".",
"rfind",
"(",
"'.'",
",",
"0",
",",
"idx",
")",
"if",
"idx",
"==",
"-",
"1",
":",
"break",
"yield... | For a string "a.b.c", yields "a.b.c", "a.b", "a". | [
"For",
"a",
"string",
"a",
".",
"b",
".",
"c",
"yields",
"a",
".",
"b",
".",
"c",
"a",
".",
"b",
"a",
"."
] | 59f893152d6eb4b7f1f62fc4b35aeeca7f26c07a | https://github.com/closeio/tasktiger/blob/59f893152d6eb4b7f1f62fc4b35aeeca7f26c07a/tasktiger/_internal.py#L93-L104 |
236,910 | closeio/tasktiger | tasktiger/logging.py | tasktiger_processor | def tasktiger_processor(logger, method_name, event_dict):
"""
TaskTiger structlog processor.
Inject the current task id for non-batch tasks.
"""
if g['current_tasks'] is not None and not g['current_task_is_batch']:
event_dict['task_id'] = g['current_tasks'][0].id
return event_dict | python | def tasktiger_processor(logger, method_name, event_dict):
"""
TaskTiger structlog processor.
Inject the current task id for non-batch tasks.
"""
if g['current_tasks'] is not None and not g['current_task_is_batch']:
event_dict['task_id'] = g['current_tasks'][0].id
return event_dict | [
"def",
"tasktiger_processor",
"(",
"logger",
",",
"method_name",
",",
"event_dict",
")",
":",
"if",
"g",
"[",
"'current_tasks'",
"]",
"is",
"not",
"None",
"and",
"not",
"g",
"[",
"'current_task_is_batch'",
"]",
":",
"event_dict",
"[",
"'task_id'",
"]",
"=",
... | TaskTiger structlog processor.
Inject the current task id for non-batch tasks. | [
"TaskTiger",
"structlog",
"processor",
"."
] | 59f893152d6eb4b7f1f62fc4b35aeeca7f26c07a | https://github.com/closeio/tasktiger/blob/59f893152d6eb4b7f1f62fc4b35aeeca7f26c07a/tasktiger/logging.py#L4-L14 |
236,911 | closeio/tasktiger | tasktiger/task.py | Task.should_retry_on | def should_retry_on(self, exception_class, logger=None):
"""
Whether this task should be retried when the given exception occurs.
"""
for n in (self.retry_on or []):
try:
if issubclass(exception_class, import_attribute(n)):
return True
... | python | def should_retry_on(self, exception_class, logger=None):
"""
Whether this task should be retried when the given exception occurs.
"""
for n in (self.retry_on or []):
try:
if issubclass(exception_class, import_attribute(n)):
return True
... | [
"def",
"should_retry_on",
"(",
"self",
",",
"exception_class",
",",
"logger",
"=",
"None",
")",
":",
"for",
"n",
"in",
"(",
"self",
".",
"retry_on",
"or",
"[",
"]",
")",
":",
"try",
":",
"if",
"issubclass",
"(",
"exception_class",
",",
"import_attribute"... | Whether this task should be retried when the given exception occurs. | [
"Whether",
"this",
"task",
"should",
"be",
"retried",
"when",
"the",
"given",
"exception",
"occurs",
"."
] | 59f893152d6eb4b7f1f62fc4b35aeeca7f26c07a | https://github.com/closeio/tasktiger/blob/59f893152d6eb4b7f1f62fc4b35aeeca7f26c07a/tasktiger/task.py#L155-L167 |
236,912 | closeio/tasktiger | tasktiger/task.py | Task.update_scheduled_time | def update_scheduled_time(self, when):
"""
Updates a scheduled task's date to the given date. If the task is not
scheduled, a TaskNotFound exception is raised.
"""
tiger = self.tiger
ts = get_timestamp(when)
assert ts
pipeline = tiger.connection.pipeline... | python | def update_scheduled_time(self, when):
"""
Updates a scheduled task's date to the given date. If the task is not
scheduled, a TaskNotFound exception is raised.
"""
tiger = self.tiger
ts = get_timestamp(when)
assert ts
pipeline = tiger.connection.pipeline... | [
"def",
"update_scheduled_time",
"(",
"self",
",",
"when",
")",
":",
"tiger",
"=",
"self",
".",
"tiger",
"ts",
"=",
"get_timestamp",
"(",
"when",
")",
"assert",
"ts",
"pipeline",
"=",
"tiger",
".",
"connection",
".",
"pipeline",
"(",
")",
"key",
"=",
"t... | Updates a scheduled task's date to the given date. If the task is not
scheduled, a TaskNotFound exception is raised. | [
"Updates",
"a",
"scheduled",
"task",
"s",
"date",
"to",
"the",
"given",
"date",
".",
"If",
"the",
"task",
"is",
"not",
"scheduled",
"a",
"TaskNotFound",
"exception",
"is",
"raised",
"."
] | 59f893152d6eb4b7f1f62fc4b35aeeca7f26c07a | https://github.com/closeio/tasktiger/blob/59f893152d6eb4b7f1f62fc4b35aeeca7f26c07a/tasktiger/task.py#L322-L342 |
236,913 | closeio/tasktiger | tasktiger/task.py | Task.n_executions | def n_executions(self):
"""
Queries and returns the number of past task executions.
"""
pipeline = self.tiger.connection.pipeline()
pipeline.exists(self.tiger._key('task', self.id))
pipeline.llen(self.tiger._key('task', self.id, 'executions'))
exists, n_executions... | python | def n_executions(self):
"""
Queries and returns the number of past task executions.
"""
pipeline = self.tiger.connection.pipeline()
pipeline.exists(self.tiger._key('task', self.id))
pipeline.llen(self.tiger._key('task', self.id, 'executions'))
exists, n_executions... | [
"def",
"n_executions",
"(",
"self",
")",
":",
"pipeline",
"=",
"self",
".",
"tiger",
".",
"connection",
".",
"pipeline",
"(",
")",
"pipeline",
".",
"exists",
"(",
"self",
".",
"tiger",
".",
"_key",
"(",
"'task'",
",",
"self",
".",
"id",
")",
")",
"... | Queries and returns the number of past task executions. | [
"Queries",
"and",
"returns",
"the",
"number",
"of",
"past",
"task",
"executions",
"."
] | 59f893152d6eb4b7f1f62fc4b35aeeca7f26c07a | https://github.com/closeio/tasktiger/blob/59f893152d6eb4b7f1f62fc4b35aeeca7f26c07a/tasktiger/task.py#L421-L433 |
236,914 | aewallin/allantools | allantools/noise_kasdin.py | Noise.set_input | def set_input(self, nr=2, qd=1, b=0):
""" Set inputs after initialization
Parameters
-------
nr: integer
length of generated time-series
number must be power of two
qd: float
discrete variance
b: float
noise type:
... | python | def set_input(self, nr=2, qd=1, b=0):
""" Set inputs after initialization
Parameters
-------
nr: integer
length of generated time-series
number must be power of two
qd: float
discrete variance
b: float
noise type:
... | [
"def",
"set_input",
"(",
"self",
",",
"nr",
"=",
"2",
",",
"qd",
"=",
"1",
",",
"b",
"=",
"0",
")",
":",
"self",
".",
"nr",
"=",
"nr",
"self",
".",
"qd",
"=",
"qd",
"self",
".",
"b",
"=",
"b"
] | Set inputs after initialization
Parameters
-------
nr: integer
length of generated time-series
number must be power of two
qd: float
discrete variance
b: float
noise type:
0 : White Phase Modulation (WPM)
... | [
"Set",
"inputs",
"after",
"initialization"
] | b5c695a5af4379fcea4d4ce93a066cb902e7ee0a | https://github.com/aewallin/allantools/blob/b5c695a5af4379fcea4d4ce93a066cb902e7ee0a/allantools/noise_kasdin.py#L81-L102 |
236,915 | aewallin/allantools | allantools/noise_kasdin.py | Noise.generateNoise | def generateNoise(self):
""" Generate noise time series based on input parameters
Returns
-------
time_series: np.array
Time series with colored noise.
len(time_series) == nr
"""
# Fill wfb array with white noise based on given discrete variance
... | python | def generateNoise(self):
""" Generate noise time series based on input parameters
Returns
-------
time_series: np.array
Time series with colored noise.
len(time_series) == nr
"""
# Fill wfb array with white noise based on given discrete variance
... | [
"def",
"generateNoise",
"(",
"self",
")",
":",
"# Fill wfb array with white noise based on given discrete variance",
"wfb",
"=",
"np",
".",
"zeros",
"(",
"self",
".",
"nr",
"*",
"2",
")",
"wfb",
"[",
":",
"self",
".",
"nr",
"]",
"=",
"np",
".",
"random",
"... | Generate noise time series based on input parameters
Returns
-------
time_series: np.array
Time series with colored noise.
len(time_series) == nr | [
"Generate",
"noise",
"time",
"series",
"based",
"on",
"input",
"parameters"
] | b5c695a5af4379fcea4d4ce93a066cb902e7ee0a | https://github.com/aewallin/allantools/blob/b5c695a5af4379fcea4d4ce93a066cb902e7ee0a/allantools/noise_kasdin.py#L104-L130 |
236,916 | aewallin/allantools | allantools/noise_kasdin.py | Noise.adev | def adev(self, tau0, tau):
""" return predicted ADEV of noise-type at given tau
"""
prefactor = self.adev_from_qd(tau0=tau0, tau=tau)
c = self.c_avar()
avar = pow(prefactor, 2)*pow(tau, c)
return np.sqrt(avar) | python | def adev(self, tau0, tau):
""" return predicted ADEV of noise-type at given tau
"""
prefactor = self.adev_from_qd(tau0=tau0, tau=tau)
c = self.c_avar()
avar = pow(prefactor, 2)*pow(tau, c)
return np.sqrt(avar) | [
"def",
"adev",
"(",
"self",
",",
"tau0",
",",
"tau",
")",
":",
"prefactor",
"=",
"self",
".",
"adev_from_qd",
"(",
"tau0",
"=",
"tau0",
",",
"tau",
"=",
"tau",
")",
"c",
"=",
"self",
".",
"c_avar",
"(",
")",
"avar",
"=",
"pow",
"(",
"prefactor",
... | return predicted ADEV of noise-type at given tau | [
"return",
"predicted",
"ADEV",
"of",
"noise",
"-",
"type",
"at",
"given",
"tau"
] | b5c695a5af4379fcea4d4ce93a066cb902e7ee0a | https://github.com/aewallin/allantools/blob/b5c695a5af4379fcea4d4ce93a066cb902e7ee0a/allantools/noise_kasdin.py#L161-L168 |
236,917 | aewallin/allantools | allantools/noise_kasdin.py | Noise.mdev | def mdev(self, tau0, tau):
""" return predicted MDEV of noise-type at given tau
"""
prefactor = self.mdev_from_qd(tau0=tau0, tau=tau)
c = self.c_mvar()
mvar = pow(prefactor, 2)*pow(tau, c)
return np.sqrt(mvar) | python | def mdev(self, tau0, tau):
""" return predicted MDEV of noise-type at given tau
"""
prefactor = self.mdev_from_qd(tau0=tau0, tau=tau)
c = self.c_mvar()
mvar = pow(prefactor, 2)*pow(tau, c)
return np.sqrt(mvar) | [
"def",
"mdev",
"(",
"self",
",",
"tau0",
",",
"tau",
")",
":",
"prefactor",
"=",
"self",
".",
"mdev_from_qd",
"(",
"tau0",
"=",
"tau0",
",",
"tau",
"=",
"tau",
")",
"c",
"=",
"self",
".",
"c_mvar",
"(",
")",
"mvar",
"=",
"pow",
"(",
"prefactor",
... | return predicted MDEV of noise-type at given tau | [
"return",
"predicted",
"MDEV",
"of",
"noise",
"-",
"type",
"at",
"given",
"tau"
] | b5c695a5af4379fcea4d4ce93a066cb902e7ee0a | https://github.com/aewallin/allantools/blob/b5c695a5af4379fcea4d4ce93a066cb902e7ee0a/allantools/noise_kasdin.py#L170-L177 |
236,918 | aewallin/allantools | allantools/noise.py | scipy_psd | def scipy_psd(x, f_sample=1.0, nr_segments=4):
""" PSD routine from scipy
we can compare our own numpy result against this one
"""
f_axis, psd_of_x = scipy.signal.welch(x, f_sample, nperseg=len(x)/nr_segments)
return f_axis, psd_of_x | python | def scipy_psd(x, f_sample=1.0, nr_segments=4):
""" PSD routine from scipy
we can compare our own numpy result against this one
"""
f_axis, psd_of_x = scipy.signal.welch(x, f_sample, nperseg=len(x)/nr_segments)
return f_axis, psd_of_x | [
"def",
"scipy_psd",
"(",
"x",
",",
"f_sample",
"=",
"1.0",
",",
"nr_segments",
"=",
"4",
")",
":",
"f_axis",
",",
"psd_of_x",
"=",
"scipy",
".",
"signal",
".",
"welch",
"(",
"x",
",",
"f_sample",
",",
"nperseg",
"=",
"len",
"(",
"x",
")",
"/",
"n... | PSD routine from scipy
we can compare our own numpy result against this one | [
"PSD",
"routine",
"from",
"scipy",
"we",
"can",
"compare",
"our",
"own",
"numpy",
"result",
"against",
"this",
"one"
] | b5c695a5af4379fcea4d4ce93a066cb902e7ee0a | https://github.com/aewallin/allantools/blob/b5c695a5af4379fcea4d4ce93a066cb902e7ee0a/allantools/noise.py#L37-L42 |
236,919 | aewallin/allantools | allantools/noise.py | iterpink | def iterpink(depth=20):
"""Generate a sequence of samples of pink noise.
pink noise generator
from http://pydoc.net/Python/lmj.sound/0.1.1/lmj.sound.noise/
Based on the Voss-McCartney algorithm, discussion and code examples at
http://www.firstpr.com.au/dsp/pink-noise/
depth: Use this many sam... | python | def iterpink(depth=20):
"""Generate a sequence of samples of pink noise.
pink noise generator
from http://pydoc.net/Python/lmj.sound/0.1.1/lmj.sound.noise/
Based on the Voss-McCartney algorithm, discussion and code examples at
http://www.firstpr.com.au/dsp/pink-noise/
depth: Use this many sam... | [
"def",
"iterpink",
"(",
"depth",
"=",
"20",
")",
":",
"values",
"=",
"numpy",
".",
"random",
".",
"randn",
"(",
"depth",
")",
"smooth",
"=",
"numpy",
".",
"random",
".",
"randn",
"(",
"depth",
")",
"source",
"=",
"numpy",
".",
"random",
".",
"randn... | Generate a sequence of samples of pink noise.
pink noise generator
from http://pydoc.net/Python/lmj.sound/0.1.1/lmj.sound.noise/
Based on the Voss-McCartney algorithm, discussion and code examples at
http://www.firstpr.com.au/dsp/pink-noise/
depth: Use this many samples of white noise to calculat... | [
"Generate",
"a",
"sequence",
"of",
"samples",
"of",
"pink",
"noise",
"."
] | b5c695a5af4379fcea4d4ce93a066cb902e7ee0a | https://github.com/aewallin/allantools/blob/b5c695a5af4379fcea4d4ce93a066cb902e7ee0a/allantools/noise.py#L85-L125 |
236,920 | aewallin/allantools | examples/noise-color-demo.py | plotline | def plotline(plt, alpha, taus, style,label=""):
""" plot a line with the slope alpha """
y = [pow(tt, alpha) for tt in taus]
plt.loglog(taus, y, style,label=label) | python | def plotline(plt, alpha, taus, style,label=""):
""" plot a line with the slope alpha """
y = [pow(tt, alpha) for tt in taus]
plt.loglog(taus, y, style,label=label) | [
"def",
"plotline",
"(",
"plt",
",",
"alpha",
",",
"taus",
",",
"style",
",",
"label",
"=",
"\"\"",
")",
":",
"y",
"=",
"[",
"pow",
"(",
"tt",
",",
"alpha",
")",
"for",
"tt",
"in",
"taus",
"]",
"plt",
".",
"loglog",
"(",
"taus",
",",
"y",
",",... | plot a line with the slope alpha | [
"plot",
"a",
"line",
"with",
"the",
"slope",
"alpha"
] | b5c695a5af4379fcea4d4ce93a066cb902e7ee0a | https://github.com/aewallin/allantools/blob/b5c695a5af4379fcea4d4ce93a066cb902e7ee0a/examples/noise-color-demo.py#L38-L41 |
236,921 | aewallin/allantools | examples/b1_noise_id_figure.py | b1_noise_id | def b1_noise_id(x, af, rate):
""" B1 ratio for noise identification
ratio of Standard Variace to AVAR
"""
(taus,devs,errs,ns) = at.adev(x,taus=[af*rate],data_type="phase", rate=rate)
oadev_x = devs[0]
y = np.diff(x)
y_cut = np.array( y[:len(y)-(len(y)%af)] ) # cut to length
ass... | python | def b1_noise_id(x, af, rate):
""" B1 ratio for noise identification
ratio of Standard Variace to AVAR
"""
(taus,devs,errs,ns) = at.adev(x,taus=[af*rate],data_type="phase", rate=rate)
oadev_x = devs[0]
y = np.diff(x)
y_cut = np.array( y[:len(y)-(len(y)%af)] ) # cut to length
ass... | [
"def",
"b1_noise_id",
"(",
"x",
",",
"af",
",",
"rate",
")",
":",
"(",
"taus",
",",
"devs",
",",
"errs",
",",
"ns",
")",
"=",
"at",
".",
"adev",
"(",
"x",
",",
"taus",
"=",
"[",
"af",
"*",
"rate",
"]",
",",
"data_type",
"=",
"\"phase\"",
",",... | B1 ratio for noise identification
ratio of Standard Variace to AVAR | [
"B1",
"ratio",
"for",
"noise",
"identification",
"ratio",
"of",
"Standard",
"Variace",
"to",
"AVAR"
] | b5c695a5af4379fcea4d4ce93a066cb902e7ee0a | https://github.com/aewallin/allantools/blob/b5c695a5af4379fcea4d4ce93a066cb902e7ee0a/examples/b1_noise_id_figure.py#L5-L19 |
236,922 | aewallin/allantools | allantools/plot.py | Plot.plot | def plot(self, atDataset,
errorbars=False,
grid=False):
""" use matplotlib methods for plotting
Parameters
----------
atDataset : allantools.Dataset()
a dataset with computed data
errorbars : boolean
Plot errorbars. Defaults to F... | python | def plot(self, atDataset,
errorbars=False,
grid=False):
""" use matplotlib methods for plotting
Parameters
----------
atDataset : allantools.Dataset()
a dataset with computed data
errorbars : boolean
Plot errorbars. Defaults to F... | [
"def",
"plot",
"(",
"self",
",",
"atDataset",
",",
"errorbars",
"=",
"False",
",",
"grid",
"=",
"False",
")",
":",
"if",
"errorbars",
":",
"self",
".",
"ax",
".",
"errorbar",
"(",
"atDataset",
".",
"out",
"[",
"\"taus\"",
"]",
",",
"atDataset",
".",
... | use matplotlib methods for plotting
Parameters
----------
atDataset : allantools.Dataset()
a dataset with computed data
errorbars : boolean
Plot errorbars. Defaults to False
grid : boolean
Plot grid. Defaults to False | [
"use",
"matplotlib",
"methods",
"for",
"plotting"
] | b5c695a5af4379fcea4d4ce93a066cb902e7ee0a | https://github.com/aewallin/allantools/blob/b5c695a5af4379fcea4d4ce93a066cb902e7ee0a/allantools/plot.py#L66-L92 |
236,923 | aewallin/allantools | allantools/ci.py | greenhall_table2 | def greenhall_table2(alpha, d):
""" Table 2 from Greenhall 2004 """
row_idx = int(-alpha+2) # map 2-> row0 and -4-> row6
assert(row_idx in [0, 1, 2, 3, 4, 5])
col_idx = int(d-1)
table2 = [[(3.0/2.0, 1.0/2.0), (35.0/18.0, 1.0), (231.0/100.0, 3.0/2.0)], # alpha=+2
[(78.6, 25.2), (790.0, ... | python | def greenhall_table2(alpha, d):
""" Table 2 from Greenhall 2004 """
row_idx = int(-alpha+2) # map 2-> row0 and -4-> row6
assert(row_idx in [0, 1, 2, 3, 4, 5])
col_idx = int(d-1)
table2 = [[(3.0/2.0, 1.0/2.0), (35.0/18.0, 1.0), (231.0/100.0, 3.0/2.0)], # alpha=+2
[(78.6, 25.2), (790.0, ... | [
"def",
"greenhall_table2",
"(",
"alpha",
",",
"d",
")",
":",
"row_idx",
"=",
"int",
"(",
"-",
"alpha",
"+",
"2",
")",
"# map 2-> row0 and -4-> row6",
"assert",
"(",
"row_idx",
"in",
"[",
"0",
",",
"1",
",",
"2",
",",
"3",
",",
"4",
",",
"5",
"]",
... | Table 2 from Greenhall 2004 | [
"Table",
"2",
"from",
"Greenhall",
"2004"
] | b5c695a5af4379fcea4d4ce93a066cb902e7ee0a | https://github.com/aewallin/allantools/blob/b5c695a5af4379fcea4d4ce93a066cb902e7ee0a/allantools/ci.py#L662-L676 |
236,924 | aewallin/allantools | allantools/ci.py | greenhall_table1 | def greenhall_table1(alpha, d):
""" Table 1 from Greenhall 2004 """
row_idx = int(-alpha+2) # map 2-> row0 and -4-> row6
col_idx = int(d-1)
table1 = [[(2.0/3.0, 1.0/3.0), (7.0/9.0, 1.0/2.0), (22.0/25.0, 2.0/3.0)], # alpha=+2
[(0.840, 0.345), (0.997, 0.616), (1.141, 0.843)],
[... | python | def greenhall_table1(alpha, d):
""" Table 1 from Greenhall 2004 """
row_idx = int(-alpha+2) # map 2-> row0 and -4-> row6
col_idx = int(d-1)
table1 = [[(2.0/3.0, 1.0/3.0), (7.0/9.0, 1.0/2.0), (22.0/25.0, 2.0/3.0)], # alpha=+2
[(0.840, 0.345), (0.997, 0.616), (1.141, 0.843)],
[... | [
"def",
"greenhall_table1",
"(",
"alpha",
",",
"d",
")",
":",
"row_idx",
"=",
"int",
"(",
"-",
"alpha",
"+",
"2",
")",
"# map 2-> row0 and -4-> row6",
"col_idx",
"=",
"int",
"(",
"d",
"-",
"1",
")",
"table1",
"=",
"[",
"[",
"(",
"2.0",
"/",
"3.0",
"... | Table 1 from Greenhall 2004 | [
"Table",
"1",
"from",
"Greenhall",
"2004"
] | b5c695a5af4379fcea4d4ce93a066cb902e7ee0a | https://github.com/aewallin/allantools/blob/b5c695a5af4379fcea4d4ce93a066cb902e7ee0a/allantools/ci.py#L678-L691 |
236,925 | aewallin/allantools | allantools/ci.py | edf_mtotdev | def edf_mtotdev(N, m, alpha):
""" Equivalent degrees of freedom for Modified Total Deviation
NIST SP1065 page 41, Table 8
"""
assert(alpha in [2, 1, 0, -1, -2])
NIST_SP1065_table8 = [(1.90, 2.1), (1.20, 1.40), (1.10, 1.2), (0.85, 0.50), (0.75, 0.31)]
#(b, c) = NIST_SP1065_table8[ abs(al... | python | def edf_mtotdev(N, m, alpha):
""" Equivalent degrees of freedom for Modified Total Deviation
NIST SP1065 page 41, Table 8
"""
assert(alpha in [2, 1, 0, -1, -2])
NIST_SP1065_table8 = [(1.90, 2.1), (1.20, 1.40), (1.10, 1.2), (0.85, 0.50), (0.75, 0.31)]
#(b, c) = NIST_SP1065_table8[ abs(al... | [
"def",
"edf_mtotdev",
"(",
"N",
",",
"m",
",",
"alpha",
")",
":",
"assert",
"(",
"alpha",
"in",
"[",
"2",
",",
"1",
",",
"0",
",",
"-",
"1",
",",
"-",
"2",
"]",
")",
"NIST_SP1065_table8",
"=",
"[",
"(",
"1.90",
",",
"2.1",
")",
",",
"(",
"1... | Equivalent degrees of freedom for Modified Total Deviation
NIST SP1065 page 41, Table 8 | [
"Equivalent",
"degrees",
"of",
"freedom",
"for",
"Modified",
"Total",
"Deviation",
"NIST",
"SP1065",
"page",
"41",
"Table",
"8"
] | b5c695a5af4379fcea4d4ce93a066cb902e7ee0a | https://github.com/aewallin/allantools/blob/b5c695a5af4379fcea4d4ce93a066cb902e7ee0a/allantools/ci.py#L710-L721 |
236,926 | aewallin/allantools | allantools/ci.py | edf_simple | def edf_simple(N, m, alpha):
"""Equivalent degrees of freedom.
Simple approximate formulae.
Parameters
----------
N : int
the number of phase samples
m : int
averaging factor, tau = m * tau0
alpha: int
exponent of f for the frequency PSD:
'wp' returns white p... | python | def edf_simple(N, m, alpha):
"""Equivalent degrees of freedom.
Simple approximate formulae.
Parameters
----------
N : int
the number of phase samples
m : int
averaging factor, tau = m * tau0
alpha: int
exponent of f for the frequency PSD:
'wp' returns white p... | [
"def",
"edf_simple",
"(",
"N",
",",
"m",
",",
"alpha",
")",
":",
"N",
"=",
"float",
"(",
"N",
")",
"m",
"=",
"float",
"(",
"m",
")",
"if",
"alpha",
"in",
"[",
"2",
",",
"1",
",",
"0",
",",
"-",
"1",
",",
"-",
"2",
"]",
":",
"# NIST SP 106... | Equivalent degrees of freedom.
Simple approximate formulae.
Parameters
----------
N : int
the number of phase samples
m : int
averaging factor, tau = m * tau0
alpha: int
exponent of f for the frequency PSD:
'wp' returns white phase noise. alpha=+2
... | [
"Equivalent",
"degrees",
"of",
"freedom",
".",
"Simple",
"approximate",
"formulae",
"."
] | b5c695a5af4379fcea4d4ce93a066cb902e7ee0a | https://github.com/aewallin/allantools/blob/b5c695a5af4379fcea4d4ce93a066cb902e7ee0a/allantools/ci.py#L723-L789 |
236,927 | aewallin/allantools | examples/gradev-demo.py | example1 | def example1():
"""
Compute the GRADEV of a white phase noise. Compares two different
scenarios. 1) The original data and 2) ADEV estimate with gap robust ADEV.
"""
N = 1000
f = 1
y = np.random.randn(1,N)[0,:]
x = [xx for xx in np.linspace(1,len(y),len(y))]
x_ax, y_ax, (err_l, err_h... | python | def example1():
"""
Compute the GRADEV of a white phase noise. Compares two different
scenarios. 1) The original data and 2) ADEV estimate with gap robust ADEV.
"""
N = 1000
f = 1
y = np.random.randn(1,N)[0,:]
x = [xx for xx in np.linspace(1,len(y),len(y))]
x_ax, y_ax, (err_l, err_h... | [
"def",
"example1",
"(",
")",
":",
"N",
"=",
"1000",
"f",
"=",
"1",
"y",
"=",
"np",
".",
"random",
".",
"randn",
"(",
"1",
",",
"N",
")",
"[",
"0",
",",
":",
"]",
"x",
"=",
"[",
"xx",
"for",
"xx",
"in",
"np",
".",
"linspace",
"(",
"1",
"... | Compute the GRADEV of a white phase noise. Compares two different
scenarios. 1) The original data and 2) ADEV estimate with gap robust ADEV. | [
"Compute",
"the",
"GRADEV",
"of",
"a",
"white",
"phase",
"noise",
".",
"Compares",
"two",
"different",
"scenarios",
".",
"1",
")",
"The",
"original",
"data",
"and",
"2",
")",
"ADEV",
"estimate",
"with",
"gap",
"robust",
"ADEV",
"."
] | b5c695a5af4379fcea4d4ce93a066cb902e7ee0a | https://github.com/aewallin/allantools/blob/b5c695a5af4379fcea4d4ce93a066cb902e7ee0a/examples/gradev-demo.py#L10-L32 |
236,928 | aewallin/allantools | examples/gradev-demo.py | example2 | def example2():
"""
Compute the GRADEV of a nonstationary white phase noise.
"""
N=1000 # number of samples
f = 1 # data samples per second
s=1+5/N*np.arange(0,N)
y=s*np.random.randn(1,N)[0,:]
x = [xx for xx in np.linspace(1,len(y),len(y))]
x_ax, y_ax, (err_l, err_h) , ns = allan.gra... | python | def example2():
"""
Compute the GRADEV of a nonstationary white phase noise.
"""
N=1000 # number of samples
f = 1 # data samples per second
s=1+5/N*np.arange(0,N)
y=s*np.random.randn(1,N)[0,:]
x = [xx for xx in np.linspace(1,len(y),len(y))]
x_ax, y_ax, (err_l, err_h) , ns = allan.gra... | [
"def",
"example2",
"(",
")",
":",
"N",
"=",
"1000",
"# number of samples",
"f",
"=",
"1",
"# data samples per second",
"s",
"=",
"1",
"+",
"5",
"/",
"N",
"*",
"np",
".",
"arange",
"(",
"0",
",",
"N",
")",
"y",
"=",
"s",
"*",
"np",
".",
"random",
... | Compute the GRADEV of a nonstationary white phase noise. | [
"Compute",
"the",
"GRADEV",
"of",
"a",
"nonstationary",
"white",
"phase",
"noise",
"."
] | b5c695a5af4379fcea4d4ce93a066cb902e7ee0a | https://github.com/aewallin/allantools/blob/b5c695a5af4379fcea4d4ce93a066cb902e7ee0a/examples/gradev-demo.py#L34-L52 |
236,929 | aewallin/allantools | allantools/allantools.py | tdev | def tdev(data, rate=1.0, data_type="phase", taus=None):
""" Time deviation.
Based on modified Allan variance.
.. math::
\\sigma^2_{TDEV}( \\tau ) = { \\tau^2 \\over 3 }
\\sigma^2_{MDEV}( \\tau )
Note that TDEV has a unit of seconds.
Parameters
----------
data: np.arra... | python | def tdev(data, rate=1.0, data_type="phase", taus=None):
""" Time deviation.
Based on modified Allan variance.
.. math::
\\sigma^2_{TDEV}( \\tau ) = { \\tau^2 \\over 3 }
\\sigma^2_{MDEV}( \\tau )
Note that TDEV has a unit of seconds.
Parameters
----------
data: np.arra... | [
"def",
"tdev",
"(",
"data",
",",
"rate",
"=",
"1.0",
",",
"data_type",
"=",
"\"phase\"",
",",
"taus",
"=",
"None",
")",
":",
"phase",
"=",
"input_to_phase",
"(",
"data",
",",
"rate",
",",
"data_type",
")",
"(",
"taus",
",",
"md",
",",
"mde",
",",
... | Time deviation.
Based on modified Allan variance.
.. math::
\\sigma^2_{TDEV}( \\tau ) = { \\tau^2 \\over 3 }
\\sigma^2_{MDEV}( \\tau )
Note that TDEV has a unit of seconds.
Parameters
----------
data: np.array
Input data. Provide either phase or frequency (fractio... | [
"Time",
"deviation",
".",
"Based",
"on",
"modified",
"Allan",
"variance",
"."
] | b5c695a5af4379fcea4d4ce93a066cb902e7ee0a | https://github.com/aewallin/allantools/blob/b5c695a5af4379fcea4d4ce93a066cb902e7ee0a/allantools/allantools.py#L109-L155 |
236,930 | aewallin/allantools | allantools/allantools.py | mdev | def mdev(data, rate=1.0, data_type="phase", taus=None):
""" Modified Allan deviation.
Used to distinguish between White and Flicker Phase Modulation.
.. math::
\\sigma^2_{MDEV}(m\\tau_0) = { 1 \\over 2 (m \\tau_0 )^2 (N-3m+1) }
\\sum_{j=1}^{N-3m+1} \\lbrace
\\sum_{i=j}^{j+m-1... | python | def mdev(data, rate=1.0, data_type="phase", taus=None):
""" Modified Allan deviation.
Used to distinguish between White and Flicker Phase Modulation.
.. math::
\\sigma^2_{MDEV}(m\\tau_0) = { 1 \\over 2 (m \\tau_0 )^2 (N-3m+1) }
\\sum_{j=1}^{N-3m+1} \\lbrace
\\sum_{i=j}^{j+m-1... | [
"def",
"mdev",
"(",
"data",
",",
"rate",
"=",
"1.0",
",",
"data_type",
"=",
"\"phase\"",
",",
"taus",
"=",
"None",
")",
":",
"phase",
"=",
"input_to_phase",
"(",
"data",
",",
"rate",
",",
"data_type",
")",
"(",
"phase",
",",
"ms",
",",
"taus_used",
... | Modified Allan deviation.
Used to distinguish between White and Flicker Phase Modulation.
.. math::
\\sigma^2_{MDEV}(m\\tau_0) = { 1 \\over 2 (m \\tau_0 )^2 (N-3m+1) }
\\sum_{j=1}^{N-3m+1} \\lbrace
\\sum_{i=j}^{j+m-1} {x}_{i+2m} - 2x_{i+m} + x_{i} \\rbrace^2
Parameters
--... | [
"Modified",
"Allan",
"deviation",
".",
"Used",
"to",
"distinguish",
"between",
"White",
"and",
"Flicker",
"Phase",
"Modulation",
"."
] | b5c695a5af4379fcea4d4ce93a066cb902e7ee0a | https://github.com/aewallin/allantools/blob/b5c695a5af4379fcea4d4ce93a066cb902e7ee0a/allantools/allantools.py#L157-L245 |
236,931 | aewallin/allantools | allantools/allantools.py | adev | def adev(data, rate=1.0, data_type="phase", taus=None):
""" Allan deviation.
Classic - use only if required - relatively poor confidence.
.. math::
\\sigma^2_{ADEV}(\\tau) = { 1 \\over 2 \\tau^2 }
\\langle ( {x}_{n+2} - 2x_{n+1} + x_{n} )^2 \\rangle
= { 1 \\over 2 (N-2) \\tau^2... | python | def adev(data, rate=1.0, data_type="phase", taus=None):
""" Allan deviation.
Classic - use only if required - relatively poor confidence.
.. math::
\\sigma^2_{ADEV}(\\tau) = { 1 \\over 2 \\tau^2 }
\\langle ( {x}_{n+2} - 2x_{n+1} + x_{n} )^2 \\rangle
= { 1 \\over 2 (N-2) \\tau^2... | [
"def",
"adev",
"(",
"data",
",",
"rate",
"=",
"1.0",
",",
"data_type",
"=",
"\"phase\"",
",",
"taus",
"=",
"None",
")",
":",
"phase",
"=",
"input_to_phase",
"(",
"data",
",",
"rate",
",",
"data_type",
")",
"(",
"phase",
",",
"m",
",",
"taus_used",
... | Allan deviation.
Classic - use only if required - relatively poor confidence.
.. math::
\\sigma^2_{ADEV}(\\tau) = { 1 \\over 2 \\tau^2 }
\\langle ( {x}_{n+2} - 2x_{n+1} + x_{n} )^2 \\rangle
= { 1 \\over 2 (N-2) \\tau^2 }
\\sum_{n=1}^{N-2} ( {x}_{n+2} - 2x_{n+1} + x_{n} )^2
... | [
"Allan",
"deviation",
".",
"Classic",
"-",
"use",
"only",
"if",
"required",
"-",
"relatively",
"poor",
"confidence",
"."
] | b5c695a5af4379fcea4d4ce93a066cb902e7ee0a | https://github.com/aewallin/allantools/blob/b5c695a5af4379fcea4d4ce93a066cb902e7ee0a/allantools/allantools.py#L247-L311 |
236,932 | aewallin/allantools | allantools/allantools.py | ohdev | def ohdev(data, rate=1.0, data_type="phase", taus=None):
""" Overlapping Hadamard deviation.
Better confidence than normal Hadamard.
.. math::
\\sigma^2_{OHDEV}(m\\tau_0) = { 1 \\over 6 (m \\tau_0 )^2 (N-3m) }
\\sum_{i=1}^{N-3m} ( {x}_{i+3m} - 3x_{i+2m} + 3x_{i+m} - x_{i} )^2
wher... | python | def ohdev(data, rate=1.0, data_type="phase", taus=None):
""" Overlapping Hadamard deviation.
Better confidence than normal Hadamard.
.. math::
\\sigma^2_{OHDEV}(m\\tau_0) = { 1 \\over 6 (m \\tau_0 )^2 (N-3m) }
\\sum_{i=1}^{N-3m} ( {x}_{i+3m} - 3x_{i+2m} + 3x_{i+m} - x_{i} )^2
wher... | [
"def",
"ohdev",
"(",
"data",
",",
"rate",
"=",
"1.0",
",",
"data_type",
"=",
"\"phase\"",
",",
"taus",
"=",
"None",
")",
":",
"phase",
"=",
"input_to_phase",
"(",
"data",
",",
"rate",
",",
"data_type",
")",
"(",
"phase",
",",
"m",
",",
"taus_used",
... | Overlapping Hadamard deviation.
Better confidence than normal Hadamard.
.. math::
\\sigma^2_{OHDEV}(m\\tau_0) = { 1 \\over 6 (m \\tau_0 )^2 (N-3m) }
\\sum_{i=1}^{N-3m} ( {x}_{i+3m} - 3x_{i+2m} + 3x_{i+m} - x_{i} )^2
where :math:`x_i` is the time-series of phase observations, spaced
... | [
"Overlapping",
"Hadamard",
"deviation",
".",
"Better",
"confidence",
"than",
"normal",
"Hadamard",
"."
] | b5c695a5af4379fcea4d4ce93a066cb902e7ee0a | https://github.com/aewallin/allantools/blob/b5c695a5af4379fcea4d4ce93a066cb902e7ee0a/allantools/allantools.py#L420-L471 |
236,933 | aewallin/allantools | allantools/allantools.py | calc_hdev_phase | def calc_hdev_phase(phase, rate, mj, stride):
""" main calculation fungtion for HDEV and OHDEV
Parameters
----------
phase: np.array
Phase data in seconds.
rate: float
The sampling rate for phase or frequency, in Hz
mj: int
M index value for stride
stride: int
... | python | def calc_hdev_phase(phase, rate, mj, stride):
""" main calculation fungtion for HDEV and OHDEV
Parameters
----------
phase: np.array
Phase data in seconds.
rate: float
The sampling rate for phase or frequency, in Hz
mj: int
M index value for stride
stride: int
... | [
"def",
"calc_hdev_phase",
"(",
"phase",
",",
"rate",
",",
"mj",
",",
"stride",
")",
":",
"tau0",
"=",
"1.0",
"/",
"float",
"(",
"rate",
")",
"mj",
"=",
"int",
"(",
"mj",
")",
"stride",
"=",
"int",
"(",
"stride",
")",
"d3",
"=",
"phase",
"[",
"3... | main calculation fungtion for HDEV and OHDEV
Parameters
----------
phase: np.array
Phase data in seconds.
rate: float
The sampling rate for phase or frequency, in Hz
mj: int
M index value for stride
stride: int
Size of stride
Returns
-------
(dev, de... | [
"main",
"calculation",
"fungtion",
"for",
"HDEV",
"and",
"OHDEV"
] | b5c695a5af4379fcea4d4ce93a066cb902e7ee0a | https://github.com/aewallin/allantools/blob/b5c695a5af4379fcea4d4ce93a066cb902e7ee0a/allantools/allantools.py#L515-L566 |
236,934 | aewallin/allantools | allantools/allantools.py | totdev | def totdev(data, rate=1.0, data_type="phase", taus=None):
""" Total deviation.
Better confidence at long averages for Allan.
.. math::
\\sigma^2_{TOTDEV}( m\\tau_0 ) = { 1 \\over 2 (m\\tau_0)^2 (N-2) }
\\sum_{i=2}^{N-1} ( {x}^*_{i-m} - 2x^*_{i} + x^*_{i+m} )^2
Where :math:`x^... | python | def totdev(data, rate=1.0, data_type="phase", taus=None):
""" Total deviation.
Better confidence at long averages for Allan.
.. math::
\\sigma^2_{TOTDEV}( m\\tau_0 ) = { 1 \\over 2 (m\\tau_0)^2 (N-2) }
\\sum_{i=2}^{N-1} ( {x}^*_{i-m} - 2x^*_{i} + x^*_{i+m} )^2
Where :math:`x^... | [
"def",
"totdev",
"(",
"data",
",",
"rate",
"=",
"1.0",
",",
"data_type",
"=",
"\"phase\"",
",",
"taus",
"=",
"None",
")",
":",
"phase",
"=",
"input_to_phase",
"(",
"data",
",",
"rate",
",",
"data_type",
")",
"(",
"phase",
",",
"m",
",",
"taus_used",
... | Total deviation.
Better confidence at long averages for Allan.
.. math::
\\sigma^2_{TOTDEV}( m\\tau_0 ) = { 1 \\over 2 (m\\tau_0)^2 (N-2) }
\\sum_{i=2}^{N-1} ( {x}^*_{i-m} - 2x^*_{i} + x^*_{i+m} )^2
Where :math:`x^*_i` is a new time-series of length :math:`3N-4`
derived from ... | [
"Total",
"deviation",
".",
"Better",
"confidence",
"at",
"long",
"averages",
"for",
"Allan",
"."
] | b5c695a5af4379fcea4d4ce93a066cb902e7ee0a | https://github.com/aewallin/allantools/blob/b5c695a5af4379fcea4d4ce93a066cb902e7ee0a/allantools/allantools.py#L568-L660 |
236,935 | aewallin/allantools | allantools/allantools.py | mtotdev | def mtotdev(data, rate=1.0, data_type="phase", taus=None):
""" PRELIMINARY - REQUIRES FURTHER TESTING.
Modified Total deviation.
Better confidence at long averages for modified Allan
FIXME: bias-correction http://www.wriley.com/CI2.pdf page 6
The variance is scaled up (divided by t... | python | def mtotdev(data, rate=1.0, data_type="phase", taus=None):
""" PRELIMINARY - REQUIRES FURTHER TESTING.
Modified Total deviation.
Better confidence at long averages for modified Allan
FIXME: bias-correction http://www.wriley.com/CI2.pdf page 6
The variance is scaled up (divided by t... | [
"def",
"mtotdev",
"(",
"data",
",",
"rate",
"=",
"1.0",
",",
"data_type",
"=",
"\"phase\"",
",",
"taus",
"=",
"None",
")",
":",
"phase",
"=",
"input_to_phase",
"(",
"data",
",",
"rate",
",",
"data_type",
")",
"(",
"phase",
",",
"ms",
",",
"taus_used"... | PRELIMINARY - REQUIRES FURTHER TESTING.
Modified Total deviation.
Better confidence at long averages for modified Allan
FIXME: bias-correction http://www.wriley.com/CI2.pdf page 6
The variance is scaled up (divided by this number) based on the
noise-type identified.
WPM... | [
"PRELIMINARY",
"-",
"REQUIRES",
"FURTHER",
"TESTING",
".",
"Modified",
"Total",
"deviation",
".",
"Better",
"confidence",
"at",
"long",
"averages",
"for",
"modified",
"Allan"
] | b5c695a5af4379fcea4d4ce93a066cb902e7ee0a | https://github.com/aewallin/allantools/blob/b5c695a5af4379fcea4d4ce93a066cb902e7ee0a/allantools/allantools.py#L674-L716 |
236,936 | aewallin/allantools | allantools/allantools.py | htotdev | def htotdev(data, rate=1.0, data_type="phase", taus=None):
""" PRELIMINARY - REQUIRES FURTHER TESTING.
Hadamard Total deviation.
Better confidence at long averages for Hadamard deviation
FIXME: bias corrections from http://www.wriley.com/CI2.pdf
W FM 0.995 alpha= 0
F... | python | def htotdev(data, rate=1.0, data_type="phase", taus=None):
""" PRELIMINARY - REQUIRES FURTHER TESTING.
Hadamard Total deviation.
Better confidence at long averages for Hadamard deviation
FIXME: bias corrections from http://www.wriley.com/CI2.pdf
W FM 0.995 alpha= 0
F... | [
"def",
"htotdev",
"(",
"data",
",",
"rate",
"=",
"1.0",
",",
"data_type",
"=",
"\"phase\"",
",",
"taus",
"=",
"None",
")",
":",
"if",
"data_type",
"==",
"\"phase\"",
":",
"phase",
"=",
"data",
"freq",
"=",
"phase2frequency",
"(",
"phase",
",",
"rate",
... | PRELIMINARY - REQUIRES FURTHER TESTING.
Hadamard Total deviation.
Better confidence at long averages for Hadamard deviation
FIXME: bias corrections from http://www.wriley.com/CI2.pdf
W FM 0.995 alpha= 0
F FM 0.851 alpha=-1
RW FM 0.771 alpha=-2
... | [
"PRELIMINARY",
"-",
"REQUIRES",
"FURTHER",
"TESTING",
".",
"Hadamard",
"Total",
"deviation",
".",
"Better",
"confidence",
"at",
"long",
"averages",
"for",
"Hadamard",
"deviation"
] | b5c695a5af4379fcea4d4ce93a066cb902e7ee0a | https://github.com/aewallin/allantools/blob/b5c695a5af4379fcea4d4ce93a066cb902e7ee0a/allantools/allantools.py#L787-L847 |
236,937 | aewallin/allantools | allantools/allantools.py | theo1 | def theo1(data, rate=1.0, data_type="phase", taus=None):
""" PRELIMINARY - REQUIRES FURTHER TESTING.
Theo1 is a two-sample variance with improved confidence and
extended averaging factor range.
.. math::
\\sigma^2_{THEO1}(m\\tau_0) = { 1 \\over (m \\tau_0 )^2 (N-m) }
... | python | def theo1(data, rate=1.0, data_type="phase", taus=None):
""" PRELIMINARY - REQUIRES FURTHER TESTING.
Theo1 is a two-sample variance with improved confidence and
extended averaging factor range.
.. math::
\\sigma^2_{THEO1}(m\\tau_0) = { 1 \\over (m \\tau_0 )^2 (N-m) }
... | [
"def",
"theo1",
"(",
"data",
",",
"rate",
"=",
"1.0",
",",
"data_type",
"=",
"\"phase\"",
",",
"taus",
"=",
"None",
")",
":",
"phase",
"=",
"input_to_phase",
"(",
"data",
",",
"rate",
",",
"data_type",
")",
"tau0",
"=",
"1.0",
"/",
"rate",
"(",
"ph... | PRELIMINARY - REQUIRES FURTHER TESTING.
Theo1 is a two-sample variance with improved confidence and
extended averaging factor range.
.. math::
\\sigma^2_{THEO1}(m\\tau_0) = { 1 \\over (m \\tau_0 )^2 (N-m) }
\\sum_{i=1}^{N-m} \\sum_{\\delta=0}^{m/2-1}
... | [
"PRELIMINARY",
"-",
"REQUIRES",
"FURTHER",
"TESTING",
".",
"Theo1",
"is",
"a",
"two",
"-",
"sample",
"variance",
"with",
"improved",
"confidence",
"and",
"extended",
"averaging",
"factor",
"range",
"."
] | b5c695a5af4379fcea4d4ce93a066cb902e7ee0a | https://github.com/aewallin/allantools/blob/b5c695a5af4379fcea4d4ce93a066cb902e7ee0a/allantools/allantools.py#L919-L987 |
236,938 | aewallin/allantools | allantools/allantools.py | tierms | def tierms(data, rate=1.0, data_type="phase", taus=None):
""" Time Interval Error RMS.
Parameters
----------
data: np.array
Input data. Provide either phase or frequency (fractional,
adimensional).
rate: float
The sampling rate for data, in Hz. Defaults to 1.0
data_type:... | python | def tierms(data, rate=1.0, data_type="phase", taus=None):
""" Time Interval Error RMS.
Parameters
----------
data: np.array
Input data. Provide either phase or frequency (fractional,
adimensional).
rate: float
The sampling rate for data, in Hz. Defaults to 1.0
data_type:... | [
"def",
"tierms",
"(",
"data",
",",
"rate",
"=",
"1.0",
",",
"data_type",
"=",
"\"phase\"",
",",
"taus",
"=",
"None",
")",
":",
"phase",
"=",
"input_to_phase",
"(",
"data",
",",
"rate",
",",
"data_type",
")",
"(",
"data",
",",
"m",
",",
"taus_used",
... | Time Interval Error RMS.
Parameters
----------
data: np.array
Input data. Provide either phase or frequency (fractional,
adimensional).
rate: float
The sampling rate for data, in Hz. Defaults to 1.0
data_type: {'phase', 'freq'}
Data type, i.e. phase or frequency. Def... | [
"Time",
"Interval",
"Error",
"RMS",
"."
] | b5c695a5af4379fcea4d4ce93a066cb902e7ee0a | https://github.com/aewallin/allantools/blob/b5c695a5af4379fcea4d4ce93a066cb902e7ee0a/allantools/allantools.py#L990-L1033 |
236,939 | aewallin/allantools | allantools/allantools.py | mtie | def mtie(data, rate=1.0, data_type="phase", taus=None):
""" Maximum Time Interval Error.
Parameters
----------
data: np.array
Input data. Provide either phase or frequency (fractional,
adimensional).
rate: float
The sampling rate for data, in Hz. Defaults to 1.0
data_typ... | python | def mtie(data, rate=1.0, data_type="phase", taus=None):
""" Maximum Time Interval Error.
Parameters
----------
data: np.array
Input data. Provide either phase or frequency (fractional,
adimensional).
rate: float
The sampling rate for data, in Hz. Defaults to 1.0
data_typ... | [
"def",
"mtie",
"(",
"data",
",",
"rate",
"=",
"1.0",
",",
"data_type",
"=",
"\"phase\"",
",",
"taus",
"=",
"None",
")",
":",
"phase",
"=",
"input_to_phase",
"(",
"data",
",",
"rate",
",",
"data_type",
")",
"(",
"phase",
",",
"m",
",",
"taus_used",
... | Maximum Time Interval Error.
Parameters
----------
data: np.array
Input data. Provide either phase or frequency (fractional,
adimensional).
rate: float
The sampling rate for data, in Hz. Defaults to 1.0
data_type: {'phase', 'freq'}
Data type, i.e. phase or frequency.... | [
"Maximum",
"Time",
"Interval",
"Error",
"."
] | b5c695a5af4379fcea4d4ce93a066cb902e7ee0a | https://github.com/aewallin/allantools/blob/b5c695a5af4379fcea4d4ce93a066cb902e7ee0a/allantools/allantools.py#L1061-L1101 |
236,940 | aewallin/allantools | allantools/allantools.py | mtie_phase_fast | def mtie_phase_fast(phase, rate=1.0, data_type="phase", taus=None):
""" fast binary decomposition algorithm for MTIE
See: STEFANO BREGNI "Fast Algorithms for TVAR and MTIE Computation in
Characterization of Network Synchronization Performance"
"""
rate = float(rate)
phase = np.asarray(p... | python | def mtie_phase_fast(phase, rate=1.0, data_type="phase", taus=None):
""" fast binary decomposition algorithm for MTIE
See: STEFANO BREGNI "Fast Algorithms for TVAR and MTIE Computation in
Characterization of Network Synchronization Performance"
"""
rate = float(rate)
phase = np.asarray(p... | [
"def",
"mtie_phase_fast",
"(",
"phase",
",",
"rate",
"=",
"1.0",
",",
"data_type",
"=",
"\"phase\"",
",",
"taus",
"=",
"None",
")",
":",
"rate",
"=",
"float",
"(",
"rate",
")",
"phase",
"=",
"np",
".",
"asarray",
"(",
"phase",
")",
"k_max",
"=",
"i... | fast binary decomposition algorithm for MTIE
See: STEFANO BREGNI "Fast Algorithms for TVAR and MTIE Computation in
Characterization of Network Synchronization Performance" | [
"fast",
"binary",
"decomposition",
"algorithm",
"for",
"MTIE"
] | b5c695a5af4379fcea4d4ce93a066cb902e7ee0a | https://github.com/aewallin/allantools/blob/b5c695a5af4379fcea4d4ce93a066cb902e7ee0a/allantools/allantools.py#L1108-L1163 |
236,941 | aewallin/allantools | allantools/allantools.py | gradev | def gradev(data, rate=1.0, data_type="phase", taus=None,
ci=0.9, noisetype='wp'):
""" gap resistant overlapping Allan deviation
Parameters
----------
data: np.array
Input data. Provide either phase or frequency (fractional,
adimensional). Warning : phase data works better (fr... | python | def gradev(data, rate=1.0, data_type="phase", taus=None,
ci=0.9, noisetype='wp'):
""" gap resistant overlapping Allan deviation
Parameters
----------
data: np.array
Input data. Provide either phase or frequency (fractional,
adimensional). Warning : phase data works better (fr... | [
"def",
"gradev",
"(",
"data",
",",
"rate",
"=",
"1.0",
",",
"data_type",
"=",
"\"phase\"",
",",
"taus",
"=",
"None",
",",
"ci",
"=",
"0.9",
",",
"noisetype",
"=",
"'wp'",
")",
":",
"if",
"(",
"data_type",
"==",
"\"freq\"",
")",
":",
"print",
"(",
... | gap resistant overlapping Allan deviation
Parameters
----------
data: np.array
Input data. Provide either phase or frequency (fractional,
adimensional). Warning : phase data works better (frequency data is
first trantformed into phase using numpy.cumsum() function, which can
... | [
"gap",
"resistant",
"overlapping",
"Allan",
"deviation"
] | b5c695a5af4379fcea4d4ce93a066cb902e7ee0a | https://github.com/aewallin/allantools/blob/b5c695a5af4379fcea4d4ce93a066cb902e7ee0a/allantools/allantools.py#L1173-L1242 |
236,942 | aewallin/allantools | allantools/allantools.py | input_to_phase | def input_to_phase(data, rate, data_type):
""" Take either phase or frequency as input and return phase
"""
if data_type == "phase":
return data
elif data_type == "freq":
return frequency2phase(data, rate)
else:
raise Exception("unknown data_type: " + data_type) | python | def input_to_phase(data, rate, data_type):
""" Take either phase or frequency as input and return phase
"""
if data_type == "phase":
return data
elif data_type == "freq":
return frequency2phase(data, rate)
else:
raise Exception("unknown data_type: " + data_type) | [
"def",
"input_to_phase",
"(",
"data",
",",
"rate",
",",
"data_type",
")",
":",
"if",
"data_type",
"==",
"\"phase\"",
":",
"return",
"data",
"elif",
"data_type",
"==",
"\"freq\"",
":",
"return",
"frequency2phase",
"(",
"data",
",",
"rate",
")",
"else",
":",... | Take either phase or frequency as input and return phase | [
"Take",
"either",
"phase",
"or",
"frequency",
"as",
"input",
"and",
"return",
"phase"
] | b5c695a5af4379fcea4d4ce93a066cb902e7ee0a | https://github.com/aewallin/allantools/blob/b5c695a5af4379fcea4d4ce93a066cb902e7ee0a/allantools/allantools.py#L1297-L1305 |
236,943 | aewallin/allantools | allantools/allantools.py | trim_data | def trim_data(x):
"""
Trim leading and trailing NaNs from dataset
This is done by browsing the array from each end and store the index of the
first non-NaN in each case, the return the appropriate slice of the array
"""
# Find indices for first and last valid data
first = 0
while np.isna... | python | def trim_data(x):
"""
Trim leading and trailing NaNs from dataset
This is done by browsing the array from each end and store the index of the
first non-NaN in each case, the return the appropriate slice of the array
"""
# Find indices for first and last valid data
first = 0
while np.isna... | [
"def",
"trim_data",
"(",
"x",
")",
":",
"# Find indices for first and last valid data",
"first",
"=",
"0",
"while",
"np",
".",
"isnan",
"(",
"x",
"[",
"first",
"]",
")",
":",
"first",
"+=",
"1",
"last",
"=",
"len",
"(",
"x",
")",
"while",
"np",
".",
... | Trim leading and trailing NaNs from dataset
This is done by browsing the array from each end and store the index of the
first non-NaN in each case, the return the appropriate slice of the array | [
"Trim",
"leading",
"and",
"trailing",
"NaNs",
"from",
"dataset",
"This",
"is",
"done",
"by",
"browsing",
"the",
"array",
"from",
"each",
"end",
"and",
"store",
"the",
"index",
"of",
"the",
"first",
"non",
"-",
"NaN",
"in",
"each",
"case",
"the",
"return"... | b5c695a5af4379fcea4d4ce93a066cb902e7ee0a | https://github.com/aewallin/allantools/blob/b5c695a5af4379fcea4d4ce93a066cb902e7ee0a/allantools/allantools.py#L1498-L1511 |
236,944 | aewallin/allantools | allantools/allantools.py | three_cornered_hat_phase | def three_cornered_hat_phase(phasedata_ab, phasedata_bc,
phasedata_ca, rate, taus, function):
"""
Three Cornered Hat Method
Given three clocks A, B, C, we seek to find their variances
:math:`\\sigma^2_A`, :math:`\\sigma^2_B`, :math:`\\sigma^2_C`.
We measure three phase ... | python | def three_cornered_hat_phase(phasedata_ab, phasedata_bc,
phasedata_ca, rate, taus, function):
"""
Three Cornered Hat Method
Given three clocks A, B, C, we seek to find their variances
:math:`\\sigma^2_A`, :math:`\\sigma^2_B`, :math:`\\sigma^2_C`.
We measure three phase ... | [
"def",
"three_cornered_hat_phase",
"(",
"phasedata_ab",
",",
"phasedata_bc",
",",
"phasedata_ca",
",",
"rate",
",",
"taus",
",",
"function",
")",
":",
"(",
"tau_ab",
",",
"dev_ab",
",",
"err_ab",
",",
"ns_ab",
")",
"=",
"function",
"(",
"phasedata_ab",
",",
... | Three Cornered Hat Method
Given three clocks A, B, C, we seek to find their variances
:math:`\\sigma^2_A`, :math:`\\sigma^2_B`, :math:`\\sigma^2_C`.
We measure three phase differences, assuming no correlation between
the clocks, the measurements have variances:
.. math::
\\sigma^2_{AB} = ... | [
"Three",
"Cornered",
"Hat",
"Method"
] | b5c695a5af4379fcea4d4ce93a066cb902e7ee0a | https://github.com/aewallin/allantools/blob/b5c695a5af4379fcea4d4ce93a066cb902e7ee0a/allantools/allantools.py#L1513-L1588 |
236,945 | aewallin/allantools | allantools/allantools.py | frequency2phase | def frequency2phase(freqdata, rate):
""" integrate fractional frequency data and output phase data
Parameters
----------
freqdata: np.array
Data array of fractional frequency measurements (nondimensional)
rate: float
The sampling rate for phase or frequency, in Hz
Returns
-... | python | def frequency2phase(freqdata, rate):
""" integrate fractional frequency data and output phase data
Parameters
----------
freqdata: np.array
Data array of fractional frequency measurements (nondimensional)
rate: float
The sampling rate for phase or frequency, in Hz
Returns
-... | [
"def",
"frequency2phase",
"(",
"freqdata",
",",
"rate",
")",
":",
"dt",
"=",
"1.0",
"/",
"float",
"(",
"rate",
")",
"# Protect against NaN values in input array (issue #60)",
"# Reintroduces data trimming as in commit 503cb82",
"freqdata",
"=",
"trim_data",
"(",
"freqdata... | integrate fractional frequency data and output phase data
Parameters
----------
freqdata: np.array
Data array of fractional frequency measurements (nondimensional)
rate: float
The sampling rate for phase or frequency, in Hz
Returns
-------
phasedata: np.array
Time i... | [
"integrate",
"fractional",
"frequency",
"data",
"and",
"output",
"phase",
"data"
] | b5c695a5af4379fcea4d4ce93a066cb902e7ee0a | https://github.com/aewallin/allantools/blob/b5c695a5af4379fcea4d4ce93a066cb902e7ee0a/allantools/allantools.py#L1595-L1619 |
236,946 | aewallin/allantools | allantools/allantools.py | phase2radians | def phase2radians(phasedata, v0):
""" Convert phase in seconds to phase in radians
Parameters
----------
phasedata: np.array
Data array of phase in seconds
v0: float
Nominal oscillator frequency in Hz
Returns
-------
fi:
phase data in radians
"""
fi = [2... | python | def phase2radians(phasedata, v0):
""" Convert phase in seconds to phase in radians
Parameters
----------
phasedata: np.array
Data array of phase in seconds
v0: float
Nominal oscillator frequency in Hz
Returns
-------
fi:
phase data in radians
"""
fi = [2... | [
"def",
"phase2radians",
"(",
"phasedata",
",",
"v0",
")",
":",
"fi",
"=",
"[",
"2",
"*",
"np",
".",
"pi",
"*",
"v0",
"*",
"xx",
"for",
"xx",
"in",
"phasedata",
"]",
"return",
"fi"
] | Convert phase in seconds to phase in radians
Parameters
----------
phasedata: np.array
Data array of phase in seconds
v0: float
Nominal oscillator frequency in Hz
Returns
-------
fi:
phase data in radians | [
"Convert",
"phase",
"in",
"seconds",
"to",
"phase",
"in",
"radians"
] | b5c695a5af4379fcea4d4ce93a066cb902e7ee0a | https://github.com/aewallin/allantools/blob/b5c695a5af4379fcea4d4ce93a066cb902e7ee0a/allantools/allantools.py#L1621-L1637 |
236,947 | aewallin/allantools | allantools/allantools.py | frequency2fractional | def frequency2fractional(frequency, mean_frequency=-1):
""" Convert frequency in Hz to fractional frequency
Parameters
----------
frequency: np.array
Data array of frequency in Hz
mean_frequency: float
(optional) The nominal mean frequency, in Hz
if omitted, defaults to mean... | python | def frequency2fractional(frequency, mean_frequency=-1):
""" Convert frequency in Hz to fractional frequency
Parameters
----------
frequency: np.array
Data array of frequency in Hz
mean_frequency: float
(optional) The nominal mean frequency, in Hz
if omitted, defaults to mean... | [
"def",
"frequency2fractional",
"(",
"frequency",
",",
"mean_frequency",
"=",
"-",
"1",
")",
":",
"if",
"mean_frequency",
"==",
"-",
"1",
":",
"mu",
"=",
"np",
".",
"mean",
"(",
"frequency",
")",
"else",
":",
"mu",
"=",
"mean_frequency",
"y",
"=",
"[",
... | Convert frequency in Hz to fractional frequency
Parameters
----------
frequency: np.array
Data array of frequency in Hz
mean_frequency: float
(optional) The nominal mean frequency, in Hz
if omitted, defaults to mean frequency=np.mean(frequency)
Returns
-------
y:
... | [
"Convert",
"frequency",
"in",
"Hz",
"to",
"fractional",
"frequency"
] | b5c695a5af4379fcea4d4ce93a066cb902e7ee0a | https://github.com/aewallin/allantools/blob/b5c695a5af4379fcea4d4ce93a066cb902e7ee0a/allantools/allantools.py#L1657-L1678 |
236,948 | aewallin/allantools | allantools/dataset.py | Dataset.set_input | def set_input(self, data,
rate=1.0, data_type="phase", taus=None):
""" Optionnal method if you chose not to set inputs on init
Parameters
----------
data: np.array
Input data. Provide either phase or frequency (fractional,
adimensional)
... | python | def set_input(self, data,
rate=1.0, data_type="phase", taus=None):
""" Optionnal method if you chose not to set inputs on init
Parameters
----------
data: np.array
Input data. Provide either phase or frequency (fractional,
adimensional)
... | [
"def",
"set_input",
"(",
"self",
",",
"data",
",",
"rate",
"=",
"1.0",
",",
"data_type",
"=",
"\"phase\"",
",",
"taus",
"=",
"None",
")",
":",
"self",
".",
"inp",
"[",
"\"data\"",
"]",
"=",
"data",
"self",
".",
"inp",
"[",
"\"rate\"",
"]",
"=",
"... | Optionnal method if you chose not to set inputs on init
Parameters
----------
data: np.array
Input data. Provide either phase or frequency (fractional,
adimensional)
rate: float
The sampling rate for data, in Hz. Defaults to 1.0
data_type: {'p... | [
"Optionnal",
"method",
"if",
"you",
"chose",
"not",
"to",
"set",
"inputs",
"on",
"init"
] | b5c695a5af4379fcea4d4ce93a066cb902e7ee0a | https://github.com/aewallin/allantools/blob/b5c695a5af4379fcea4d4ce93a066cb902e7ee0a/allantools/dataset.py#L93-L113 |
236,949 | aewallin/allantools | allantools/dataset.py | Dataset.compute | def compute(self, function):
"""Evaluate the passed function with the supplied data.
Stores result in self.out.
Parameters
----------
function: str
Name of the :mod:`allantools` function to evaluate
Returns
-------
result: dict
T... | python | def compute(self, function):
"""Evaluate the passed function with the supplied data.
Stores result in self.out.
Parameters
----------
function: str
Name of the :mod:`allantools` function to evaluate
Returns
-------
result: dict
T... | [
"def",
"compute",
"(",
"self",
",",
"function",
")",
":",
"try",
":",
"func",
"=",
"getattr",
"(",
"allantools",
",",
"function",
")",
"except",
"AttributeError",
":",
"raise",
"AttributeError",
"(",
"\"function must be defined in allantools\"",
")",
"whitelisted"... | Evaluate the passed function with the supplied data.
Stores result in self.out.
Parameters
----------
function: str
Name of the :mod:`allantools` function to evaluate
Returns
-------
result: dict
The results of the calculation. | [
"Evaluate",
"the",
"passed",
"function",
"with",
"the",
"supplied",
"data",
"."
] | b5c695a5af4379fcea4d4ce93a066cb902e7ee0a | https://github.com/aewallin/allantools/blob/b5c695a5af4379fcea4d4ce93a066cb902e7ee0a/allantools/dataset.py#L115-L148 |
236,950 | aewallin/allantools | examples/noise-color_and_PSD.py | many_psds | def many_psds(k=2,fs=1.0, b0=1.0, N=1024):
""" compute average of many PSDs """
psd=[]
for j in range(k):
print j
x = noise.white(N=2*4096,b0=b0,fs=fs)
f, tmp = noise.numpy_psd(x,fs)
if j==0:
psd = tmp
else:
psd = psd + tmp
return f, psd/k | python | def many_psds(k=2,fs=1.0, b0=1.0, N=1024):
""" compute average of many PSDs """
psd=[]
for j in range(k):
print j
x = noise.white(N=2*4096,b0=b0,fs=fs)
f, tmp = noise.numpy_psd(x,fs)
if j==0:
psd = tmp
else:
psd = psd + tmp
return f, psd/k | [
"def",
"many_psds",
"(",
"k",
"=",
"2",
",",
"fs",
"=",
"1.0",
",",
"b0",
"=",
"1.0",
",",
"N",
"=",
"1024",
")",
":",
"psd",
"=",
"[",
"]",
"for",
"j",
"in",
"range",
"(",
"k",
")",
":",
"print",
"j",
"x",
"=",
"noise",
".",
"white",
"("... | compute average of many PSDs | [
"compute",
"average",
"of",
"many",
"PSDs"
] | b5c695a5af4379fcea4d4ce93a066cb902e7ee0a | https://github.com/aewallin/allantools/blob/b5c695a5af4379fcea4d4ce93a066cb902e7ee0a/examples/noise-color_and_PSD.py#L7-L18 |
236,951 | singnet/snet-cli | snet_cli/commands.py | OrganizationCommand.list_my | def list_my(self):
""" Find organization that has the current identity as the owner or as the member """
org_list = self.call_contract_command("Registry", "listOrganizations", [])
rez_owner = []
rez_member = []
for idx, org_id in enumerate(org_list):
(found, org_id,... | python | def list_my(self):
""" Find organization that has the current identity as the owner or as the member """
org_list = self.call_contract_command("Registry", "listOrganizations", [])
rez_owner = []
rez_member = []
for idx, org_id in enumerate(org_list):
(found, org_id,... | [
"def",
"list_my",
"(",
"self",
")",
":",
"org_list",
"=",
"self",
".",
"call_contract_command",
"(",
"\"Registry\"",
",",
"\"listOrganizations\"",
",",
"[",
"]",
")",
"rez_owner",
"=",
"[",
"]",
"rez_member",
"=",
"[",
"]",
"for",
"idx",
",",
"org_id",
"... | Find organization that has the current identity as the owner or as the member | [
"Find",
"organization",
"that",
"has",
"the",
"current",
"identity",
"as",
"the",
"owner",
"or",
"as",
"the",
"member"
] | 1b5ac98cb9a64211c861ead9fcfe6208f2749032 | https://github.com/singnet/snet-cli/blob/1b5ac98cb9a64211c861ead9fcfe6208f2749032/snet_cli/commands.py#L541-L567 |
236,952 | singnet/snet-cli | snet_cli/mpe_service_metadata.py | MPEServiceMetadata.add_group | def add_group(self, group_name, payment_address):
""" Return new group_id in base64 """
if (self.is_group_name_exists(group_name)):
raise Exception("the group \"%s\" is already present"%str(group_name))
group_id_base64 = base64.b64encode(secrets.token_bytes(32))
self.m["group... | python | def add_group(self, group_name, payment_address):
""" Return new group_id in base64 """
if (self.is_group_name_exists(group_name)):
raise Exception("the group \"%s\" is already present"%str(group_name))
group_id_base64 = base64.b64encode(secrets.token_bytes(32))
self.m["group... | [
"def",
"add_group",
"(",
"self",
",",
"group_name",
",",
"payment_address",
")",
":",
"if",
"(",
"self",
".",
"is_group_name_exists",
"(",
"group_name",
")",
")",
":",
"raise",
"Exception",
"(",
"\"the group \\\"%s\\\" is already present\"",
"%",
"str",
"(",
"gr... | Return new group_id in base64 | [
"Return",
"new",
"group_id",
"in",
"base64"
] | 1b5ac98cb9a64211c861ead9fcfe6208f2749032 | https://github.com/singnet/snet-cli/blob/1b5ac98cb9a64211c861ead9fcfe6208f2749032/snet_cli/mpe_service_metadata.py#L75-L83 |
236,953 | singnet/snet-cli | snet_cli/mpe_service_metadata.py | MPEServiceMetadata.is_group_name_exists | def is_group_name_exists(self, group_name):
""" check if group with given name is already exists """
groups = self.m["groups"]
for g in groups:
if (g["group_name"] == group_name):
return True
return False | python | def is_group_name_exists(self, group_name):
""" check if group with given name is already exists """
groups = self.m["groups"]
for g in groups:
if (g["group_name"] == group_name):
return True
return False | [
"def",
"is_group_name_exists",
"(",
"self",
",",
"group_name",
")",
":",
"groups",
"=",
"self",
".",
"m",
"[",
"\"groups\"",
"]",
"for",
"g",
"in",
"groups",
":",
"if",
"(",
"g",
"[",
"\"group_name\"",
"]",
"==",
"group_name",
")",
":",
"return",
"True... | check if group with given name is already exists | [
"check",
"if",
"group",
"with",
"given",
"name",
"is",
"already",
"exists"
] | 1b5ac98cb9a64211c861ead9fcfe6208f2749032 | https://github.com/singnet/snet-cli/blob/1b5ac98cb9a64211c861ead9fcfe6208f2749032/snet_cli/mpe_service_metadata.py#L103-L109 |
236,954 | singnet/snet-cli | snet_cli/mpe_service_metadata.py | MPEServiceMetadata.get_group_name_nonetrick | def get_group_name_nonetrick(self, group_name = None):
""" In all getter function in case of single payment group, group_name can be None """
groups = self.m["groups"]
if (len(groups) == 0):
raise Exception("Cannot find any groups in metadata")
if (not group_name):
... | python | def get_group_name_nonetrick(self, group_name = None):
""" In all getter function in case of single payment group, group_name can be None """
groups = self.m["groups"]
if (len(groups) == 0):
raise Exception("Cannot find any groups in metadata")
if (not group_name):
... | [
"def",
"get_group_name_nonetrick",
"(",
"self",
",",
"group_name",
"=",
"None",
")",
":",
"groups",
"=",
"self",
".",
"m",
"[",
"\"groups\"",
"]",
"if",
"(",
"len",
"(",
"groups",
")",
"==",
"0",
")",
":",
"raise",
"Exception",
"(",
"\"Cannot find any gr... | In all getter function in case of single payment group, group_name can be None | [
"In",
"all",
"getter",
"function",
"in",
"case",
"of",
"single",
"payment",
"group",
"group_name",
"can",
"be",
"None"
] | 1b5ac98cb9a64211c861ead9fcfe6208f2749032 | https://github.com/singnet/snet-cli/blob/1b5ac98cb9a64211c861ead9fcfe6208f2749032/snet_cli/mpe_service_metadata.py#L145-L154 |
236,955 | singnet/snet-cli | snet_cli/utils_ipfs.py | get_from_ipfs_and_checkhash | def get_from_ipfs_and_checkhash(ipfs_client, ipfs_hash_base58, validate=True):
"""
Get file from ipfs
We must check the hash becasue we cannot believe that ipfs_client wasn't been compromise
"""
if validate:
from snet_cli.resources.proto.unixfs_pb2 import Data
from snet_cli.resources... | python | def get_from_ipfs_and_checkhash(ipfs_client, ipfs_hash_base58, validate=True):
"""
Get file from ipfs
We must check the hash becasue we cannot believe that ipfs_client wasn't been compromise
"""
if validate:
from snet_cli.resources.proto.unixfs_pb2 import Data
from snet_cli.resources... | [
"def",
"get_from_ipfs_and_checkhash",
"(",
"ipfs_client",
",",
"ipfs_hash_base58",
",",
"validate",
"=",
"True",
")",
":",
"if",
"validate",
":",
"from",
"snet_cli",
".",
"resources",
".",
"proto",
".",
"unixfs_pb2",
"import",
"Data",
"from",
"snet_cli",
".",
... | Get file from ipfs
We must check the hash becasue we cannot believe that ipfs_client wasn't been compromise | [
"Get",
"file",
"from",
"ipfs",
"We",
"must",
"check",
"the",
"hash",
"becasue",
"we",
"cannot",
"believe",
"that",
"ipfs_client",
"wasn",
"t",
"been",
"compromise"
] | 1b5ac98cb9a64211c861ead9fcfe6208f2749032 | https://github.com/singnet/snet-cli/blob/1b5ac98cb9a64211c861ead9fcfe6208f2749032/snet_cli/utils_ipfs.py#L35-L63 |
236,956 | singnet/snet-cli | snet_cli/utils_ipfs.py | hash_to_bytesuri | def hash_to_bytesuri(s):
"""
Convert in and from bytes uri format used in Registry contract
"""
# TODO: we should pad string with zeros till closest 32 bytes word because of a bug in processReceipt (in snet_cli.contract.process_receipt)
s = "ipfs://" + s
return s.encode("ascii").ljust(32 * (len(... | python | def hash_to_bytesuri(s):
"""
Convert in and from bytes uri format used in Registry contract
"""
# TODO: we should pad string with zeros till closest 32 bytes word because of a bug in processReceipt (in snet_cli.contract.process_receipt)
s = "ipfs://" + s
return s.encode("ascii").ljust(32 * (len(... | [
"def",
"hash_to_bytesuri",
"(",
"s",
")",
":",
"# TODO: we should pad string with zeros till closest 32 bytes word because of a bug in processReceipt (in snet_cli.contract.process_receipt)",
"s",
"=",
"\"ipfs://\"",
"+",
"s",
"return",
"s",
".",
"encode",
"(",
"\"ascii\"",
")",
... | Convert in and from bytes uri format used in Registry contract | [
"Convert",
"in",
"and",
"from",
"bytes",
"uri",
"format",
"used",
"in",
"Registry",
"contract"
] | 1b5ac98cb9a64211c861ead9fcfe6208f2749032 | https://github.com/singnet/snet-cli/blob/1b5ac98cb9a64211c861ead9fcfe6208f2749032/snet_cli/utils_ipfs.py#L65-L71 |
236,957 | singnet/snet-cli | snet_cli/mpe_treasurer_command.py | MPETreasurerCommand._get_stub_and_request_classes | def _get_stub_and_request_classes(self, service_name):
""" import protobuf and return stub and request class """
# Compile protobuf if needed
codegen_dir = Path.home().joinpath(".snet", "mpe_client", "control_service")
proto_dir = Path(__file__).absolute().parent.joinpath("resources", ... | python | def _get_stub_and_request_classes(self, service_name):
""" import protobuf and return stub and request class """
# Compile protobuf if needed
codegen_dir = Path.home().joinpath(".snet", "mpe_client", "control_service")
proto_dir = Path(__file__).absolute().parent.joinpath("resources", ... | [
"def",
"_get_stub_and_request_classes",
"(",
"self",
",",
"service_name",
")",
":",
"# Compile protobuf if needed",
"codegen_dir",
"=",
"Path",
".",
"home",
"(",
")",
".",
"joinpath",
"(",
"\".snet\"",
",",
"\"mpe_client\"",
",",
"\"control_service\"",
")",
"proto_d... | import protobuf and return stub and request class | [
"import",
"protobuf",
"and",
"return",
"stub",
"and",
"request",
"class"
] | 1b5ac98cb9a64211c861ead9fcfe6208f2749032 | https://github.com/singnet/snet-cli/blob/1b5ac98cb9a64211c861ead9fcfe6208f2749032/snet_cli/mpe_treasurer_command.py#L31-L40 |
236,958 | singnet/snet-cli | snet_cli/mpe_treasurer_command.py | MPETreasurerCommand._start_claim_channels | def _start_claim_channels(self, grpc_channel, channels_ids):
""" Safely run StartClaim for given channels """
unclaimed_payments = self._call_GetListUnclaimed(grpc_channel)
unclaimed_payments_dict = {p["channel_id"] : p for p in unclaimed_payments}
to_claim = []
for channel_id i... | python | def _start_claim_channels(self, grpc_channel, channels_ids):
""" Safely run StartClaim for given channels """
unclaimed_payments = self._call_GetListUnclaimed(grpc_channel)
unclaimed_payments_dict = {p["channel_id"] : p for p in unclaimed_payments}
to_claim = []
for channel_id i... | [
"def",
"_start_claim_channels",
"(",
"self",
",",
"grpc_channel",
",",
"channels_ids",
")",
":",
"unclaimed_payments",
"=",
"self",
".",
"_call_GetListUnclaimed",
"(",
"grpc_channel",
")",
"unclaimed_payments_dict",
"=",
"{",
"p",
"[",
"\"channel_id\"",
"]",
":",
... | Safely run StartClaim for given channels | [
"Safely",
"run",
"StartClaim",
"for",
"given",
"channels"
] | 1b5ac98cb9a64211c861ead9fcfe6208f2749032 | https://github.com/singnet/snet-cli/blob/1b5ac98cb9a64211c861ead9fcfe6208f2749032/snet_cli/mpe_treasurer_command.py#L103-L120 |
236,959 | singnet/snet-cli | snet_cli/mpe_treasurer_command.py | MPETreasurerCommand._claim_in_progress_and_claim_channels | def _claim_in_progress_and_claim_channels(self, grpc_channel, channels):
""" Claim all 'pending' payments in progress and after we claim given channels """
# first we get the list of all 'payments in progress' in case we 'lost' some payments.
payments = self._call_GetListInProgress(grpc_channel)... | python | def _claim_in_progress_and_claim_channels(self, grpc_channel, channels):
""" Claim all 'pending' payments in progress and after we claim given channels """
# first we get the list of all 'payments in progress' in case we 'lost' some payments.
payments = self._call_GetListInProgress(grpc_channel)... | [
"def",
"_claim_in_progress_and_claim_channels",
"(",
"self",
",",
"grpc_channel",
",",
"channels",
")",
":",
"# first we get the list of all 'payments in progress' in case we 'lost' some payments.",
"payments",
"=",
"self",
".",
"_call_GetListInProgress",
"(",
"grpc_channel",
")"... | Claim all 'pending' payments in progress and after we claim given channels | [
"Claim",
"all",
"pending",
"payments",
"in",
"progress",
"and",
"after",
"we",
"claim",
"given",
"channels"
] | 1b5ac98cb9a64211c861ead9fcfe6208f2749032 | https://github.com/singnet/snet-cli/blob/1b5ac98cb9a64211c861ead9fcfe6208f2749032/snet_cli/mpe_treasurer_command.py#L122-L130 |
236,960 | singnet/snet-cli | snet_cli/config.py | Config.create_default_config | def create_default_config(self):
""" Create default configuration if config file does not exist """
# make config directory with the minimal possible permission
self._config_file.parent.mkdir(mode=0o700, exist_ok=True)
self["network.kovan"] = {"default_eth_rpc_endpoint": "https://kovan... | python | def create_default_config(self):
""" Create default configuration if config file does not exist """
# make config directory with the minimal possible permission
self._config_file.parent.mkdir(mode=0o700, exist_ok=True)
self["network.kovan"] = {"default_eth_rpc_endpoint": "https://kovan... | [
"def",
"create_default_config",
"(",
"self",
")",
":",
"# make config directory with the minimal possible permission",
"self",
".",
"_config_file",
".",
"parent",
".",
"mkdir",
"(",
"mode",
"=",
"0o700",
",",
"exist_ok",
"=",
"True",
")",
"self",
"[",
"\"network.kov... | Create default configuration if config file does not exist | [
"Create",
"default",
"configuration",
"if",
"config",
"file",
"does",
"not",
"exist"
] | 1b5ac98cb9a64211c861ead9fcfe6208f2749032 | https://github.com/singnet/snet-cli/blob/1b5ac98cb9a64211c861ead9fcfe6208f2749032/snet_cli/config.py#L175-L187 |
236,961 | singnet/snet-cli | snet_cli/utils_proto.py | switch_to_json_payload_encoding | def switch_to_json_payload_encoding(call_fn, response_class):
""" Switch payload encoding to JSON for GRPC call """
def json_serializer(*args, **kwargs):
return bytes(json_format.MessageToJson(args[0], True, preserving_proto_field_name=True), "utf-8")
def json_deserializer(*args, **kwargs):
... | python | def switch_to_json_payload_encoding(call_fn, response_class):
""" Switch payload encoding to JSON for GRPC call """
def json_serializer(*args, **kwargs):
return bytes(json_format.MessageToJson(args[0], True, preserving_proto_field_name=True), "utf-8")
def json_deserializer(*args, **kwargs):
... | [
"def",
"switch_to_json_payload_encoding",
"(",
"call_fn",
",",
"response_class",
")",
":",
"def",
"json_serializer",
"(",
"*",
"args",
",",
"*",
"*",
"kwargs",
")",
":",
"return",
"bytes",
"(",
"json_format",
".",
"MessageToJson",
"(",
"args",
"[",
"0",
"]",... | Switch payload encoding to JSON for GRPC call | [
"Switch",
"payload",
"encoding",
"to",
"JSON",
"for",
"GRPC",
"call"
] | 1b5ac98cb9a64211c861ead9fcfe6208f2749032 | https://github.com/singnet/snet-cli/blob/1b5ac98cb9a64211c861ead9fcfe6208f2749032/snet_cli/utils_proto.py#L72-L81 |
236,962 | singnet/snet-cli | snet_cli/mpe_account_command.py | MPEAccountCommand.print_agi_and_mpe_balances | def print_agi_and_mpe_balances(self):
""" Print balance of ETH, AGI, and MPE wallet """
if (self.args.account):
account = self.args.account
else:
account = self.ident.address
eth_wei = self.w3.eth.getBalance(account)
agi_cogs = self.call_contract_command(... | python | def print_agi_and_mpe_balances(self):
""" Print balance of ETH, AGI, and MPE wallet """
if (self.args.account):
account = self.args.account
else:
account = self.ident.address
eth_wei = self.w3.eth.getBalance(account)
agi_cogs = self.call_contract_command(... | [
"def",
"print_agi_and_mpe_balances",
"(",
"self",
")",
":",
"if",
"(",
"self",
".",
"args",
".",
"account",
")",
":",
"account",
"=",
"self",
".",
"args",
".",
"account",
"else",
":",
"account",
"=",
"self",
".",
"ident",
".",
"address",
"eth_wei",
"="... | Print balance of ETH, AGI, and MPE wallet | [
"Print",
"balance",
"of",
"ETH",
"AGI",
"and",
"MPE",
"wallet"
] | 1b5ac98cb9a64211c861ead9fcfe6208f2749032 | https://github.com/singnet/snet-cli/blob/1b5ac98cb9a64211c861ead9fcfe6208f2749032/snet_cli/mpe_account_command.py#L10-L24 |
236,963 | singnet/snet-cli | snet_cli/mpe_service_command.py | MPEServiceCommand.publish_proto_in_ipfs | def publish_proto_in_ipfs(self):
""" Publish proto files in ipfs and print hash """
ipfs_hash_base58 = utils_ipfs.publish_proto_in_ipfs(self._get_ipfs_client(), self.args.protodir)
self._printout(ipfs_hash_base58) | python | def publish_proto_in_ipfs(self):
""" Publish proto files in ipfs and print hash """
ipfs_hash_base58 = utils_ipfs.publish_proto_in_ipfs(self._get_ipfs_client(), self.args.protodir)
self._printout(ipfs_hash_base58) | [
"def",
"publish_proto_in_ipfs",
"(",
"self",
")",
":",
"ipfs_hash_base58",
"=",
"utils_ipfs",
".",
"publish_proto_in_ipfs",
"(",
"self",
".",
"_get_ipfs_client",
"(",
")",
",",
"self",
".",
"args",
".",
"protodir",
")",
"self",
".",
"_printout",
"(",
"ipfs_has... | Publish proto files in ipfs and print hash | [
"Publish",
"proto",
"files",
"in",
"ipfs",
"and",
"print",
"hash"
] | 1b5ac98cb9a64211c861ead9fcfe6208f2749032 | https://github.com/singnet/snet-cli/blob/1b5ac98cb9a64211c861ead9fcfe6208f2749032/snet_cli/mpe_service_command.py#L15-L18 |
236,964 | singnet/snet-cli | snet_cli/mpe_service_command.py | MPEServiceCommand.publish_proto_metadata_update | def publish_proto_metadata_update(self):
""" Publish protobuf model in ipfs and update existing metadata file """
metadata = load_mpe_service_metadata(self.args.metadata_file)
ipfs_hash_base58 = utils_ipfs.publish_proto_in_ipfs(self._get_ipfs_client(), self.args.protodir)
metadata.set_si... | python | def publish_proto_metadata_update(self):
""" Publish protobuf model in ipfs and update existing metadata file """
metadata = load_mpe_service_metadata(self.args.metadata_file)
ipfs_hash_base58 = utils_ipfs.publish_proto_in_ipfs(self._get_ipfs_client(), self.args.protodir)
metadata.set_si... | [
"def",
"publish_proto_metadata_update",
"(",
"self",
")",
":",
"metadata",
"=",
"load_mpe_service_metadata",
"(",
"self",
".",
"args",
".",
"metadata_file",
")",
"ipfs_hash_base58",
"=",
"utils_ipfs",
".",
"publish_proto_in_ipfs",
"(",
"self",
".",
"_get_ipfs_client",... | Publish protobuf model in ipfs and update existing metadata file | [
"Publish",
"protobuf",
"model",
"in",
"ipfs",
"and",
"update",
"existing",
"metadata",
"file"
] | 1b5ac98cb9a64211c861ead9fcfe6208f2749032 | https://github.com/singnet/snet-cli/blob/1b5ac98cb9a64211c861ead9fcfe6208f2749032/snet_cli/mpe_service_command.py#L37-L42 |
236,965 | singnet/snet-cli | snet_cli/mpe_channel_command.py | MPEChannelCommand._get_persistent_mpe_dir | def _get_persistent_mpe_dir(self):
""" get persistent storage for mpe """
mpe_address = self.get_mpe_address().lower()
registry_address = self.get_registry_address().lower()
return Path.home().joinpath(".snet", "mpe_client", "%s_%s"%(mpe_address, registry_address)) | python | def _get_persistent_mpe_dir(self):
""" get persistent storage for mpe """
mpe_address = self.get_mpe_address().lower()
registry_address = self.get_registry_address().lower()
return Path.home().joinpath(".snet", "mpe_client", "%s_%s"%(mpe_address, registry_address)) | [
"def",
"_get_persistent_mpe_dir",
"(",
"self",
")",
":",
"mpe_address",
"=",
"self",
".",
"get_mpe_address",
"(",
")",
".",
"lower",
"(",
")",
"registry_address",
"=",
"self",
".",
"get_registry_address",
"(",
")",
".",
"lower",
"(",
")",
"return",
"Path",
... | get persistent storage for mpe | [
"get",
"persistent",
"storage",
"for",
"mpe"
] | 1b5ac98cb9a64211c861ead9fcfe6208f2749032 | https://github.com/singnet/snet-cli/blob/1b5ac98cb9a64211c861ead9fcfe6208f2749032/snet_cli/mpe_channel_command.py#L21-L25 |
236,966 | singnet/snet-cli | snet_cli/mpe_channel_command.py | MPEChannelCommand._check_mpe_address_metadata | def _check_mpe_address_metadata(self, metadata):
""" we make sure that MultiPartyEscrow address from metadata is correct """
mpe_address = self.get_mpe_address()
if (str(mpe_address).lower() != str(metadata["mpe_address"]).lower()):
raise Exception("MultiPartyEscrow contract address ... | python | def _check_mpe_address_metadata(self, metadata):
""" we make sure that MultiPartyEscrow address from metadata is correct """
mpe_address = self.get_mpe_address()
if (str(mpe_address).lower() != str(metadata["mpe_address"]).lower()):
raise Exception("MultiPartyEscrow contract address ... | [
"def",
"_check_mpe_address_metadata",
"(",
"self",
",",
"metadata",
")",
":",
"mpe_address",
"=",
"self",
".",
"get_mpe_address",
"(",
")",
"if",
"(",
"str",
"(",
"mpe_address",
")",
".",
"lower",
"(",
")",
"!=",
"str",
"(",
"metadata",
"[",
"\"mpe_address... | we make sure that MultiPartyEscrow address from metadata is correct | [
"we",
"make",
"sure",
"that",
"MultiPartyEscrow",
"address",
"from",
"metadata",
"is",
"correct"
] | 1b5ac98cb9a64211c861ead9fcfe6208f2749032 | https://github.com/singnet/snet-cli/blob/1b5ac98cb9a64211c861ead9fcfe6208f2749032/snet_cli/mpe_channel_command.py#L74-L78 |
236,967 | singnet/snet-cli | snet_cli/mpe_channel_command.py | MPEChannelCommand._init_or_update_registered_service_if_needed | def _init_or_update_registered_service_if_needed(self):
'''
similar to _init_or_update_service_if_needed but we get service_registraion from registry,
so we can update only registered services
'''
if (self.is_service_initialized()):
old_reg = self._read_service_info(s... | python | def _init_or_update_registered_service_if_needed(self):
'''
similar to _init_or_update_service_if_needed but we get service_registraion from registry,
so we can update only registered services
'''
if (self.is_service_initialized()):
old_reg = self._read_service_info(s... | [
"def",
"_init_or_update_registered_service_if_needed",
"(",
"self",
")",
":",
"if",
"(",
"self",
".",
"is_service_initialized",
"(",
")",
")",
":",
"old_reg",
"=",
"self",
".",
"_read_service_info",
"(",
"self",
".",
"args",
".",
"org_id",
",",
"self",
".",
... | similar to _init_or_update_service_if_needed but we get service_registraion from registry,
so we can update only registered services | [
"similar",
"to",
"_init_or_update_service_if_needed",
"but",
"we",
"get",
"service_registraion",
"from",
"registry",
"so",
"we",
"can",
"update",
"only",
"registered",
"services"
] | 1b5ac98cb9a64211c861ead9fcfe6208f2749032 | https://github.com/singnet/snet-cli/blob/1b5ac98cb9a64211c861ead9fcfe6208f2749032/snet_cli/mpe_channel_command.py#L115-L136 |
236,968 | singnet/snet-cli | snet_cli/mpe_channel_command.py | MPEChannelCommand._smart_get_initialized_channel_for_service | def _smart_get_initialized_channel_for_service(self, metadata, filter_by, is_try_initailize = True):
'''
- filter_by can be sender or signer
'''
channels = self._get_initialized_channels_for_service(self.args.org_id, self.args.service_id)
group_id = metadata.get_group_id(self.ar... | python | def _smart_get_initialized_channel_for_service(self, metadata, filter_by, is_try_initailize = True):
'''
- filter_by can be sender or signer
'''
channels = self._get_initialized_channels_for_service(self.args.org_id, self.args.service_id)
group_id = metadata.get_group_id(self.ar... | [
"def",
"_smart_get_initialized_channel_for_service",
"(",
"self",
",",
"metadata",
",",
"filter_by",
",",
"is_try_initailize",
"=",
"True",
")",
":",
"channels",
"=",
"self",
".",
"_get_initialized_channels_for_service",
"(",
"self",
".",
"args",
".",
"org_id",
",",... | - filter_by can be sender or signer | [
"-",
"filter_by",
"can",
"be",
"sender",
"or",
"signer"
] | 1b5ac98cb9a64211c861ead9fcfe6208f2749032 | https://github.com/singnet/snet-cli/blob/1b5ac98cb9a64211c861ead9fcfe6208f2749032/snet_cli/mpe_channel_command.py#L297-L320 |
236,969 | singnet/snet-cli | snet_cli/mpe_channel_command.py | MPEChannelCommand._get_all_filtered_channels | def _get_all_filtered_channels(self, topics_without_signature):
""" get all filtered chanels from blockchain logs """
mpe_address = self.get_mpe_address()
event_signature = self.ident.w3.sha3(text="ChannelOpen(uint256,uint256,address,address,address,bytes32,uint256,uint256)").hex()
t... | python | def _get_all_filtered_channels(self, topics_without_signature):
""" get all filtered chanels from blockchain logs """
mpe_address = self.get_mpe_address()
event_signature = self.ident.w3.sha3(text="ChannelOpen(uint256,uint256,address,address,address,bytes32,uint256,uint256)").hex()
t... | [
"def",
"_get_all_filtered_channels",
"(",
"self",
",",
"topics_without_signature",
")",
":",
"mpe_address",
"=",
"self",
".",
"get_mpe_address",
"(",
")",
"event_signature",
"=",
"self",
".",
"ident",
".",
"w3",
".",
"sha3",
"(",
"text",
"=",
"\"ChannelOpen(uint... | get all filtered chanels from blockchain logs | [
"get",
"all",
"filtered",
"chanels",
"from",
"blockchain",
"logs"
] | 1b5ac98cb9a64211c861ead9fcfe6208f2749032 | https://github.com/singnet/snet-cli/blob/1b5ac98cb9a64211c861ead9fcfe6208f2749032/snet_cli/mpe_channel_command.py#L415-L424 |
236,970 | pachyderm/python-pachyderm | src/python_pachyderm/pfs_client.py | PfsClient.list_repo | def list_repo(self):
"""
Returns info about all Repos.
"""
req = proto.ListRepoRequest()
res = self.stub.ListRepo(req, metadata=self.metadata)
if hasattr(res, 'repo_info'):
return res.repo_info
return [] | python | def list_repo(self):
"""
Returns info about all Repos.
"""
req = proto.ListRepoRequest()
res = self.stub.ListRepo(req, metadata=self.metadata)
if hasattr(res, 'repo_info'):
return res.repo_info
return [] | [
"def",
"list_repo",
"(",
"self",
")",
":",
"req",
"=",
"proto",
".",
"ListRepoRequest",
"(",
")",
"res",
"=",
"self",
".",
"stub",
".",
"ListRepo",
"(",
"req",
",",
"metadata",
"=",
"self",
".",
"metadata",
")",
"if",
"hasattr",
"(",
"res",
",",
"'... | Returns info about all Repos. | [
"Returns",
"info",
"about",
"all",
"Repos",
"."
] | 1c58cf91d30e03716a4f45213989e890f7b8a78c | https://github.com/pachyderm/python-pachyderm/blob/1c58cf91d30e03716a4f45213989e890f7b8a78c/src/python_pachyderm/pfs_client.py#L71-L79 |
236,971 | pachyderm/python-pachyderm | src/python_pachyderm/pfs_client.py | PfsClient.delete_repo | def delete_repo(self, repo_name=None, force=False, all=False):
"""
Deletes a repo and reclaims the storage space it was using.
Params:
* repo_name: The name of the repo.
* force: If set to true, the repo will be removed regardless of
errors. This argument should be used ... | python | def delete_repo(self, repo_name=None, force=False, all=False):
"""
Deletes a repo and reclaims the storage space it was using.
Params:
* repo_name: The name of the repo.
* force: If set to true, the repo will be removed regardless of
errors. This argument should be used ... | [
"def",
"delete_repo",
"(",
"self",
",",
"repo_name",
"=",
"None",
",",
"force",
"=",
"False",
",",
"all",
"=",
"False",
")",
":",
"if",
"not",
"all",
":",
"if",
"repo_name",
":",
"req",
"=",
"proto",
".",
"DeleteRepoRequest",
"(",
"repo",
"=",
"proto... | Deletes a repo and reclaims the storage space it was using.
Params:
* repo_name: The name of the repo.
* force: If set to true, the repo will be removed regardless of
errors. This argument should be used with care.
* all: Delete all repos. | [
"Deletes",
"a",
"repo",
"and",
"reclaims",
"the",
"storage",
"space",
"it",
"was",
"using",
"."
] | 1c58cf91d30e03716a4f45213989e890f7b8a78c | https://github.com/pachyderm/python-pachyderm/blob/1c58cf91d30e03716a4f45213989e890f7b8a78c/src/python_pachyderm/pfs_client.py#L81-L102 |
236,972 | pachyderm/python-pachyderm | src/python_pachyderm/pfs_client.py | PfsClient.start_commit | def start_commit(self, repo_name, branch=None, parent=None, description=None):
"""
Begins the process of committing data to a Repo. Once started you can
write to the Commit with PutFile and when all the data has been
written you must finish the Commit with FinishCommit. NOTE, data is
... | python | def start_commit(self, repo_name, branch=None, parent=None, description=None):
"""
Begins the process of committing data to a Repo. Once started you can
write to the Commit with PutFile and when all the data has been
written you must finish the Commit with FinishCommit. NOTE, data is
... | [
"def",
"start_commit",
"(",
"self",
",",
"repo_name",
",",
"branch",
"=",
"None",
",",
"parent",
"=",
"None",
",",
"description",
"=",
"None",
")",
":",
"req",
"=",
"proto",
".",
"StartCommitRequest",
"(",
"parent",
"=",
"proto",
".",
"Commit",
"(",
"r... | Begins the process of committing data to a Repo. Once started you can
write to the Commit with PutFile and when all the data has been
written you must finish the Commit with FinishCommit. NOTE, data is
not persisted until FinishCommit is called. A Commit object is
returned.
Para... | [
"Begins",
"the",
"process",
"of",
"committing",
"data",
"to",
"a",
"Repo",
".",
"Once",
"started",
"you",
"can",
"write",
"to",
"the",
"Commit",
"with",
"PutFile",
"and",
"when",
"all",
"the",
"data",
"has",
"been",
"written",
"you",
"must",
"finish",
"t... | 1c58cf91d30e03716a4f45213989e890f7b8a78c | https://github.com/pachyderm/python-pachyderm/blob/1c58cf91d30e03716a4f45213989e890f7b8a78c/src/python_pachyderm/pfs_client.py#L104-L129 |
236,973 | pachyderm/python-pachyderm | src/python_pachyderm/pfs_client.py | PfsClient.finish_commit | def finish_commit(self, commit):
"""
Ends the process of committing data to a Repo and persists the
Commit. Once a Commit is finished the data becomes immutable and
future attempts to write to it with PutFile will error.
Params:
* commit: A tuple, string, or Commit objec... | python | def finish_commit(self, commit):
"""
Ends the process of committing data to a Repo and persists the
Commit. Once a Commit is finished the data becomes immutable and
future attempts to write to it with PutFile will error.
Params:
* commit: A tuple, string, or Commit objec... | [
"def",
"finish_commit",
"(",
"self",
",",
"commit",
")",
":",
"req",
"=",
"proto",
".",
"FinishCommitRequest",
"(",
"commit",
"=",
"commit_from",
"(",
"commit",
")",
")",
"res",
"=",
"self",
".",
"stub",
".",
"FinishCommit",
"(",
"req",
",",
"metadata",
... | Ends the process of committing data to a Repo and persists the
Commit. Once a Commit is finished the data becomes immutable and
future attempts to write to it with PutFile will error.
Params:
* commit: A tuple, string, or Commit object representing the commit. | [
"Ends",
"the",
"process",
"of",
"committing",
"data",
"to",
"a",
"Repo",
"and",
"persists",
"the",
"Commit",
".",
"Once",
"a",
"Commit",
"is",
"finished",
"the",
"data",
"becomes",
"immutable",
"and",
"future",
"attempts",
"to",
"write",
"to",
"it",
"with"... | 1c58cf91d30e03716a4f45213989e890f7b8a78c | https://github.com/pachyderm/python-pachyderm/blob/1c58cf91d30e03716a4f45213989e890f7b8a78c/src/python_pachyderm/pfs_client.py#L131-L142 |
236,974 | pachyderm/python-pachyderm | src/python_pachyderm/pfs_client.py | PfsClient.commit | def commit(self, repo_name, branch=None, parent=None, description=None):
"""A context manager for doing stuff inside a commit."""
commit = self.start_commit(repo_name, branch, parent, description)
try:
yield commit
except Exception as e:
print("An exception occurr... | python | def commit(self, repo_name, branch=None, parent=None, description=None):
"""A context manager for doing stuff inside a commit."""
commit = self.start_commit(repo_name, branch, parent, description)
try:
yield commit
except Exception as e:
print("An exception occurr... | [
"def",
"commit",
"(",
"self",
",",
"repo_name",
",",
"branch",
"=",
"None",
",",
"parent",
"=",
"None",
",",
"description",
"=",
"None",
")",
":",
"commit",
"=",
"self",
".",
"start_commit",
"(",
"repo_name",
",",
"branch",
",",
"parent",
",",
"descrip... | A context manager for doing stuff inside a commit. | [
"A",
"context",
"manager",
"for",
"doing",
"stuff",
"inside",
"a",
"commit",
"."
] | 1c58cf91d30e03716a4f45213989e890f7b8a78c | https://github.com/pachyderm/python-pachyderm/blob/1c58cf91d30e03716a4f45213989e890f7b8a78c/src/python_pachyderm/pfs_client.py#L145-L155 |
236,975 | pachyderm/python-pachyderm | src/python_pachyderm/pfs_client.py | PfsClient.inspect_commit | def inspect_commit(self, commit):
"""
Returns info about a specific Commit.
Params:
* commit: A tuple, string, or Commit object representing the commit.
"""
req = proto.InspectCommitRequest(commit=commit_from(commit))
return self.stub.InspectCommit(req, metadata=... | python | def inspect_commit(self, commit):
"""
Returns info about a specific Commit.
Params:
* commit: A tuple, string, or Commit object representing the commit.
"""
req = proto.InspectCommitRequest(commit=commit_from(commit))
return self.stub.InspectCommit(req, metadata=... | [
"def",
"inspect_commit",
"(",
"self",
",",
"commit",
")",
":",
"req",
"=",
"proto",
".",
"InspectCommitRequest",
"(",
"commit",
"=",
"commit_from",
"(",
"commit",
")",
")",
"return",
"self",
".",
"stub",
".",
"InspectCommit",
"(",
"req",
",",
"metadata",
... | Returns info about a specific Commit.
Params:
* commit: A tuple, string, or Commit object representing the commit. | [
"Returns",
"info",
"about",
"a",
"specific",
"Commit",
"."
] | 1c58cf91d30e03716a4f45213989e890f7b8a78c | https://github.com/pachyderm/python-pachyderm/blob/1c58cf91d30e03716a4f45213989e890f7b8a78c/src/python_pachyderm/pfs_client.py#L157-L165 |
236,976 | pachyderm/python-pachyderm | src/python_pachyderm/pfs_client.py | PfsClient.list_commit | def list_commit(self, repo_name, to_commit=None, from_commit=None, number=0):
"""
Gets a list of CommitInfo objects.
Params:
* repo_name: If only `repo_name` is given, all commits in the repo are
returned.
* to_commit: Optional. Only the ancestors of `to`, including `to`... | python | def list_commit(self, repo_name, to_commit=None, from_commit=None, number=0):
"""
Gets a list of CommitInfo objects.
Params:
* repo_name: If only `repo_name` is given, all commits in the repo are
returned.
* to_commit: Optional. Only the ancestors of `to`, including `to`... | [
"def",
"list_commit",
"(",
"self",
",",
"repo_name",
",",
"to_commit",
"=",
"None",
",",
"from_commit",
"=",
"None",
",",
"number",
"=",
"0",
")",
":",
"req",
"=",
"proto",
".",
"ListCommitRequest",
"(",
"repo",
"=",
"proto",
".",
"Repo",
"(",
"name",
... | Gets a list of CommitInfo objects.
Params:
* repo_name: If only `repo_name` is given, all commits in the repo are
returned.
* to_commit: Optional. Only the ancestors of `to`, including `to`
itself, are considered.
* from_commit: Optional. Only the descendants of `from`, ... | [
"Gets",
"a",
"list",
"of",
"CommitInfo",
"objects",
"."
] | 1c58cf91d30e03716a4f45213989e890f7b8a78c | https://github.com/pachyderm/python-pachyderm/blob/1c58cf91d30e03716a4f45213989e890f7b8a78c/src/python_pachyderm/pfs_client.py#L177-L200 |
236,977 | pachyderm/python-pachyderm | src/python_pachyderm/pfs_client.py | PfsClient.delete_commit | def delete_commit(self, commit):
"""
Deletes a commit.
Params:
* commit: A tuple, string, or Commit object representing the commit.
"""
req = proto.DeleteCommitRequest(commit=commit_from(commit))
self.stub.DeleteCommit(req, metadata=self.metadata) | python | def delete_commit(self, commit):
"""
Deletes a commit.
Params:
* commit: A tuple, string, or Commit object representing the commit.
"""
req = proto.DeleteCommitRequest(commit=commit_from(commit))
self.stub.DeleteCommit(req, metadata=self.metadata) | [
"def",
"delete_commit",
"(",
"self",
",",
"commit",
")",
":",
"req",
"=",
"proto",
".",
"DeleteCommitRequest",
"(",
"commit",
"=",
"commit_from",
"(",
"commit",
")",
")",
"self",
".",
"stub",
".",
"DeleteCommit",
"(",
"req",
",",
"metadata",
"=",
"self",... | Deletes a commit.
Params:
* commit: A tuple, string, or Commit object representing the commit. | [
"Deletes",
"a",
"commit",
"."
] | 1c58cf91d30e03716a4f45213989e890f7b8a78c | https://github.com/pachyderm/python-pachyderm/blob/1c58cf91d30e03716a4f45213989e890f7b8a78c/src/python_pachyderm/pfs_client.py#L202-L210 |
236,978 | pachyderm/python-pachyderm | src/python_pachyderm/pfs_client.py | PfsClient.flush_commit | def flush_commit(self, commits, repos=tuple()):
"""
Blocks until all of the commits which have a set of commits as
provenance have finished. For commits to be considered they must have
all of the specified commits as provenance. This in effect waits for
all of the jobs that are t... | python | def flush_commit(self, commits, repos=tuple()):
"""
Blocks until all of the commits which have a set of commits as
provenance have finished. For commits to be considered they must have
all of the specified commits as provenance. This in effect waits for
all of the jobs that are t... | [
"def",
"flush_commit",
"(",
"self",
",",
"commits",
",",
"repos",
"=",
"tuple",
"(",
")",
")",
":",
"req",
"=",
"proto",
".",
"FlushCommitRequest",
"(",
"commit",
"=",
"[",
"commit_from",
"(",
"c",
")",
"for",
"c",
"in",
"commits",
"]",
",",
"to_repo... | Blocks until all of the commits which have a set of commits as
provenance have finished. For commits to be considered they must have
all of the specified commits as provenance. This in effect waits for
all of the jobs that are triggered by a set of commits to complete.
It returns an erro... | [
"Blocks",
"until",
"all",
"of",
"the",
"commits",
"which",
"have",
"a",
"set",
"of",
"commits",
"as",
"provenance",
"have",
"finished",
".",
"For",
"commits",
"to",
"be",
"considered",
"they",
"must",
"have",
"all",
"of",
"the",
"specified",
"commits",
"as... | 1c58cf91d30e03716a4f45213989e890f7b8a78c | https://github.com/pachyderm/python-pachyderm/blob/1c58cf91d30e03716a4f45213989e890f7b8a78c/src/python_pachyderm/pfs_client.py#L212-L233 |
236,979 | pachyderm/python-pachyderm | src/python_pachyderm/pfs_client.py | PfsClient.subscribe_commit | def subscribe_commit(self, repo_name, branch, from_commit_id=None):
"""
SubscribeCommit is like ListCommit but it keeps listening for commits as
they come in. This returns an iterator Commit objects.
Params:
* repo_name: Name of the repo.
* branch: Branch to subscribe to... | python | def subscribe_commit(self, repo_name, branch, from_commit_id=None):
"""
SubscribeCommit is like ListCommit but it keeps listening for commits as
they come in. This returns an iterator Commit objects.
Params:
* repo_name: Name of the repo.
* branch: Branch to subscribe to... | [
"def",
"subscribe_commit",
"(",
"self",
",",
"repo_name",
",",
"branch",
",",
"from_commit_id",
"=",
"None",
")",
":",
"repo",
"=",
"proto",
".",
"Repo",
"(",
"name",
"=",
"repo_name",
")",
"req",
"=",
"proto",
".",
"SubscribeCommitRequest",
"(",
"repo",
... | SubscribeCommit is like ListCommit but it keeps listening for commits as
they come in. This returns an iterator Commit objects.
Params:
* repo_name: Name of the repo.
* branch: Branch to subscribe to.
* from_commit_id: Optional. Only commits created since this commit
are... | [
"SubscribeCommit",
"is",
"like",
"ListCommit",
"but",
"it",
"keeps",
"listening",
"for",
"commits",
"as",
"they",
"come",
"in",
".",
"This",
"returns",
"an",
"iterator",
"Commit",
"objects",
"."
] | 1c58cf91d30e03716a4f45213989e890f7b8a78c | https://github.com/pachyderm/python-pachyderm/blob/1c58cf91d30e03716a4f45213989e890f7b8a78c/src/python_pachyderm/pfs_client.py#L235-L251 |
236,980 | pachyderm/python-pachyderm | src/python_pachyderm/pfs_client.py | PfsClient.list_branch | def list_branch(self, repo_name):
"""
Lists the active Branch objects on a Repo.
Params:
* repo_name: The name of the repo.
"""
req = proto.ListBranchRequest(repo=proto.Repo(name=repo_name))
res = self.stub.ListBranch(req, metadata=self.metadata)
if hasat... | python | def list_branch(self, repo_name):
"""
Lists the active Branch objects on a Repo.
Params:
* repo_name: The name of the repo.
"""
req = proto.ListBranchRequest(repo=proto.Repo(name=repo_name))
res = self.stub.ListBranch(req, metadata=self.metadata)
if hasat... | [
"def",
"list_branch",
"(",
"self",
",",
"repo_name",
")",
":",
"req",
"=",
"proto",
".",
"ListBranchRequest",
"(",
"repo",
"=",
"proto",
".",
"Repo",
"(",
"name",
"=",
"repo_name",
")",
")",
"res",
"=",
"self",
".",
"stub",
".",
"ListBranch",
"(",
"r... | Lists the active Branch objects on a Repo.
Params:
* repo_name: The name of the repo. | [
"Lists",
"the",
"active",
"Branch",
"objects",
"on",
"a",
"Repo",
"."
] | 1c58cf91d30e03716a4f45213989e890f7b8a78c | https://github.com/pachyderm/python-pachyderm/blob/1c58cf91d30e03716a4f45213989e890f7b8a78c/src/python_pachyderm/pfs_client.py#L253-L264 |
236,981 | pachyderm/python-pachyderm | src/python_pachyderm/pfs_client.py | PfsClient.set_branch | def set_branch(self, commit, branch_name):
"""
Sets a commit and its ancestors as a branch.
Params:
* commit: A tuple, string, or Commit object representing the commit.
* branch_name: The name for the branch to set.
"""
res = proto.SetBranchRequest(commit=commit_... | python | def set_branch(self, commit, branch_name):
"""
Sets a commit and its ancestors as a branch.
Params:
* commit: A tuple, string, or Commit object representing the commit.
* branch_name: The name for the branch to set.
"""
res = proto.SetBranchRequest(commit=commit_... | [
"def",
"set_branch",
"(",
"self",
",",
"commit",
",",
"branch_name",
")",
":",
"res",
"=",
"proto",
".",
"SetBranchRequest",
"(",
"commit",
"=",
"commit_from",
"(",
"commit",
")",
",",
"branch",
"=",
"branch_name",
")",
"self",
".",
"stub",
".",
"SetBran... | Sets a commit and its ancestors as a branch.
Params:
* commit: A tuple, string, or Commit object representing the commit.
* branch_name: The name for the branch to set. | [
"Sets",
"a",
"commit",
"and",
"its",
"ancestors",
"as",
"a",
"branch",
"."
] | 1c58cf91d30e03716a4f45213989e890f7b8a78c | https://github.com/pachyderm/python-pachyderm/blob/1c58cf91d30e03716a4f45213989e890f7b8a78c/src/python_pachyderm/pfs_client.py#L266-L275 |
236,982 | pachyderm/python-pachyderm | src/python_pachyderm/pfs_client.py | PfsClient.delete_branch | def delete_branch(self, repo_name, branch_name):
"""
Deletes a branch, but leaves the commits themselves intact. In other
words, those commits can still be accessed via commit IDs and other
branches they happen to be on.
Params:
* repo_name: The name of the repo.
... | python | def delete_branch(self, repo_name, branch_name):
"""
Deletes a branch, but leaves the commits themselves intact. In other
words, those commits can still be accessed via commit IDs and other
branches they happen to be on.
Params:
* repo_name: The name of the repo.
... | [
"def",
"delete_branch",
"(",
"self",
",",
"repo_name",
",",
"branch_name",
")",
":",
"res",
"=",
"proto",
".",
"DeleteBranchRequest",
"(",
"repo",
"=",
"Repo",
"(",
"name",
"=",
"repo_name",
")",
",",
"branch",
"=",
"branch_name",
")",
"self",
".",
"stub... | Deletes a branch, but leaves the commits themselves intact. In other
words, those commits can still be accessed via commit IDs and other
branches they happen to be on.
Params:
* repo_name: The name of the repo.
* branch_name: The name of the branch to delete. | [
"Deletes",
"a",
"branch",
"but",
"leaves",
"the",
"commits",
"themselves",
"intact",
".",
"In",
"other",
"words",
"those",
"commits",
"can",
"still",
"be",
"accessed",
"via",
"commit",
"IDs",
"and",
"other",
"branches",
"they",
"happen",
"to",
"be",
"on",
... | 1c58cf91d30e03716a4f45213989e890f7b8a78c | https://github.com/pachyderm/python-pachyderm/blob/1c58cf91d30e03716a4f45213989e890f7b8a78c/src/python_pachyderm/pfs_client.py#L277-L288 |
236,983 | pachyderm/python-pachyderm | src/python_pachyderm/pfs_client.py | PfsClient.put_file_url | def put_file_url(self, commit, path, url, recursive=False):
"""
Puts a file using the content found at a URL. The URL is sent to the
server which performs the request.
Params:
* commit: A tuple, string, or Commit object representing the commit.
* path: The path to the fi... | python | def put_file_url(self, commit, path, url, recursive=False):
"""
Puts a file using the content found at a URL. The URL is sent to the
server which performs the request.
Params:
* commit: A tuple, string, or Commit object representing the commit.
* path: The path to the fi... | [
"def",
"put_file_url",
"(",
"self",
",",
"commit",
",",
"path",
",",
"url",
",",
"recursive",
"=",
"False",
")",
":",
"req",
"=",
"iter",
"(",
"[",
"proto",
".",
"PutFileRequest",
"(",
"file",
"=",
"proto",
".",
"File",
"(",
"commit",
"=",
"commit_fr... | Puts a file using the content found at a URL. The URL is sent to the
server which performs the request.
Params:
* commit: A tuple, string, or Commit object representing the commit.
* path: The path to the file.
* url: The url of the file to put.
* recursive: allow for re... | [
"Puts",
"a",
"file",
"using",
"the",
"content",
"found",
"at",
"a",
"URL",
".",
"The",
"URL",
"is",
"sent",
"to",
"the",
"server",
"which",
"performs",
"the",
"request",
"."
] | 1c58cf91d30e03716a4f45213989e890f7b8a78c | https://github.com/pachyderm/python-pachyderm/blob/1c58cf91d30e03716a4f45213989e890f7b8a78c/src/python_pachyderm/pfs_client.py#L363-L382 |
236,984 | pachyderm/python-pachyderm | src/python_pachyderm/pfs_client.py | PfsClient.get_file | def get_file(self, commit, path, offset_bytes=0, size_bytes=0, extract_value=True):
"""
Returns an iterator of the contents contents of a file at a specific Commit.
Params:
* commit: A tuple, string, or Commit object representing the commit.
* path: The path of the file.
... | python | def get_file(self, commit, path, offset_bytes=0, size_bytes=0, extract_value=True):
"""
Returns an iterator of the contents contents of a file at a specific Commit.
Params:
* commit: A tuple, string, or Commit object representing the commit.
* path: The path of the file.
... | [
"def",
"get_file",
"(",
"self",
",",
"commit",
",",
"path",
",",
"offset_bytes",
"=",
"0",
",",
"size_bytes",
"=",
"0",
",",
"extract_value",
"=",
"True",
")",
":",
"req",
"=",
"proto",
".",
"GetFileRequest",
"(",
"file",
"=",
"proto",
".",
"File",
"... | Returns an iterator of the contents contents of a file at a specific Commit.
Params:
* commit: A tuple, string, or Commit object representing the commit.
* path: The path of the file.
* offset_bytes: Optional. specifies a number of bytes that should be
skipped in the beginning o... | [
"Returns",
"an",
"iterator",
"of",
"the",
"contents",
"contents",
"of",
"a",
"file",
"at",
"a",
"specific",
"Commit",
"."
] | 1c58cf91d30e03716a4f45213989e890f7b8a78c | https://github.com/pachyderm/python-pachyderm/blob/1c58cf91d30e03716a4f45213989e890f7b8a78c/src/python_pachyderm/pfs_client.py#L384-L409 |
236,985 | pachyderm/python-pachyderm | src/python_pachyderm/pfs_client.py | PfsClient.get_files | def get_files(self, commit, paths, recursive=False):
"""
Returns the contents of a list of files at a specific Commit as a
dictionary of file paths to data.
Params:
* commit: A tuple, string, or Commit object representing the commit.
* paths: A list of paths to retrieve.... | python | def get_files(self, commit, paths, recursive=False):
"""
Returns the contents of a list of files at a specific Commit as a
dictionary of file paths to data.
Params:
* commit: A tuple, string, or Commit object representing the commit.
* paths: A list of paths to retrieve.... | [
"def",
"get_files",
"(",
"self",
",",
"commit",
",",
"paths",
",",
"recursive",
"=",
"False",
")",
":",
"filtered_file_infos",
"=",
"[",
"]",
"for",
"path",
"in",
"paths",
":",
"fi",
"=",
"self",
".",
"inspect_file",
"(",
"commit",
",",
"path",
")",
... | Returns the contents of a list of files at a specific Commit as a
dictionary of file paths to data.
Params:
* commit: A tuple, string, or Commit object representing the commit.
* paths: A list of paths to retrieve.
* recursive: If True, will go into each directory in the list
... | [
"Returns",
"the",
"contents",
"of",
"a",
"list",
"of",
"files",
"at",
"a",
"specific",
"Commit",
"as",
"a",
"dictionary",
"of",
"file",
"paths",
"to",
"data",
"."
] | 1c58cf91d30e03716a4f45213989e890f7b8a78c | https://github.com/pachyderm/python-pachyderm/blob/1c58cf91d30e03716a4f45213989e890f7b8a78c/src/python_pachyderm/pfs_client.py#L411-L432 |
236,986 | pachyderm/python-pachyderm | src/python_pachyderm/pfs_client.py | PfsClient.inspect_file | def inspect_file(self, commit, path):
"""
Returns info about a specific file.
Params:
* commit: A tuple, string, or Commit object representing the commit.
* path: Path to file.
"""
req = proto.InspectFileRequest(file=proto.File(commit=commit_from(commit), path=pa... | python | def inspect_file(self, commit, path):
"""
Returns info about a specific file.
Params:
* commit: A tuple, string, or Commit object representing the commit.
* path: Path to file.
"""
req = proto.InspectFileRequest(file=proto.File(commit=commit_from(commit), path=pa... | [
"def",
"inspect_file",
"(",
"self",
",",
"commit",
",",
"path",
")",
":",
"req",
"=",
"proto",
".",
"InspectFileRequest",
"(",
"file",
"=",
"proto",
".",
"File",
"(",
"commit",
"=",
"commit_from",
"(",
"commit",
")",
",",
"path",
"=",
"path",
")",
")... | Returns info about a specific file.
Params:
* commit: A tuple, string, or Commit object representing the commit.
* path: Path to file. | [
"Returns",
"info",
"about",
"a",
"specific",
"file",
"."
] | 1c58cf91d30e03716a4f45213989e890f7b8a78c | https://github.com/pachyderm/python-pachyderm/blob/1c58cf91d30e03716a4f45213989e890f7b8a78c/src/python_pachyderm/pfs_client.py#L434-L444 |
236,987 | pachyderm/python-pachyderm | src/python_pachyderm/pfs_client.py | PfsClient.list_file | def list_file(self, commit, path, recursive=False):
"""
Lists the files in a directory.
Params:
* commit: A tuple, string, or Commit object representing the commit.
* path: The path to the directory.
* recursive: If True, continue listing the files for sub-directories.
... | python | def list_file(self, commit, path, recursive=False):
"""
Lists the files in a directory.
Params:
* commit: A tuple, string, or Commit object representing the commit.
* path: The path to the directory.
* recursive: If True, continue listing the files for sub-directories.
... | [
"def",
"list_file",
"(",
"self",
",",
"commit",
",",
"path",
",",
"recursive",
"=",
"False",
")",
":",
"req",
"=",
"proto",
".",
"ListFileRequest",
"(",
"file",
"=",
"proto",
".",
"File",
"(",
"commit",
"=",
"commit_from",
"(",
"commit",
")",
",",
"p... | Lists the files in a directory.
Params:
* commit: A tuple, string, or Commit object representing the commit.
* path: The path to the directory.
* recursive: If True, continue listing the files for sub-directories. | [
"Lists",
"the",
"files",
"in",
"a",
"directory",
"."
] | 1c58cf91d30e03716a4f45213989e890f7b8a78c | https://github.com/pachyderm/python-pachyderm/blob/1c58cf91d30e03716a4f45213989e890f7b8a78c/src/python_pachyderm/pfs_client.py#L446-L466 |
236,988 | pachyderm/python-pachyderm | src/python_pachyderm/pfs_client.py | PfsClient.delete_file | def delete_file(self, commit, path):
"""
Deletes a file from a Commit. DeleteFile leaves a tombstone in the
Commit, assuming the file isn't written to later attempting to get the
file from the finished commit will result in not found error. The file
will of course remain intact i... | python | def delete_file(self, commit, path):
"""
Deletes a file from a Commit. DeleteFile leaves a tombstone in the
Commit, assuming the file isn't written to later attempting to get the
file from the finished commit will result in not found error. The file
will of course remain intact i... | [
"def",
"delete_file",
"(",
"self",
",",
"commit",
",",
"path",
")",
":",
"req",
"=",
"proto",
".",
"DeleteFileRequest",
"(",
"file",
"=",
"proto",
".",
"File",
"(",
"commit",
"=",
"commit_from",
"(",
"commit",
")",
",",
"path",
"=",
"path",
")",
")",... | Deletes a file from a Commit. DeleteFile leaves a tombstone in the
Commit, assuming the file isn't written to later attempting to get the
file from the finished commit will result in not found error. The file
will of course remain intact in the Commit's parent.
Params:
* commit:... | [
"Deletes",
"a",
"file",
"from",
"a",
"Commit",
".",
"DeleteFile",
"leaves",
"a",
"tombstone",
"in",
"the",
"Commit",
"assuming",
"the",
"file",
"isn",
"t",
"written",
"to",
"later",
"attempting",
"to",
"get",
"the",
"file",
"from",
"the",
"finished",
"comm... | 1c58cf91d30e03716a4f45213989e890f7b8a78c | https://github.com/pachyderm/python-pachyderm/blob/1c58cf91d30e03716a4f45213989e890f7b8a78c/src/python_pachyderm/pfs_client.py#L475-L487 |
236,989 | IdentityPython/SATOSA | src/satosa/frontends/saml2.py | SAMLFrontend.handle_authn_request | def handle_authn_request(self, context, binding_in):
"""
This method is bound to the starting endpoint of the authentication.
:type context: satosa.context.Context
:type binding_in: str
:rtype: satosa.response.Response
:param context: The current context
:param ... | python | def handle_authn_request(self, context, binding_in):
"""
This method is bound to the starting endpoint of the authentication.
:type context: satosa.context.Context
:type binding_in: str
:rtype: satosa.response.Response
:param context: The current context
:param ... | [
"def",
"handle_authn_request",
"(",
"self",
",",
"context",
",",
"binding_in",
")",
":",
"return",
"self",
".",
"_handle_authn_request",
"(",
"context",
",",
"binding_in",
",",
"self",
".",
"idp",
")"
] | This method is bound to the starting endpoint of the authentication.
:type context: satosa.context.Context
:type binding_in: str
:rtype: satosa.response.Response
:param context: The current context
:param binding_in: The binding type (http post, http redirect, ...)
:ret... | [
"This",
"method",
"is",
"bound",
"to",
"the",
"starting",
"endpoint",
"of",
"the",
"authentication",
"."
] | 49da5d4c0ac1a5ebf1a71b4f7aaf04f0e52d8fdb | https://github.com/IdentityPython/SATOSA/blob/49da5d4c0ac1a5ebf1a71b4f7aaf04f0e52d8fdb/src/satosa/frontends/saml2.py#L90-L102 |
236,990 | IdentityPython/SATOSA | src/satosa/frontends/saml2.py | SAMLFrontend._create_state_data | def _create_state_data(self, context, resp_args, relay_state):
"""
Returns a dict containing the state needed in the response flow.
:type context: satosa.context.Context
:type resp_args: dict[str, str | saml2.samlp.NameIDPolicy]
:type relay_state: str
:rtype: dict[str, d... | python | def _create_state_data(self, context, resp_args, relay_state):
"""
Returns a dict containing the state needed in the response flow.
:type context: satosa.context.Context
:type resp_args: dict[str, str | saml2.samlp.NameIDPolicy]
:type relay_state: str
:rtype: dict[str, d... | [
"def",
"_create_state_data",
"(",
"self",
",",
"context",
",",
"resp_args",
",",
"relay_state",
")",
":",
"if",
"\"name_id_policy\"",
"in",
"resp_args",
"and",
"resp_args",
"[",
"\"name_id_policy\"",
"]",
"is",
"not",
"None",
":",
"resp_args",
"[",
"\"name_id_po... | Returns a dict containing the state needed in the response flow.
:type context: satosa.context.Context
:type resp_args: dict[str, str | saml2.samlp.NameIDPolicy]
:type relay_state: str
:rtype: dict[str, dict[str, str] | str]
:param context: The current context
:param re... | [
"Returns",
"a",
"dict",
"containing",
"the",
"state",
"needed",
"in",
"the",
"response",
"flow",
"."
] | 49da5d4c0ac1a5ebf1a71b4f7aaf04f0e52d8fdb | https://github.com/IdentityPython/SATOSA/blob/49da5d4c0ac1a5ebf1a71b4f7aaf04f0e52d8fdb/src/satosa/frontends/saml2.py#L125-L141 |
236,991 | IdentityPython/SATOSA | src/satosa/frontends/saml2.py | SAMLFrontend._handle_authn_request | def _handle_authn_request(self, context, binding_in, idp):
"""
See doc for handle_authn_request method.
:type context: satosa.context.Context
:type binding_in: str
:type idp: saml.server.Server
:rtype: satosa.response.Response
:param context: The current context... | python | def _handle_authn_request(self, context, binding_in, idp):
"""
See doc for handle_authn_request method.
:type context: satosa.context.Context
:type binding_in: str
:type idp: saml.server.Server
:rtype: satosa.response.Response
:param context: The current context... | [
"def",
"_handle_authn_request",
"(",
"self",
",",
"context",
",",
"binding_in",
",",
"idp",
")",
":",
"req_info",
"=",
"idp",
".",
"parse_authn_request",
"(",
"context",
".",
"request",
"[",
"\"SAMLRequest\"",
"]",
",",
"binding_in",
")",
"authn_req",
"=",
"... | See doc for handle_authn_request method.
:type context: satosa.context.Context
:type binding_in: str
:type idp: saml.server.Server
:rtype: satosa.response.Response
:param context: The current context
:param binding_in: The pysaml binding type
:param idp: The sam... | [
"See",
"doc",
"for",
"handle_authn_request",
"method",
"."
] | 49da5d4c0ac1a5ebf1a71b4f7aaf04f0e52d8fdb | https://github.com/IdentityPython/SATOSA/blob/49da5d4c0ac1a5ebf1a71b4f7aaf04f0e52d8fdb/src/satosa/frontends/saml2.py#L177-L245 |
236,992 | IdentityPython/SATOSA | src/satosa/frontends/saml2.py | SAMLFrontend._get_approved_attributes | def _get_approved_attributes(self, idp, idp_policy, sp_entity_id, state):
"""
Returns a list of approved attributes
:type idp: saml.server.Server
:type idp_policy: saml2.assertion.Policy
:type sp_entity_id: str
:type state: satosa.state.State
:rtype: list[str]
... | python | def _get_approved_attributes(self, idp, idp_policy, sp_entity_id, state):
"""
Returns a list of approved attributes
:type idp: saml.server.Server
:type idp_policy: saml2.assertion.Policy
:type sp_entity_id: str
:type state: satosa.state.State
:rtype: list[str]
... | [
"def",
"_get_approved_attributes",
"(",
"self",
",",
"idp",
",",
"idp_policy",
",",
"sp_entity_id",
",",
"state",
")",
":",
"name_format",
"=",
"idp_policy",
".",
"get_name_form",
"(",
"sp_entity_id",
")",
"attrconvs",
"=",
"idp",
".",
"config",
".",
"attribut... | Returns a list of approved attributes
:type idp: saml.server.Server
:type idp_policy: saml2.assertion.Policy
:type sp_entity_id: str
:type state: satosa.state.State
:rtype: list[str]
:param idp: The saml frontend idp server
:param idp_policy: The idp policy
... | [
"Returns",
"a",
"list",
"of",
"approved",
"attributes"
] | 49da5d4c0ac1a5ebf1a71b4f7aaf04f0e52d8fdb | https://github.com/IdentityPython/SATOSA/blob/49da5d4c0ac1a5ebf1a71b4f7aaf04f0e52d8fdb/src/satosa/frontends/saml2.py#L247-L274 |
236,993 | IdentityPython/SATOSA | src/satosa/frontends/saml2.py | SAMLFrontend._build_idp_config_endpoints | def _build_idp_config_endpoints(self, config, providers):
"""
Builds the final frontend module config
:type config: dict[str, Any]
:type providers: list[str]
:rtype: dict[str, Any]
:param config: The module config
:param providers: A list of backend names
... | python | def _build_idp_config_endpoints(self, config, providers):
"""
Builds the final frontend module config
:type config: dict[str, Any]
:type providers: list[str]
:rtype: dict[str, Any]
:param config: The module config
:param providers: A list of backend names
... | [
"def",
"_build_idp_config_endpoints",
"(",
"self",
",",
"config",
",",
"providers",
")",
":",
"# Add an endpoint to each provider",
"idp_endpoints",
"=",
"[",
"]",
"for",
"endp_category",
"in",
"self",
".",
"endpoints",
":",
"for",
"func",
",",
"endpoint",
"in",
... | Builds the final frontend module config
:type config: dict[str, Any]
:type providers: list[str]
:rtype: dict[str, Any]
:param config: The module config
:param providers: A list of backend names
:return: The final config | [
"Builds",
"the",
"final",
"frontend",
"module",
"config"
] | 49da5d4c0ac1a5ebf1a71b4f7aaf04f0e52d8fdb | https://github.com/IdentityPython/SATOSA/blob/49da5d4c0ac1a5ebf1a71b4f7aaf04f0e52d8fdb/src/satosa/frontends/saml2.py#L522-L544 |
236,994 | IdentityPython/SATOSA | src/satosa/frontends/saml2.py | SAMLMirrorFrontend._load_endpoints_to_config | def _load_endpoints_to_config(self, provider, target_entity_id, config=None):
"""
Loads approved endpoints to the config.
:type url_base: str
:type provider: str
:type target_entity_id: str
:rtype: dict[str, Any]
:param url_base: The proxy base url
:para... | python | def _load_endpoints_to_config(self, provider, target_entity_id, config=None):
"""
Loads approved endpoints to the config.
:type url_base: str
:type provider: str
:type target_entity_id: str
:rtype: dict[str, Any]
:param url_base: The proxy base url
:para... | [
"def",
"_load_endpoints_to_config",
"(",
"self",
",",
"provider",
",",
"target_entity_id",
",",
"config",
"=",
"None",
")",
":",
"idp_conf",
"=",
"copy",
".",
"deepcopy",
"(",
"config",
"or",
"self",
".",
"idp_config",
")",
"for",
"service",
",",
"endpoint",... | Loads approved endpoints to the config.
:type url_base: str
:type provider: str
:type target_entity_id: str
:rtype: dict[str, Any]
:param url_base: The proxy base url
:param provider: target backend name
:param target_entity_id: frontend target entity id
... | [
"Loads",
"approved",
"endpoints",
"to",
"the",
"config",
"."
] | 49da5d4c0ac1a5ebf1a71b4f7aaf04f0e52d8fdb | https://github.com/IdentityPython/SATOSA/blob/49da5d4c0ac1a5ebf1a71b4f7aaf04f0e52d8fdb/src/satosa/frontends/saml2.py#L564-L587 |
236,995 | IdentityPython/SATOSA | src/satosa/frontends/saml2.py | SAMLMirrorFrontend._load_idp_dynamic_entity_id | def _load_idp_dynamic_entity_id(self, state):
"""
Loads an idp server with the entity id saved in state
:type state: satosa.state.State
:rtype: saml.server.Server
:param state: The current state
:return: An idp server
"""
# Change the idp entity id dynam... | python | def _load_idp_dynamic_entity_id(self, state):
"""
Loads an idp server with the entity id saved in state
:type state: satosa.state.State
:rtype: saml.server.Server
:param state: The current state
:return: An idp server
"""
# Change the idp entity id dynam... | [
"def",
"_load_idp_dynamic_entity_id",
"(",
"self",
",",
"state",
")",
":",
"# Change the idp entity id dynamically",
"idp_config_file",
"=",
"copy",
".",
"deepcopy",
"(",
"self",
".",
"idp_config",
")",
"idp_config_file",
"[",
"\"entityid\"",
"]",
"=",
"\"{}/{}\"",
... | Loads an idp server with the entity id saved in state
:type state: satosa.state.State
:rtype: saml.server.Server
:param state: The current state
:return: An idp server | [
"Loads",
"an",
"idp",
"server",
"with",
"the",
"entity",
"id",
"saved",
"in",
"state"
] | 49da5d4c0ac1a5ebf1a71b4f7aaf04f0e52d8fdb | https://github.com/IdentityPython/SATOSA/blob/49da5d4c0ac1a5ebf1a71b4f7aaf04f0e52d8fdb/src/satosa/frontends/saml2.py#L605-L619 |
236,996 | IdentityPython/SATOSA | src/satosa/frontends/saml2.py | SAMLVirtualCoFrontend._get_co_name_from_path | def _get_co_name_from_path(self, context):
"""
The CO name is URL encoded and obtained from the request path
for a request coming into one of the standard binding endpoints.
For example the HTTP-Redirect binding request path will have the
format
{base}/{backend}/{co_name... | python | def _get_co_name_from_path(self, context):
"""
The CO name is URL encoded and obtained from the request path
for a request coming into one of the standard binding endpoints.
For example the HTTP-Redirect binding request path will have the
format
{base}/{backend}/{co_name... | [
"def",
"_get_co_name_from_path",
"(",
"self",
",",
"context",
")",
":",
"url_encoded_co_name",
"=",
"context",
".",
"path",
".",
"split",
"(",
"\"/\"",
")",
"[",
"1",
"]",
"co_name",
"=",
"unquote_plus",
"(",
"url_encoded_co_name",
")",
"return",
"co_name"
] | The CO name is URL encoded and obtained from the request path
for a request coming into one of the standard binding endpoints.
For example the HTTP-Redirect binding request path will have the
format
{base}/{backend}/{co_name}/sso/redirect
:type context: satosa.context.Context
... | [
"The",
"CO",
"name",
"is",
"URL",
"encoded",
"and",
"obtained",
"from",
"the",
"request",
"path",
"for",
"a",
"request",
"coming",
"into",
"one",
"of",
"the",
"standard",
"binding",
"endpoints",
".",
"For",
"example",
"the",
"HTTP",
"-",
"Redirect",
"bindi... | 49da5d4c0ac1a5ebf1a71b4f7aaf04f0e52d8fdb | https://github.com/IdentityPython/SATOSA/blob/49da5d4c0ac1a5ebf1a71b4f7aaf04f0e52d8fdb/src/satosa/frontends/saml2.py#L750-L768 |
236,997 | IdentityPython/SATOSA | src/satosa/frontends/saml2.py | SAMLVirtualCoFrontend._get_co_name | def _get_co_name(self, context):
"""
Obtain the CO name previously saved in the request state, or if not set
use the request path obtained from the current context to determine
the target CO.
:type context: The current context
:rtype: string
:param context: The ... | python | def _get_co_name(self, context):
"""
Obtain the CO name previously saved in the request state, or if not set
use the request path obtained from the current context to determine
the target CO.
:type context: The current context
:rtype: string
:param context: The ... | [
"def",
"_get_co_name",
"(",
"self",
",",
"context",
")",
":",
"try",
":",
"co_name",
"=",
"context",
".",
"state",
"[",
"self",
".",
"name",
"]",
"[",
"self",
".",
"KEY_CO_NAME",
"]",
"logger",
".",
"debug",
"(",
"\"Found CO {} from state\"",
".",
"forma... | Obtain the CO name previously saved in the request state, or if not set
use the request path obtained from the current context to determine
the target CO.
:type context: The current context
:rtype: string
:param context: The current context
:return: CO name | [
"Obtain",
"the",
"CO",
"name",
"previously",
"saved",
"in",
"the",
"request",
"state",
"or",
"if",
"not",
"set",
"use",
"the",
"request",
"path",
"obtained",
"from",
"the",
"current",
"context",
"to",
"determine",
"the",
"target",
"CO",
"."
] | 49da5d4c0ac1a5ebf1a71b4f7aaf04f0e52d8fdb | https://github.com/IdentityPython/SATOSA/blob/49da5d4c0ac1a5ebf1a71b4f7aaf04f0e52d8fdb/src/satosa/frontends/saml2.py#L770-L789 |
236,998 | IdentityPython/SATOSA | src/satosa/frontends/saml2.py | SAMLVirtualCoFrontend._add_endpoints_to_config | def _add_endpoints_to_config(self, config, co_name, backend_name):
"""
Use the request path from the context to determine the target backend,
then construct mappings from bindings to endpoints for the virtual
IdP for the CO.
The endpoint URLs have the form
{base}/{backe... | python | def _add_endpoints_to_config(self, config, co_name, backend_name):
"""
Use the request path from the context to determine the target backend,
then construct mappings from bindings to endpoints for the virtual
IdP for the CO.
The endpoint URLs have the form
{base}/{backe... | [
"def",
"_add_endpoints_to_config",
"(",
"self",
",",
"config",
",",
"co_name",
",",
"backend_name",
")",
":",
"for",
"service",
",",
"endpoint",
"in",
"self",
".",
"endpoints",
".",
"items",
"(",
")",
":",
"idp_endpoints",
"=",
"[",
"]",
"for",
"binding",
... | Use the request path from the context to determine the target backend,
then construct mappings from bindings to endpoints for the virtual
IdP for the CO.
The endpoint URLs have the form
{base}/{backend}/{co_name}/{path}
:type config: satosa.satosa_config.SATOSAConfig
:... | [
"Use",
"the",
"request",
"path",
"from",
"the",
"context",
"to",
"determine",
"the",
"target",
"backend",
"then",
"construct",
"mappings",
"from",
"bindings",
"to",
"endpoints",
"for",
"the",
"virtual",
"IdP",
"for",
"the",
"CO",
"."
] | 49da5d4c0ac1a5ebf1a71b4f7aaf04f0e52d8fdb | https://github.com/IdentityPython/SATOSA/blob/49da5d4c0ac1a5ebf1a71b4f7aaf04f0e52d8fdb/src/satosa/frontends/saml2.py#L791-L828 |
236,999 | IdentityPython/SATOSA | src/satosa/frontends/saml2.py | SAMLVirtualCoFrontend._add_entity_id | def _add_entity_id(self, config, co_name):
"""
Use the CO name to construct the entity ID for the virtual IdP
for the CO.
The entity ID has the form
{base_entity_id}/{co_name}
:type config: satosa.satosa_config.SATOSAConfig
:type co_name: str
:rtype: sa... | python | def _add_entity_id(self, config, co_name):
"""
Use the CO name to construct the entity ID for the virtual IdP
for the CO.
The entity ID has the form
{base_entity_id}/{co_name}
:type config: satosa.satosa_config.SATOSAConfig
:type co_name: str
:rtype: sa... | [
"def",
"_add_entity_id",
"(",
"self",
",",
"config",
",",
"co_name",
")",
":",
"base_entity_id",
"=",
"config",
"[",
"'entityid'",
"]",
"co_entity_id",
"=",
"\"{}/{}\"",
".",
"format",
"(",
"base_entity_id",
",",
"quote_plus",
"(",
"co_name",
")",
")",
"conf... | Use the CO name to construct the entity ID for the virtual IdP
for the CO.
The entity ID has the form
{base_entity_id}/{co_name}
:type config: satosa.satosa_config.SATOSAConfig
:type co_name: str
:rtype: satosa.satosa_config.SATOSAConfig
:param config: satosa ... | [
"Use",
"the",
"CO",
"name",
"to",
"construct",
"the",
"entity",
"ID",
"for",
"the",
"virtual",
"IdP",
"for",
"the",
"CO",
"."
] | 49da5d4c0ac1a5ebf1a71b4f7aaf04f0e52d8fdb | https://github.com/IdentityPython/SATOSA/blob/49da5d4c0ac1a5ebf1a71b4f7aaf04f0e52d8fdb/src/satosa/frontends/saml2.py#L830-L852 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.