body stringlengths 26 98.2k | body_hash int64 -9,222,864,604,528,158,000 9,221,803,474B | docstring stringlengths 1 16.8k | path stringlengths 5 230 | name stringlengths 1 96 | repository_name stringlengths 7 89 | lang stringclasses 1
value | body_without_docstring stringlengths 20 98.2k |
|---|---|---|---|---|---|---|---|
def blackbody(nu, ref_freq=353.0):
'\n The ratio of the blackbody function for dust at frequency nu\n over the value for reference frequency ref_freq\n\n Arguments\n ---------\n nu : float\n Frequency in GHz.\n ref_freq : float\n Reference frequency in GHz.\n\n Returns\n ------... | -3,061,004,010,963,920,400 | The ratio of the blackbody function for dust at frequency nu
over the value for reference frequency ref_freq
Arguments
---------
nu : float
Frequency in GHz.
ref_freq : float
Reference frequency in GHz.
Returns
-------
blackbody_ratio : float
B(nu, T_dust) / B(nu_ref, T_dust) | xfaster/spec_tools.py | blackbody | SPIDER-CMB/xfaster | python | def blackbody(nu, ref_freq=353.0):
'\n The ratio of the blackbody function for dust at frequency nu\n over the value for reference frequency ref_freq\n\n Arguments\n ---------\n nu : float\n Frequency in GHz.\n ref_freq : float\n Reference frequency in GHz.\n\n Returns\n ------... |
def rj2cmb(nu_in):
'\n Conversion from Rayleigh-Jeans units to CMB temperature units\n\n Arguments\n ---------\n nu_in : float\n Frequency in GHz.\n\n Returns\n -------\n cal_fac : float\n Number by which to multiply a RJ temperature to get a CMB temp\n '
k = 1.38064852e-23... | 2,812,329,766,020,827,000 | Conversion from Rayleigh-Jeans units to CMB temperature units
Arguments
---------
nu_in : float
Frequency in GHz.
Returns
-------
cal_fac : float
Number by which to multiply a RJ temperature to get a CMB temp | xfaster/spec_tools.py | rj2cmb | SPIDER-CMB/xfaster | python | def rj2cmb(nu_in):
'\n Conversion from Rayleigh-Jeans units to CMB temperature units\n\n Arguments\n ---------\n nu_in : float\n Frequency in GHz.\n\n Returns\n -------\n cal_fac : float\n Number by which to multiply a RJ temperature to get a CMB temp\n '
k = 1.38064852e-23... |
def scale_dust(freq0, freq1, ref_freq, beta, delta_beta=None, deriv=False):
'\n Get the factor by which you must multiply the cross spectrum from maps of\n frequencies freq0 and freq1 to match the dust power at ref_freq given\n spectra index beta.\n\n If deriv is True, return the frequency scaling at th... | 8,050,293,239,967,490,000 | Get the factor by which you must multiply the cross spectrum from maps of
frequencies freq0 and freq1 to match the dust power at ref_freq given
spectra index beta.
If deriv is True, return the frequency scaling at the reference beta,
and the first derivative w.r.t. beta.
Otherwise if delta_beta is given, return the s... | xfaster/spec_tools.py | scale_dust | SPIDER-CMB/xfaster | python | def scale_dust(freq0, freq1, ref_freq, beta, delta_beta=None, deriv=False):
'\n Get the factor by which you must multiply the cross spectrum from maps of\n frequencies freq0 and freq1 to match the dust power at ref_freq given\n spectra index beta.\n\n If deriv is True, return the frequency scaling at th... |
def wigner3j(l2, m2, l3, m3):
'\n Wigner 3j symbols computed for all valid values of ``L``, as in:\n\n .. math::\n\n \\begin{pmatrix}\n \\ell_2 & \\ell_3 & L \\\\\n m_2 & m_3 & 0 \\\\\n \\end{pmatrix}\n\n Arguments\n ---------\n l2, m2, l3, m3 : int\n The ell and ... | -2,767,139,856,052,830,000 | Wigner 3j symbols computed for all valid values of ``L``, as in:
.. math::
\begin{pmatrix}
\ell_2 & \ell_3 & L \\
m_2 & m_3 & 0 \\
\end{pmatrix}
Arguments
---------
l2, m2, l3, m3 : int
The ell and m values for which to compute the symbols.
Returns
-------
fj : array_like
Array of size ``l... | xfaster/spec_tools.py | wigner3j | SPIDER-CMB/xfaster | python | def wigner3j(l2, m2, l3, m3):
'\n Wigner 3j symbols computed for all valid values of ``L``, as in:\n\n .. math::\n\n \\begin{pmatrix}\n \\ell_2 & \\ell_3 & L \\\\\n m_2 & m_3 & 0 \\\\\n \\end{pmatrix}\n\n Arguments\n ---------\n l2, m2, l3, m3 : int\n The ell and ... |
def get_camb_cl(r, lmax, nt=None, spec='total', lfac=True):
"\n Compute camb spectrum with tensors and lensing.\n\n Parameter values are from arXiv:1807.06209 Table 1 Plik best fit\n\n Arguments\n ---------\n r : float\n Tensor-to-scalar ratio\n lmax : int\n Maximum ell for which to ... | 6,037,776,520,534,010,000 | Compute camb spectrum with tensors and lensing.
Parameter values are from arXiv:1807.06209 Table 1 Plik best fit
Arguments
---------
r : float
Tensor-to-scalar ratio
lmax : int
Maximum ell for which to compute spectra
nt : scalar, optional
Tensor spectral index. If not supplied, assumes
slow-roll con... | xfaster/spec_tools.py | get_camb_cl | SPIDER-CMB/xfaster | python | def get_camb_cl(r, lmax, nt=None, spec='total', lfac=True):
"\n Compute camb spectrum with tensors and lensing.\n\n Parameter values are from arXiv:1807.06209 Table 1 Plik best fit\n\n Arguments\n ---------\n r : float\n Tensor-to-scalar ratio\n lmax : int\n Maximum ell for which to ... |
def raise_does_not_exist(msg):
'Decorator to turn a function that get a http 404 response to a\n :exc:`DoesNotExist` exception.'
def decorator(func):
@wraps(func)
def wrapped(*args, **kwargs):
try:
return func(*args, **kwargs)
except ClientHttpError a... | -7,812,388,451,997,881,000 | Decorator to turn a function that get a http 404 response to a
:exc:`DoesNotExist` exception. | seafileapi/utils.py | raise_does_not_exist | AdriCueGim/python-seafile | python | def raise_does_not_exist(msg):
'Decorator to turn a function that get a http 404 response to a\n :exc:`DoesNotExist` exception.'
def decorator(func):
@wraps(func)
def wrapped(*args, **kwargs):
try:
return func(*args, **kwargs)
except ClientHttpError a... |
def __init__(self):
'\n :param InstanceIds: 实例ID列表,格式如:cmgo-p8vnipr5。与云数据库控制台页面中显示的实例ID相同\n :type InstanceIds: list of str\n :param ProjectId: 项目ID\n :type ProjectId: int\n '
self.InstanceIds = None
self.ProjectId = None | -6,788,372,924,002,334,000 | :param InstanceIds: 实例ID列表,格式如:cmgo-p8vnipr5。与云数据库控制台页面中显示的实例ID相同
:type InstanceIds: list of str
:param ProjectId: 项目ID
:type ProjectId: int | tencentcloud/mongodb/v20180408/models.py | __init__ | qin5506/tencentcloud-sdk-python | python | def __init__(self):
'\n :param InstanceIds: 实例ID列表,格式如:cmgo-p8vnipr5。与云数据库控制台页面中显示的实例ID相同\n :type InstanceIds: list of str\n :param ProjectId: 项目ID\n :type ProjectId: int\n '
self.InstanceIds = None
self.ProjectId = None |
def __init__(self):
'\n :param FlowIds: 返回的异步任务ID列表\n :type FlowIds: list of int non-negative\n :param RequestId: 唯一请求 ID,每次请求都会返回。定位问题时需要提供该次请求的 RequestId。\n :type RequestId: str\n '
self.FlowIds = None
self.RequestId = None | 9,070,933,027,787,485,000 | :param FlowIds: 返回的异步任务ID列表
:type FlowIds: list of int non-negative
:param RequestId: 唯一请求 ID,每次请求都会返回。定位问题时需要提供该次请求的 RequestId。
:type RequestId: str | tencentcloud/mongodb/v20180408/models.py | __init__ | qin5506/tencentcloud-sdk-python | python | def __init__(self):
'\n :param FlowIds: 返回的异步任务ID列表\n :type FlowIds: list of int non-negative\n :param RequestId: 唯一请求 ID,每次请求都会返回。定位问题时需要提供该次请求的 RequestId。\n :type RequestId: str\n '
self.FlowIds = None
self.RequestId = None |
def __init__(self):
'\n :param IP: 连接的客户端IP\n :type IP: str\n :param Count: 对应客户端IP的连接数\n :type Count: int\n '
self.IP = None
self.Count = None | 8,443,674,525,877,007,000 | :param IP: 连接的客户端IP
:type IP: str
:param Count: 对应客户端IP的连接数
:type Count: int | tencentcloud/mongodb/v20180408/models.py | __init__ | qin5506/tencentcloud-sdk-python | python | def __init__(self):
'\n :param IP: 连接的客户端IP\n :type IP: str\n :param Count: 对应客户端IP的连接数\n :type Count: int\n '
self.IP = None
self.Count = None |
def __init__(self):
'\n :param Memory: 实例内存大小,单位:GB\n :type Memory: int\n :param Volume: 实例硬盘大小,单位:GB\n :type Volume: int\n :param ReplicateSetNum: 副本集个数,1为单副本集实例,大于1为分片集群实例,最大不超过10\n :type ReplicateSetNum: int\n :param SecondaryNum: 每个副本集内从节点个数,目前只支持从节点数为2\n ... | 9,074,244,070,652,575,000 | :param Memory: 实例内存大小,单位:GB
:type Memory: int
:param Volume: 实例硬盘大小,单位:GB
:type Volume: int
:param ReplicateSetNum: 副本集个数,1为单副本集实例,大于1为分片集群实例,最大不超过10
:type ReplicateSetNum: int
:param SecondaryNum: 每个副本集内从节点个数,目前只支持从节点数为2
:type SecondaryNum: int
:param EngineVersion: MongoDB引擎版本,值包括MONGO_3_WT 、MONGO_3_ROCKS和MONGO_36_WT... | tencentcloud/mongodb/v20180408/models.py | __init__ | qin5506/tencentcloud-sdk-python | python | def __init__(self):
'\n :param Memory: 实例内存大小,单位:GB\n :type Memory: int\n :param Volume: 实例硬盘大小,单位:GB\n :type Volume: int\n :param ReplicateSetNum: 副本集个数,1为单副本集实例,大于1为分片集群实例,最大不超过10\n :type ReplicateSetNum: int\n :param SecondaryNum: 每个副本集内从节点个数,目前只支持从节点数为2\n ... |
def __init__(self):
'\n :param DealId: 订单ID\n :type DealId: str\n :param InstanceIds: 创建的实例ID列表\n :type InstanceIds: list of str\n :param RequestId: 唯一请求 ID,每次请求都会返回。定位问题时需要提供该次请求的 RequestId。\n :type RequestId: str\n '
self.DealId = None
self.InstanceIds = No... | 1,054,841,195,029,218,000 | :param DealId: 订单ID
:type DealId: str
:param InstanceIds: 创建的实例ID列表
:type InstanceIds: list of str
:param RequestId: 唯一请求 ID,每次请求都会返回。定位问题时需要提供该次请求的 RequestId。
:type RequestId: str | tencentcloud/mongodb/v20180408/models.py | __init__ | qin5506/tencentcloud-sdk-python | python | def __init__(self):
'\n :param DealId: 订单ID\n :type DealId: str\n :param InstanceIds: 创建的实例ID列表\n :type InstanceIds: list of str\n :param RequestId: 唯一请求 ID,每次请求都会返回。定位问题时需要提供该次请求的 RequestId。\n :type RequestId: str\n '
self.DealId = None
self.InstanceIds = No... |
def __init__(self):
'\n :param SecondaryNum: 每个副本集内从节点个数\n :type SecondaryNum: int\n :param Memory: 实例内存大小,单位:GB\n :type Memory: int\n :param Volume: 实例硬盘大小,单位:GB\n :type Volume: int\n :param MongoVersion: 版本号,当前支持 MONGO_3_WT、MONGO_3_ROCKS、MONGO_36_WT\n :type ... | 7,096,470,038,123,483,000 | :param SecondaryNum: 每个副本集内从节点个数
:type SecondaryNum: int
:param Memory: 实例内存大小,单位:GB
:type Memory: int
:param Volume: 实例硬盘大小,单位:GB
:type Volume: int
:param MongoVersion: 版本号,当前支持 MONGO_3_WT、MONGO_3_ROCKS、MONGO_36_WT
:type MongoVersion: str
:param MachineCode: 机器类型,GIO:高IO版;TGIO:高IO万兆
:type MachineCode: str
:param Goods... | tencentcloud/mongodb/v20180408/models.py | __init__ | qin5506/tencentcloud-sdk-python | python | def __init__(self):
'\n :param SecondaryNum: 每个副本集内从节点个数\n :type SecondaryNum: int\n :param Memory: 实例内存大小,单位:GB\n :type Memory: int\n :param Volume: 实例硬盘大小,单位:GB\n :type Volume: int\n :param MongoVersion: 版本号,当前支持 MONGO_3_WT、MONGO_3_ROCKS、MONGO_36_WT\n :type ... |
def __init__(self):
'\n :param DealId: 订单ID\n :type DealId: str\n :param InstanceIds: 创建的实例ID列表\n :type InstanceIds: list of str\n :param RequestId: 唯一请求 ID,每次请求都会返回。定位问题时需要提供该次请求的 RequestId。\n :type RequestId: str\n '
self.DealId = None
self.InstanceIds = No... | 1,054,841,195,029,218,000 | :param DealId: 订单ID
:type DealId: str
:param InstanceIds: 创建的实例ID列表
:type InstanceIds: list of str
:param RequestId: 唯一请求 ID,每次请求都会返回。定位问题时需要提供该次请求的 RequestId。
:type RequestId: str | tencentcloud/mongodb/v20180408/models.py | __init__ | qin5506/tencentcloud-sdk-python | python | def __init__(self):
'\n :param DealId: 订单ID\n :type DealId: str\n :param InstanceIds: 创建的实例ID列表\n :type InstanceIds: list of str\n :param RequestId: 唯一请求 ID,每次请求都会返回。定位问题时需要提供该次请求的 RequestId。\n :type RequestId: str\n '
self.DealId = None
self.InstanceIds = No... |
def __init__(self):
'\n :param InstanceId: 实例ID,格式如:cmgo-p8vnipr5。与云数据库控制台页面中显示的实例ID相同\n :type InstanceId: str\n '
self.InstanceId = None | 7,466,592,164,755,291,000 | :param InstanceId: 实例ID,格式如:cmgo-p8vnipr5。与云数据库控制台页面中显示的实例ID相同
:type InstanceId: str | tencentcloud/mongodb/v20180408/models.py | __init__ | qin5506/tencentcloud-sdk-python | python | def __init__(self):
'\n :param InstanceId: 实例ID,格式如:cmgo-p8vnipr5。与云数据库控制台页面中显示的实例ID相同\n :type InstanceId: str\n '
self.InstanceId = None |
def __init__(self):
'\n :param Clients: 客户端连接信息,包括客户端IP和对应IP的连接数量\n注意:此字段可能返回 null,表示取不到有效值。\n :type Clients: list of ClientConnection\n :param RequestId: 唯一请求 ID,每次请求都会返回。定位问题时需要提供该次请求的 RequestId。\n :type RequestId: str\n '
self.Clients = None
self.RequestId = None | 4,556,269,938,670,427,600 | :param Clients: 客户端连接信息,包括客户端IP和对应IP的连接数量
注意:此字段可能返回 null,表示取不到有效值。
:type Clients: list of ClientConnection
:param RequestId: 唯一请求 ID,每次请求都会返回。定位问题时需要提供该次请求的 RequestId。
:type RequestId: str | tencentcloud/mongodb/v20180408/models.py | __init__ | qin5506/tencentcloud-sdk-python | python | def __init__(self):
'\n :param Clients: 客户端连接信息,包括客户端IP和对应IP的连接数量\n注意:此字段可能返回 null,表示取不到有效值。\n :type Clients: list of ClientConnection\n :param RequestId: 唯一请求 ID,每次请求都会返回。定位问题时需要提供该次请求的 RequestId。\n :type RequestId: str\n '
self.Clients = None
self.RequestId = None |
def __init__(self):
'\n :param InstanceIds: 实例ID列表,格式如:cmgo-p8vnipr5。与云数据库控制台页面中显示的实例ID相同\n :type InstanceIds: list of str\n :param InstanceType: 实例类型,取值范围:0-所有实例,1-正式实例,2-临时实例, 3-只读实例,-1-正式实例+只读+灾备实例\n :type InstanceType: int\n :param ClusterType: 集群类型,取值范围:0-副本集实例,1-分片实例,-1-所有实例... | -7,437,992,389,170,799,000 | :param InstanceIds: 实例ID列表,格式如:cmgo-p8vnipr5。与云数据库控制台页面中显示的实例ID相同
:type InstanceIds: list of str
:param InstanceType: 实例类型,取值范围:0-所有实例,1-正式实例,2-临时实例, 3-只读实例,-1-正式实例+只读+灾备实例
:type InstanceType: int
:param ClusterType: 集群类型,取值范围:0-副本集实例,1-分片实例,-1-所有实例
:type ClusterType: int
:param Status: 实例状态,取值范围:0-待初始化,1-流程执行中,2-实例有效,... | tencentcloud/mongodb/v20180408/models.py | __init__ | qin5506/tencentcloud-sdk-python | python | def __init__(self):
'\n :param InstanceIds: 实例ID列表,格式如:cmgo-p8vnipr5。与云数据库控制台页面中显示的实例ID相同\n :type InstanceIds: list of str\n :param InstanceType: 实例类型,取值范围:0-所有实例,1-正式实例,2-临时实例, 3-只读实例,-1-正式实例+只读+灾备实例\n :type InstanceType: int\n :param ClusterType: 集群类型,取值范围:0-副本集实例,1-分片实例,-1-所有实例... |
def __init__(self):
'\n :param TotalCount: 符合查询条件的实例总数\n :type TotalCount: int\n :param InstanceDetails: 实例详细信息\n :type InstanceDetails: list of MongoDBInstanceDetail\n :param RequestId: 唯一请求 ID,每次请求都会返回。定位问题时需要提供该次请求的 RequestId。\n :type RequestId: str\n '
self.T... | -5,344,254,023,169,226,000 | :param TotalCount: 符合查询条件的实例总数
:type TotalCount: int
:param InstanceDetails: 实例详细信息
:type InstanceDetails: list of MongoDBInstanceDetail
:param RequestId: 唯一请求 ID,每次请求都会返回。定位问题时需要提供该次请求的 RequestId。
:type RequestId: str | tencentcloud/mongodb/v20180408/models.py | __init__ | qin5506/tencentcloud-sdk-python | python | def __init__(self):
'\n :param TotalCount: 符合查询条件的实例总数\n :type TotalCount: int\n :param InstanceDetails: 实例详细信息\n :type InstanceDetails: list of MongoDBInstanceDetail\n :param RequestId: 唯一请求 ID,每次请求都会返回。定位问题时需要提供该次请求的 RequestId。\n :type RequestId: str\n '
self.T... |
def __init__(self):
'\n :param InstanceId: 实例ID,格式如:cmgo-p8vnipr5。与云数据库控制台页面中显示的实例ID相同\n :type InstanceId: str\n :param StartTime: 慢日志起始时间,格式:yyyy-mm-dd hh:mm:ss,如:2019-06-01 10:00:00。查询起止时间间隔不能超过24小时,只允许查询最近7天内慢日志。\n :type StartTime: str\n :param EndTime: 慢日志终止时间,格式:yyyy-mm-dd hh... | 3,319,906,674,030,196,700 | :param InstanceId: 实例ID,格式如:cmgo-p8vnipr5。与云数据库控制台页面中显示的实例ID相同
:type InstanceId: str
:param StartTime: 慢日志起始时间,格式:yyyy-mm-dd hh:mm:ss,如:2019-06-01 10:00:00。查询起止时间间隔不能超过24小时,只允许查询最近7天内慢日志。
:type StartTime: str
:param EndTime: 慢日志终止时间,格式:yyyy-mm-dd hh:mm:ss,如:2019-06-02 12:00:00。查询起止时间间隔不能超过24小时,只允许查询最近7天内慢日志。
:type EndT... | tencentcloud/mongodb/v20180408/models.py | __init__ | qin5506/tencentcloud-sdk-python | python | def __init__(self):
'\n :param InstanceId: 实例ID,格式如:cmgo-p8vnipr5。与云数据库控制台页面中显示的实例ID相同\n :type InstanceId: str\n :param StartTime: 慢日志起始时间,格式:yyyy-mm-dd hh:mm:ss,如:2019-06-01 10:00:00。查询起止时间间隔不能超过24小时,只允许查询最近7天内慢日志。\n :type StartTime: str\n :param EndTime: 慢日志终止时间,格式:yyyy-mm-dd hh... |
def __init__(self):
'\n :param TotalCount: 符合查询条件的慢查询日志总数。\n :type TotalCount: int\n :param SlowLogList: 符合查询条件的慢查询日志详情。\n :type SlowLogList: list of str\n :param RequestId: 唯一请求 ID,每次请求都会返回。定位问题时需要提供该次请求的 RequestId。\n :type RequestId: str\n '
self.TotalCount = N... | -7,829,219,154,739,202,000 | :param TotalCount: 符合查询条件的慢查询日志总数。
:type TotalCount: int
:param SlowLogList: 符合查询条件的慢查询日志详情。
:type SlowLogList: list of str
:param RequestId: 唯一请求 ID,每次请求都会返回。定位问题时需要提供该次请求的 RequestId。
:type RequestId: str | tencentcloud/mongodb/v20180408/models.py | __init__ | qin5506/tencentcloud-sdk-python | python | def __init__(self):
'\n :param TotalCount: 符合查询条件的慢查询日志总数。\n :type TotalCount: int\n :param SlowLogList: 符合查询条件的慢查询日志详情。\n :type SlowLogList: list of str\n :param RequestId: 唯一请求 ID,每次请求都会返回。定位问题时需要提供该次请求的 RequestId。\n :type RequestId: str\n '
self.TotalCount = N... |
def __init__(self):
'\n :param Zone: 可用区\n :type Zone: str\n '
self.Zone = None | -5,430,548,088,093,569,000 | :param Zone: 可用区
:type Zone: str | tencentcloud/mongodb/v20180408/models.py | __init__ | qin5506/tencentcloud-sdk-python | python | def __init__(self):
'\n :param Zone: 可用区\n :type Zone: str\n '
self.Zone = None |
def __init__(self):
'\n :param SpecInfoList: 实例售卖规格信息列表\n :type SpecInfoList: list of SpecificationInfo\n :param RequestId: 唯一请求 ID,每次请求都会返回。定位问题时需要提供该次请求的 RequestId。\n :type RequestId: str\n '
self.SpecInfoList = None
self.RequestId = None | -308,205,219,353,436,860 | :param SpecInfoList: 实例售卖规格信息列表
:type SpecInfoList: list of SpecificationInfo
:param RequestId: 唯一请求 ID,每次请求都会返回。定位问题时需要提供该次请求的 RequestId。
:type RequestId: str | tencentcloud/mongodb/v20180408/models.py | __init__ | qin5506/tencentcloud-sdk-python | python | def __init__(self):
'\n :param SpecInfoList: 实例售卖规格信息列表\n :type SpecInfoList: list of SpecificationInfo\n :param RequestId: 唯一请求 ID,每次请求都会返回。定位问题时需要提供该次请求的 RequestId。\n :type RequestId: str\n '
self.SpecInfoList = None
self.RequestId = None |
def __init__(self):
'\n :param InstanceId: 实例ID\n :type InstanceId: str\n :param Region: 地域信息\n :type Region: str\n '
self.InstanceId = None
self.Region = None | -8,440,805,245,017,172,000 | :param InstanceId: 实例ID
:type InstanceId: str
:param Region: 地域信息
:type Region: str | tencentcloud/mongodb/v20180408/models.py | __init__ | qin5506/tencentcloud-sdk-python | python | def __init__(self):
'\n :param InstanceId: 实例ID\n :type InstanceId: str\n :param Region: 地域信息\n :type Region: str\n '
self.InstanceId = None
self.Region = None |
def __init__(self):
'\n :param InstanceId: 实例ID\n :type InstanceId: str\n :param InstanceName: 实例名称\n :type InstanceName: str\n :param PayMode: 付费类型,可能的返回值:1-包年包月;0-按量计费\n :type PayMode: int\n :param ProjectId: 项目ID\n :type ProjectId: int\n :param Clust... | 4,477,843,219,725,253,000 | :param InstanceId: 实例ID
:type InstanceId: str
:param InstanceName: 实例名称
:type InstanceName: str
:param PayMode: 付费类型,可能的返回值:1-包年包月;0-按量计费
:type PayMode: int
:param ProjectId: 项目ID
:type ProjectId: int
:param ClusterType: 集群类型,可能的返回值:0-副本集实例,1-分片实例,
... | tencentcloud/mongodb/v20180408/models.py | __init__ | qin5506/tencentcloud-sdk-python | python | def __init__(self):
'\n :param InstanceId: 实例ID\n :type InstanceId: str\n :param InstanceName: 实例名称\n :type InstanceName: str\n :param PayMode: 付费类型,可能的返回值:1-包年包月;0-按量计费\n :type PayMode: int\n :param ProjectId: 项目ID\n :type ProjectId: int\n :param Clust... |
def __init__(self):
'\n :param UsedVolume: 分片已使用容量\n :type UsedVolume: float\n :param ReplicaSetId: 分片ID\n :type ReplicaSetId: str\n :param ReplicaSetName: 分片名\n :type ReplicaSetName: str\n :param Memory: 分片内存规格,单位为MB\n :type Memory: int\n :param Volume... | -5,998,285,072,832,786,000 | :param UsedVolume: 分片已使用容量
:type UsedVolume: float
:param ReplicaSetId: 分片ID
:type ReplicaSetId: str
:param ReplicaSetName: 分片名
:type ReplicaSetName: str
:param Memory: 分片内存规格,单位为MB
:type Memory: int
:param Volume: 分片磁盘规格,单位为MB
:type Volume: int
:param OplogSize: 分片Oplog大小,单位为MB
:type OplogSize: int
:param SecondaryNum... | tencentcloud/mongodb/v20180408/models.py | __init__ | qin5506/tencentcloud-sdk-python | python | def __init__(self):
'\n :param UsedVolume: 分片已使用容量\n :type UsedVolume: float\n :param ReplicaSetId: 分片ID\n :type ReplicaSetId: str\n :param ReplicaSetName: 分片名\n :type ReplicaSetName: str\n :param Memory: 分片内存规格,单位为MB\n :type Memory: int\n :param Volume... |
def __init__(self):
'\n :param InstanceId: 实例ID,格式如:cmgo-p8vnipr5。与云数据库控制台页面中显示的实例ID相同\n :type InstanceId: str\n :param NewName: 实例名称\n :type NewName: str\n '
self.InstanceId = None
self.NewName = None | -1,895,953,465,478,440,000 | :param InstanceId: 实例ID,格式如:cmgo-p8vnipr5。与云数据库控制台页面中显示的实例ID相同
:type InstanceId: str
:param NewName: 实例名称
:type NewName: str | tencentcloud/mongodb/v20180408/models.py | __init__ | qin5506/tencentcloud-sdk-python | python | def __init__(self):
'\n :param InstanceId: 实例ID,格式如:cmgo-p8vnipr5。与云数据库控制台页面中显示的实例ID相同\n :type InstanceId: str\n :param NewName: 实例名称\n :type NewName: str\n '
self.InstanceId = None
self.NewName = None |
def __init__(self):
'\n :param RequestId: 唯一请求 ID,每次请求都会返回。定位问题时需要提供该次请求的 RequestId。\n :type RequestId: str\n '
self.RequestId = None | -5,957,967,262,820,529,000 | :param RequestId: 唯一请求 ID,每次请求都会返回。定位问题时需要提供该次请求的 RequestId。
:type RequestId: str | tencentcloud/mongodb/v20180408/models.py | __init__ | qin5506/tencentcloud-sdk-python | python | def __init__(self):
'\n :param RequestId: 唯一请求 ID,每次请求都会返回。定位问题时需要提供该次请求的 RequestId。\n :type RequestId: str\n '
self.RequestId = None |
def __init__(self):
'\n :param InstanceIds: 实例ID列表,格式如:cmgo-p8vnipr5。与云数据库控制台页面中显示的实例ID相同\n :type InstanceIds: list of str\n :param AutoRenewFlag: 续费选项,取值范围:0-手动续费,1-自动续费,2-确认不续费\n :type AutoRenewFlag: int\n '
self.InstanceIds = None
self.AutoRenewFlag = None | -1,801,866,595,669,366,500 | :param InstanceIds: 实例ID列表,格式如:cmgo-p8vnipr5。与云数据库控制台页面中显示的实例ID相同
:type InstanceIds: list of str
:param AutoRenewFlag: 续费选项,取值范围:0-手动续费,1-自动续费,2-确认不续费
:type AutoRenewFlag: int | tencentcloud/mongodb/v20180408/models.py | __init__ | qin5506/tencentcloud-sdk-python | python | def __init__(self):
'\n :param InstanceIds: 实例ID列表,格式如:cmgo-p8vnipr5。与云数据库控制台页面中显示的实例ID相同\n :type InstanceIds: list of str\n :param AutoRenewFlag: 续费选项,取值范围:0-手动续费,1-自动续费,2-确认不续费\n :type AutoRenewFlag: int\n '
self.InstanceIds = None
self.AutoRenewFlag = None |
def __init__(self):
'\n :param RequestId: 唯一请求 ID,每次请求都会返回。定位问题时需要提供该次请求的 RequestId。\n :type RequestId: str\n '
self.RequestId = None | -5,957,967,262,820,529,000 | :param RequestId: 唯一请求 ID,每次请求都会返回。定位问题时需要提供该次请求的 RequestId。
:type RequestId: str | tencentcloud/mongodb/v20180408/models.py | __init__ | qin5506/tencentcloud-sdk-python | python | def __init__(self):
'\n :param RequestId: 唯一请求 ID,每次请求都会返回。定位问题时需要提供该次请求的 RequestId。\n :type RequestId: str\n '
self.RequestId = None |
def __init__(self):
'\n :param InstanceId: 实例ID,格式如:cmgo-p8vnipr5。与云数据库控制台页面中显示的实例ID相同\n :type InstanceId: str\n :param UserName: 实例账户名称\n :type UserName: str\n :param Password: 实例新密码,至少包含字母、数字和字符(!@#%^*())中的两种,长度为8-16个字符\n :type Password: str\n '
self.InstanceId... | 2,673,499,813,344,423,400 | :param InstanceId: 实例ID,格式如:cmgo-p8vnipr5。与云数据库控制台页面中显示的实例ID相同
:type InstanceId: str
:param UserName: 实例账户名称
:type UserName: str
:param Password: 实例新密码,至少包含字母、数字和字符(!@#%^*())中的两种,长度为8-16个字符
:type Password: str | tencentcloud/mongodb/v20180408/models.py | __init__ | qin5506/tencentcloud-sdk-python | python | def __init__(self):
'\n :param InstanceId: 实例ID,格式如:cmgo-p8vnipr5。与云数据库控制台页面中显示的实例ID相同\n :type InstanceId: str\n :param UserName: 实例账户名称\n :type UserName: str\n :param Password: 实例新密码,至少包含字母、数字和字符(!@#%^*())中的两种,长度为8-16个字符\n :type Password: str\n '
self.InstanceId... |
def __init__(self):
'\n :param FlowId: 返回的异步任务ID\n :type FlowId: int\n :param RequestId: 唯一请求 ID,每次请求都会返回。定位问题时需要提供该次请求的 RequestId。\n :type RequestId: str\n '
self.FlowId = None
self.RequestId = None | 168,431,123,442,788,260 | :param FlowId: 返回的异步任务ID
:type FlowId: int
:param RequestId: 唯一请求 ID,每次请求都会返回。定位问题时需要提供该次请求的 RequestId。
:type RequestId: str | tencentcloud/mongodb/v20180408/models.py | __init__ | qin5506/tencentcloud-sdk-python | python | def __init__(self):
'\n :param FlowId: 返回的异步任务ID\n :type FlowId: int\n :param RequestId: 唯一请求 ID,每次请求都会返回。定位问题时需要提供该次请求的 RequestId。\n :type RequestId: str\n '
self.FlowId = None
self.RequestId = None |
def __init__(self):
'\n :param SpecCode: 规格信息标识\n :type SpecCode: str\n :param Status: 规格有效标志,取值:0-停止售卖,1-开放售卖\n :type Status: int\n :param MachineType: 机器类型,取值:0-HIO,4-HIO10G\n :type MachineType: str\n :param Cpu: cpu核心数\n :type Cpu: int\n :param Memor... | 342,077,246,768,402,050 | :param SpecCode: 规格信息标识
:type SpecCode: str
:param Status: 规格有效标志,取值:0-停止售卖,1-开放售卖
:type Status: int
:param MachineType: 机器类型,取值:0-HIO,4-HIO10G
:type MachineType: str
:param Cpu: cpu核心数
:type Cpu: int
:param Memory: 内存规格,单位为MB
:type Memory: int
:param DefaultStorage: 默认磁盘规格,单位MB
:type DefaultStorage: int
:param MaxStor... | tencentcloud/mongodb/v20180408/models.py | __init__ | qin5506/tencentcloud-sdk-python | python | def __init__(self):
'\n :param SpecCode: 规格信息标识\n :type SpecCode: str\n :param Status: 规格有效标志,取值:0-停止售卖,1-开放售卖\n :type Status: int\n :param MachineType: 机器类型,取值:0-HIO,4-HIO10G\n :type MachineType: str\n :param Cpu: cpu核心数\n :type Cpu: int\n :param Memor... |
def __init__(self):
'\n :param Region: 地域信息\n :type Region: str\n :param Zone: 可用区信息\n :type Zone: str\n :param SpecItems: 售卖规格信息\n :type SpecItems: list of SpecItem\n '
self.Region = None
self.Zone = None
self.SpecItems = None | 2,896,008,479,591,454,700 | :param Region: 地域信息
:type Region: str
:param Zone: 可用区信息
:type Zone: str
:param SpecItems: 售卖规格信息
:type SpecItems: list of SpecItem | tencentcloud/mongodb/v20180408/models.py | __init__ | qin5506/tencentcloud-sdk-python | python | def __init__(self):
'\n :param Region: 地域信息\n :type Region: str\n :param Zone: 可用区信息\n :type Zone: str\n :param SpecItems: 售卖规格信息\n :type SpecItems: list of SpecItem\n '
self.Region = None
self.Zone = None
self.SpecItems = None |
def __init__(self):
'\n :param TagKey: 标签Key值\n :type TagKey: str\n :param TagValue: 标签值\n :type TagValue: str\n '
self.TagKey = None
self.TagValue = None | 2,818,798,211,660,525,600 | :param TagKey: 标签Key值
:type TagKey: str
:param TagValue: 标签值
:type TagValue: str | tencentcloud/mongodb/v20180408/models.py | __init__ | qin5506/tencentcloud-sdk-python | python | def __init__(self):
'\n :param TagKey: 标签Key值\n :type TagKey: str\n :param TagValue: 标签值\n :type TagValue: str\n '
self.TagKey = None
self.TagValue = None |
def __init__(self):
'\n :param InstanceId: 实例ID,格式如:cmgo-p8vnipr5。\n :type InstanceId: str\n '
self.InstanceId = None | -4,814,981,435,062,103,000 | :param InstanceId: 实例ID,格式如:cmgo-p8vnipr5。
:type InstanceId: str | tencentcloud/mongodb/v20180408/models.py | __init__ | qin5506/tencentcloud-sdk-python | python | def __init__(self):
'\n :param InstanceId: 实例ID,格式如:cmgo-p8vnipr5。\n :type InstanceId: str\n '
self.InstanceId = None |
def __init__(self):
'\n :param AsyncRequestId: 订单ID,表示注销实例成功\n :type AsyncRequestId: str\n :param RequestId: 唯一请求 ID,每次请求都会返回。定位问题时需要提供该次请求的 RequestId。\n :type RequestId: str\n '
self.AsyncRequestId = None
self.RequestId = None | 1,323,675,753,685,545,500 | :param AsyncRequestId: 订单ID,表示注销实例成功
:type AsyncRequestId: str
:param RequestId: 唯一请求 ID,每次请求都会返回。定位问题时需要提供该次请求的 RequestId。
:type RequestId: str | tencentcloud/mongodb/v20180408/models.py | __init__ | qin5506/tencentcloud-sdk-python | python | def __init__(self):
'\n :param AsyncRequestId: 订单ID,表示注销实例成功\n :type AsyncRequestId: str\n :param RequestId: 唯一请求 ID,每次请求都会返回。定位问题时需要提供该次请求的 RequestId。\n :type RequestId: str\n '
self.AsyncRequestId = None
self.RequestId = None |
def __init__(self):
'\n :param InstanceId: 实例ID,格式如:cmgo-p8vnipr5\n :type InstanceId: str\n :param Memory: 升级后的内存大小,单位:GB\n :type Memory: int\n :param Volume: 升级后的硬盘大小,单位:GB\n :type Volume: int\n :param OplogSize: 升级后oplog的大小,单位:GB,默认为磁盘空间的10%,允许设置的最小值为磁盘的10%,最大值为磁盘的... | -4,952,864,791,188,179,000 | :param InstanceId: 实例ID,格式如:cmgo-p8vnipr5
:type InstanceId: str
:param Memory: 升级后的内存大小,单位:GB
:type Memory: int
:param Volume: 升级后的硬盘大小,单位:GB
:type Volume: int
:param OplogSize: 升级后oplog的大小,单位:GB,默认为磁盘空间的10%,允许设置的最小值为磁盘的10%,最大值为磁盘的90%
:type OplogSize: int | tencentcloud/mongodb/v20180408/models.py | __init__ | qin5506/tencentcloud-sdk-python | python | def __init__(self):
'\n :param InstanceId: 实例ID,格式如:cmgo-p8vnipr5\n :type InstanceId: str\n :param Memory: 升级后的内存大小,单位:GB\n :type Memory: int\n :param Volume: 升级后的硬盘大小,单位:GB\n :type Volume: int\n :param OplogSize: 升级后oplog的大小,单位:GB,默认为磁盘空间的10%,允许设置的最小值为磁盘的10%,最大值为磁盘的... |
def __init__(self):
'\n :param DealId: 订单ID\n :type DealId: str\n :param RequestId: 唯一请求 ID,每次请求都会返回。定位问题时需要提供该次请求的 RequestId。\n :type RequestId: str\n '
self.DealId = None
self.RequestId = None | 4,609,315,430,297,468,400 | :param DealId: 订单ID
:type DealId: str
:param RequestId: 唯一请求 ID,每次请求都会返回。定位问题时需要提供该次请求的 RequestId。
:type RequestId: str | tencentcloud/mongodb/v20180408/models.py | __init__ | qin5506/tencentcloud-sdk-python | python | def __init__(self):
'\n :param DealId: 订单ID\n :type DealId: str\n :param RequestId: 唯一请求 ID,每次请求都会返回。定位问题时需要提供该次请求的 RequestId。\n :type RequestId: str\n '
self.DealId = None
self.RequestId = None |
def __init__(self):
'\n :param InstanceId: 实例ID,格式如:cmgo-p8vnipr5。与云数据库控制台页面中显示的实例ID相同\n :type InstanceId: str\n :param Memory: 升级后的内存大小,单位:GB\n :type Memory: int\n :param Volume: 升级后的硬盘大小,单位:GB\n :type Volume: int\n :param OplogSize: 升级后oplog的大小,单位:GB,默认为磁盘空间的10%,允许... | 3,858,098,735,489,418,000 | :param InstanceId: 实例ID,格式如:cmgo-p8vnipr5。与云数据库控制台页面中显示的实例ID相同
:type InstanceId: str
:param Memory: 升级后的内存大小,单位:GB
:type Memory: int
:param Volume: 升级后的硬盘大小,单位:GB
:type Volume: int
:param OplogSize: 升级后oplog的大小,单位:GB,默认为磁盘空间的10%,允许设置的最小值为磁盘的10%,最大值为磁盘的90%
:type OplogSize: int | tencentcloud/mongodb/v20180408/models.py | __init__ | qin5506/tencentcloud-sdk-python | python | def __init__(self):
'\n :param InstanceId: 实例ID,格式如:cmgo-p8vnipr5。与云数据库控制台页面中显示的实例ID相同\n :type InstanceId: str\n :param Memory: 升级后的内存大小,单位:GB\n :type Memory: int\n :param Volume: 升级后的硬盘大小,单位:GB\n :type Volume: int\n :param OplogSize: 升级后oplog的大小,单位:GB,默认为磁盘空间的10%,允许... |
def __init__(self):
'\n :param DealId: 订单ID\n :type DealId: str\n :param RequestId: 唯一请求 ID,每次请求都会返回。定位问题时需要提供该次请求的 RequestId。\n :type RequestId: str\n '
self.DealId = None
self.RequestId = None | 4,609,315,430,297,468,400 | :param DealId: 订单ID
:type DealId: str
:param RequestId: 唯一请求 ID,每次请求都会返回。定位问题时需要提供该次请求的 RequestId。
:type RequestId: str | tencentcloud/mongodb/v20180408/models.py | __init__ | qin5506/tencentcloud-sdk-python | python | def __init__(self):
'\n :param DealId: 订单ID\n :type DealId: str\n :param RequestId: 唯一请求 ID,每次请求都会返回。定位问题时需要提供该次请求的 RequestId。\n :type RequestId: str\n '
self.DealId = None
self.RequestId = None |
def __init__(self, row, to_account):
'Build a Skype Event from a single row.\n\n Args:\n row: A row object (instance of sqlite3.Row) that contains the\n extracted data from a single row in the database.\n to_account: A string containing the accounts (excluding the\n author) o... | -2,209,459,207,477,835,500 | Build a Skype Event from a single row.
Args:
row: A row object (instance of sqlite3.Row) that contains the
extracted data from a single row in the database.
to_account: A string containing the accounts (excluding the
author) of the conversation. | plaso/parsers/sqlite_plugins/skype.py | __init__ | Defense-Cyber-Crime-Center/plaso | python | def __init__(self, row, to_account):
'Build a Skype Event from a single row.\n\n Args:\n row: A row object (instance of sqlite3.Row) that contains the\n extracted data from a single row in the database.\n to_account: A string containing the accounts (excluding the\n author) o... |
def __init__(self, timestamp, usage, identifier, full_name, display_name, email, country):
'Initialize the event.\n\n Args:\n timestamp: The POSIX timestamp value.\n usage: A string containing the description string of the timestamp.\n identifier: The row identifier.\n full_name: A string con... | -1,848,867,815,453,986,800 | Initialize the event.
Args:
timestamp: The POSIX timestamp value.
usage: A string containing the description string of the timestamp.
identifier: The row identifier.
full_name: A string containing the full name of the Skype account holder.
display_name: A string containing the chosen display name of the acco... | plaso/parsers/sqlite_plugins/skype.py | __init__ | Defense-Cyber-Crime-Center/plaso | python | def __init__(self, timestamp, usage, identifier, full_name, display_name, email, country):
'Initialize the event.\n\n Args:\n timestamp: The POSIX timestamp value.\n usage: A string containing the description string of the timestamp.\n identifier: The row identifier.\n full_name: A string con... |
def __init__(self, row, dst_number):
"Read the information related with the SMS.\n\n Args:\n row: row form the sql query.\n row['time_sms']: timestamp when the sms was send.\n row['dstnum_sms']: number which receives the sms.\n row['msg_sms']: text send to this sms.\n d... | 7,238,326,909,042,559,000 | Read the information related with the SMS.
Args:
row: row form the sql query.
row['time_sms']: timestamp when the sms was send.
row['dstnum_sms']: number which receives the sms.
row['msg_sms']: text send to this sms.
dst_number: phone number where the user send the sms. | plaso/parsers/sqlite_plugins/skype.py | __init__ | Defense-Cyber-Crime-Center/plaso | python | def __init__(self, row, dst_number):
"Read the information related with the SMS.\n\n Args:\n row: row form the sql query.\n row['time_sms']: timestamp when the sms was send.\n row['dstnum_sms']: number which receives the sms.\n row['msg_sms']: text send to this sms.\n d... |
def __init__(self, timestamp, call_type, user_start_call, source, destination, video_conference):
'Contains information if the call was cancelled, accepted or finished.\n\n Args:\n timestamp: the timestamp of the event.\n call_type: WAITING, STARTED, FINISHED.\n user_start_call: boolean, t... | 3,729,223,191,627,515,400 | Contains information if the call was cancelled, accepted or finished.
Args:
timestamp: the timestamp of the event.
call_type: WAITING, STARTED, FINISHED.
user_start_call: boolean, true indicates that the owner
account started the call.
source: the account which started the call.
destinatio... | plaso/parsers/sqlite_plugins/skype.py | __init__ | Defense-Cyber-Crime-Center/plaso | python | def __init__(self, timestamp, call_type, user_start_call, source, destination, video_conference):
'Contains information if the call was cancelled, accepted or finished.\n\n Args:\n timestamp: the timestamp of the event.\n call_type: WAITING, STARTED, FINISHED.\n user_start_call: boolean, t... |
def __init__(self, row, timestamp, action_type, source, destination):
'Actions related with sending files.\n\n Args:\n row:\n filepath: path from the file.\n filename: name of the file.\n filesize: size of the file.\n timestamp: when the action happens.\n action_... | -365,589,398,346,313,150 | Actions related with sending files.
Args:
row:
filepath: path from the file.
filename: name of the file.
filesize: size of the file.
timestamp: when the action happens.
action_type: GETSOLICITUDE, SENDSOLICITUDE, ACCEPTED, FINISHED.
source: The account that sent the file.
destination: The account... | plaso/parsers/sqlite_plugins/skype.py | __init__ | Defense-Cyber-Crime-Center/plaso | python | def __init__(self, row, timestamp, action_type, source, destination):
'Actions related with sending files.\n\n Args:\n row:\n filepath: path from the file.\n filename: name of the file.\n filesize: size of the file.\n timestamp: when the action happens.\n action_... |
def ParseAccountInformation(self, parser_mediator, row, query=None, **unused_kwargs):
'Parses the Accounts database.\n\n Args:\n parser_mediator: A parser mediator object (instance of ParserMediator).\n row: The row resulting from the query.\n query: Optional query string. The default is None.\n ... | 1,106,927,174,727,200,300 | Parses the Accounts database.
Args:
parser_mediator: A parser mediator object (instance of ParserMediator).
row: The row resulting from the query.
query: Optional query string. The default is None. | plaso/parsers/sqlite_plugins/skype.py | ParseAccountInformation | Defense-Cyber-Crime-Center/plaso | python | def ParseAccountInformation(self, parser_mediator, row, query=None, **unused_kwargs):
'Parses the Accounts database.\n\n Args:\n parser_mediator: A parser mediator object (instance of ParserMediator).\n row: The row resulting from the query.\n query: Optional query string. The default is None.\n ... |
def ParseChat(self, parser_mediator, row, query=None, **unused_kwargs):
'Parses a chat message row.\n\n Args:\n parser_mediator: A parser mediator object (instance of ParserMediator).\n row: The row resulting from the query.\n query: Optional query string. The default is None.\n '
to_accoun... | 4,747,559,496,911,788,000 | Parses a chat message row.
Args:
parser_mediator: A parser mediator object (instance of ParserMediator).
row: The row resulting from the query.
query: Optional query string. The default is None. | plaso/parsers/sqlite_plugins/skype.py | ParseChat | Defense-Cyber-Crime-Center/plaso | python | def ParseChat(self, parser_mediator, row, query=None, **unused_kwargs):
'Parses a chat message row.\n\n Args:\n parser_mediator: A parser mediator object (instance of ParserMediator).\n row: The row resulting from the query.\n query: Optional query string. The default is None.\n '
to_accoun... |
def ParseSMS(self, parser_mediator, row, query=None, **unused_kwargs):
'Parse SMS.\n\n Args:\n parser_mediator: A parser mediator object (instance of ParserMediator).\n row: The row resulting from the query.\n query: Optional query string. The default is None.\n '
dst_number = row['dstnum_s... | 8,635,166,024,142,017,000 | Parse SMS.
Args:
parser_mediator: A parser mediator object (instance of ParserMediator).
row: The row resulting from the query.
query: Optional query string. The default is None. | plaso/parsers/sqlite_plugins/skype.py | ParseSMS | Defense-Cyber-Crime-Center/plaso | python | def ParseSMS(self, parser_mediator, row, query=None, **unused_kwargs):
'Parse SMS.\n\n Args:\n parser_mediator: A parser mediator object (instance of ParserMediator).\n row: The row resulting from the query.\n query: Optional query string. The default is None.\n '
dst_number = row['dstnum_s... |
def ParseCall(self, parser_mediator, row, query=None, **unused_kwargs):
'Parse the calls taking into accounts some rows.\n\n Args:\n parser_mediator: A parser mediator object (instance of ParserMediator).\n row: The row resulting from the query.\n query: Optional query string. The default is None.... | 3,052,790,095,354,990,600 | Parse the calls taking into accounts some rows.
Args:
parser_mediator: A parser mediator object (instance of ParserMediator).
row: The row resulting from the query.
query: Optional query string. The default is None. | plaso/parsers/sqlite_plugins/skype.py | ParseCall | Defense-Cyber-Crime-Center/plaso | python | def ParseCall(self, parser_mediator, row, query=None, **unused_kwargs):
'Parse the calls taking into accounts some rows.\n\n Args:\n parser_mediator: A parser mediator object (instance of ParserMediator).\n row: The row resulting from the query.\n query: Optional query string. The default is None.... |
def ParseFileTransfer(self, parser_mediator, row, cache=None, database=None, query=None, **unused_kwargs):
'Parse the transfer files.\n\n There is no direct relationship between who sends the file and\n who accepts the file.\n\n Args:\n parser_mediator: A parser mediator object (instance of ParserMe... | -4,675,259,592,888,954,000 | Parse the transfer files.
There is no direct relationship between who sends the file and
who accepts the file.
Args:
parser_mediator: A parser mediator object (instance of ParserMediator).
row: the row with all information related with the file transfers.
query: Optional query string. The default is None.
c... | plaso/parsers/sqlite_plugins/skype.py | ParseFileTransfer | Defense-Cyber-Crime-Center/plaso | python | def ParseFileTransfer(self, parser_mediator, row, cache=None, database=None, query=None, **unused_kwargs):
'Parse the transfer files.\n\n There is no direct relationship between who sends the file and\n who accepts the file.\n\n Args:\n parser_mediator: A parser mediator object (instance of ParserMe... |
def searchable(self):
'Enable search line edit visible.'
self._search_line_edit.setVisible(True)
return self | -8,225,402,888,302,331,000 | Enable search line edit visible. | dayu_widgets/item_view_set.py | searchable | kanbang/dayu_widgets | python | def searchable(self):
self._search_line_edit.setVisible(True)
return self |
def setUp(self):
'\n Set up method to run before each test case\n '
self.new_user = credentials('Paul', '123') | 4,424,100,375,332,029,000 | Set up method to run before each test case | credentials_test.py | setUp | paulmunyao/Password-Locker | python | def setUp(self):
'\n \n '
self.new_user = credentials('Paul', '123') |
def test__init__(self):
'\n test__init__ test case to test if the object is initialized properly\n '
self.assertEqual(self.new_user.user_name, 'Paul')
self.assertEqual(self.new_user.password, '123') | 734,204,525,461,232,000 | test__init__ test case to test if the object is initialized properly | credentials_test.py | test__init__ | paulmunyao/Password-Locker | python | def test__init__(self):
'\n \n '
self.assertEqual(self.new_user.user_name, 'Paul')
self.assertEqual(self.new_user.password, '123') |
def test__save_user(self):
'\n test to see if the user is saved\n '
self.new_credentials.save_credentials()
self.assertEqual(len(credentials.user_list), 1) | -8,293,255,632,897,053,000 | test to see if the user is saved | credentials_test.py | test__save_user | paulmunyao/Password-Locker | python | def test__save_user(self):
'\n \n '
self.new_credentials.save_credentials()
self.assertEqual(len(credentials.user_list), 1) |
@property
def verbosity(self) -> Verbosity:
"\n Verbosity level (default `warning`)\n\n Level 0: only show 'error' messages.\n Level 1: also show 'warning' messages.\n Level 2: also show 'info' messages.\n Level 3: also show 'hint' messages.\n Level 4: also show very detail... | 6,424,720,869,344,444,000 | Verbosity level (default `warning`)
Level 0: only show 'error' messages.
Level 1: also show 'warning' messages.
Level 2: also show 'info' messages.
Level 3: also show 'hint' messages.
Level 4: also show very detailed progress for 'debug'ging. | scanpy/_settings.py | verbosity | gamazeps/scanpy | python | @property
def verbosity(self) -> Verbosity:
"\n Verbosity level (default `warning`)\n\n Level 0: only show 'error' messages.\n Level 1: also show 'warning' messages.\n Level 2: also show 'info' messages.\n Level 3: also show 'hint' messages.\n Level 4: also show very detail... |
@property
def plot_suffix(self) -> str:
'Global suffix that is appended to figure filenames.\n '
return self._plot_suffix | 3,859,983,572,578,482,000 | Global suffix that is appended to figure filenames. | scanpy/_settings.py | plot_suffix | gamazeps/scanpy | python | @property
def plot_suffix(self) -> str:
'\n '
return self._plot_suffix |
@property
def file_format_data(self) -> str:
"File format for saving AnnData objects.\n\n Allowed are 'txt', 'csv' (comma separated value file) for exporting and 'h5ad'\n (hdf5) for lossless saving.\n "
return self._file_format_data | -3,070,800,732,118,796,300 | File format for saving AnnData objects.
Allowed are 'txt', 'csv' (comma separated value file) for exporting and 'h5ad'
(hdf5) for lossless saving. | scanpy/_settings.py | file_format_data | gamazeps/scanpy | python | @property
def file_format_data(self) -> str:
"File format for saving AnnData objects.\n\n Allowed are 'txt', 'csv' (comma separated value file) for exporting and 'h5ad'\n (hdf5) for lossless saving.\n "
return self._file_format_data |
@property
def file_format_figs(self) -> str:
"File format for saving figures.\n\n For example 'png', 'pdf' or 'svg'. Many other formats work as well (see\n `matplotlib.pyplot.savefig`).\n "
return self._file_format_figs | -1,319,306,077,014,287,400 | File format for saving figures.
For example 'png', 'pdf' or 'svg'. Many other formats work as well (see
`matplotlib.pyplot.savefig`). | scanpy/_settings.py | file_format_figs | gamazeps/scanpy | python | @property
def file_format_figs(self) -> str:
"File format for saving figures.\n\n For example 'png', 'pdf' or 'svg'. Many other formats work as well (see\n `matplotlib.pyplot.savefig`).\n "
return self._file_format_figs |
@property
def autosave(self) -> bool:
' Automatically save figures in :attr:`~scanpy._settings.ScanpyConfig.figdir` (default `False`).\n\n Do not show plots/figures interactively.\n '
return self._autosave | 3,660,266,334,677,051,400 | Automatically save figures in :attr:`~scanpy._settings.ScanpyConfig.figdir` (default `False`).
Do not show plots/figures interactively. | scanpy/_settings.py | autosave | gamazeps/scanpy | python | @property
def autosave(self) -> bool:
' Automatically save figures in :attr:`~scanpy._settings.ScanpyConfig.figdir` (default `False`).\n\n Do not show plots/figures interactively.\n '
return self._autosave |
@property
def autoshow(self) -> bool:
' Automatically show figures if `autosave == False` (default `True`).\n\n There is no need to call the matplotlib pl.show() in this case.\n '
return self._autoshow | -6,447,307,756,049,594,000 | Automatically show figures if `autosave == False` (default `True`).
There is no need to call the matplotlib pl.show() in this case. | scanpy/_settings.py | autoshow | gamazeps/scanpy | python | @property
def autoshow(self) -> bool:
' Automatically show figures if `autosave == False` (default `True`).\n\n There is no need to call the matplotlib pl.show() in this case.\n '
return self._autoshow |
@property
def writedir(self) -> Path:
' Directory where the function scanpy.write writes to by default.\n '
return self._writedir | -5,245,418,655,141,521,000 | Directory where the function scanpy.write writes to by default. | scanpy/_settings.py | writedir | gamazeps/scanpy | python | @property
def writedir(self) -> Path:
' \n '
return self._writedir |
@property
def cachedir(self) -> Path:
" Directory for cache files (default `'./cache/'`).\n "
return self._cachedir | -1,021,373,160,847,789,800 | Directory for cache files (default `'./cache/'`). | scanpy/_settings.py | cachedir | gamazeps/scanpy | python | @property
def cachedir(self) -> Path:
" \n "
return self._cachedir |
@property
def datasetdir(self) -> Path:
" Directory for example :mod:`~scanpy.datasets` (default `'./data/'`).\n "
return self._datasetdir | 6,038,991,707,708,268,000 | Directory for example :mod:`~scanpy.datasets` (default `'./data/'`). | scanpy/_settings.py | datasetdir | gamazeps/scanpy | python | @property
def datasetdir(self) -> Path:
" \n "
return self._datasetdir |
@property
def figdir(self) -> Path:
" Directory for saving figures (default `'./figures/'`).\n "
return self._figdir | 3,064,606,170,553,432,600 | Directory for saving figures (default `'./figures/'`). | scanpy/_settings.py | figdir | gamazeps/scanpy | python | @property
def figdir(self) -> Path:
" \n "
return self._figdir |
@property
def max_memory(self) -> Union[(int, float)]:
' Maximal memory usage in Gigabyte.\n\n Is currently not well respected....\n '
return self._max_memory | -7,489,614,085,946,220,000 | Maximal memory usage in Gigabyte.
Is currently not well respected.... | scanpy/_settings.py | max_memory | gamazeps/scanpy | python | @property
def max_memory(self) -> Union[(int, float)]:
' Maximal memory usage in Gigabyte.\n\n Is currently not well respected....\n '
return self._max_memory |
@property
def n_jobs(self) -> int:
' Default number of jobs/ CPUs to use for parallel computing.\n '
return self._n_jobs | 1,803,948,937,692,783,000 | Default number of jobs/ CPUs to use for parallel computing. | scanpy/_settings.py | n_jobs | gamazeps/scanpy | python | @property
def n_jobs(self) -> int:
' \n '
return self._n_jobs |
@property
def logpath(self) -> Optional[Path]:
' The file path `logfile` was set to.\n '
return self._logpath | -2,058,415,471,124,060,200 | The file path `logfile` was set to. | scanpy/_settings.py | logpath | gamazeps/scanpy | python | @property
def logpath(self) -> Optional[Path]:
' \n '
return self._logpath |
@property
def logfile(self) -> TextIO:
" The open file to write logs to.\n\n Set it to a :class:`~pathlib.Path` or :class:`str` to open a new one.\n The default `None` corresponds to :obj:`sys.stdout` in jupyter notebooks\n and to :obj:`sys.stderr` otherwise.\n\n For backwards com... | 629,563,856,985,492,100 | The open file to write logs to.
Set it to a :class:`~pathlib.Path` or :class:`str` to open a new one.
The default `None` corresponds to :obj:`sys.stdout` in jupyter notebooks
and to :obj:`sys.stderr` otherwise.
For backwards compatibility, setting it to `''` behaves like setting it to `None`. | scanpy/_settings.py | logfile | gamazeps/scanpy | python | @property
def logfile(self) -> TextIO:
" The open file to write logs to.\n\n Set it to a :class:`~pathlib.Path` or :class:`str` to open a new one.\n The default `None` corresponds to :obj:`sys.stdout` in jupyter notebooks\n and to :obj:`sys.stderr` otherwise.\n\n For backwards com... |
@property
def categories_to_ignore(self) -> List[str]:
' Categories that are omitted in plotting etc.\n '
return self._categories_to_ignore | -7,675,846,271,189,168,000 | Categories that are omitted in plotting etc. | scanpy/_settings.py | categories_to_ignore | gamazeps/scanpy | python | @property
def categories_to_ignore(self) -> List[str]:
' \n '
return self._categories_to_ignore |
def set_figure_params(self, scanpy: bool=True, dpi: int=80, dpi_save: int=150, frameon: bool=True, vector_friendly: bool=True, fontsize: int=14, color_map: Optional[str]=None, format: Union[(str, Iterable[str])]='pdf', transparent: bool=False, ipython_format: str='png2x'):
" Set resolution/size, styling and ... | 7,115,684,889,461,385,000 | Set resolution/size, styling and format of figures.
Parameters
----------
scanpy
Init default values for :obj:`matplotlib.rcParams` suited for Scanpy.
dpi
Resolution of rendered figures - this influences the size of figures in notebooks.
dpi_save
Resolution of saved figures. This should typically be higher... | scanpy/_settings.py | set_figure_params | gamazeps/scanpy | python | def set_figure_params(self, scanpy: bool=True, dpi: int=80, dpi_save: int=150, frameon: bool=True, vector_friendly: bool=True, fontsize: int=14, color_map: Optional[str]=None, format: Union[(str, Iterable[str])]='pdf', transparent: bool=False, ipython_format: str='png2x'):
" Set resolution/size, styling and ... |
@staticmethod
def _is_run_from_ipython():
'Determines whether run from Ipython.\n\n Only affects progress bars.\n '
try:
__IPYTHON__
return True
except NameError:
return False | -3,831,951,316,772,162,000 | Determines whether run from Ipython.
Only affects progress bars. | scanpy/_settings.py | _is_run_from_ipython | gamazeps/scanpy | python | @staticmethod
def _is_run_from_ipython():
'Determines whether run from Ipython.\n\n Only affects progress bars.\n '
try:
__IPYTHON__
return True
except NameError:
return False |
def fps(src: torch.Tensor, batch=None, ratio=None, random_start=True):
'"A sampling algorithm from the `"PointNet++: Deep Hierarchical Feature\n Learning on Point Sets in a Metric Space"\n <https://arxiv.org/abs/1706.02413>`_ paper, which iteratively samples the\n most distant point with regard to the rest... | -3,016,408,835,132,523,000 | "A sampling algorithm from the `"PointNet++: Deep Hierarchical Feature
Learning on Point Sets in a Metric Space"
<https://arxiv.org/abs/1706.02413>`_ paper, which iteratively samples the
most distant point with regard to the rest points.
Args:
src (Tensor): Point feature matrix
:math:`\mathbf{X} \in \mathb... | torch_cluster/fps.py | fps | Hacky-DH/pytorch_cluster | python | def fps(src: torch.Tensor, batch=None, ratio=None, random_start=True):
'"A sampling algorithm from the `"PointNet++: Deep Hierarchical Feature\n Learning on Point Sets in a Metric Space"\n <https://arxiv.org/abs/1706.02413>`_ paper, which iteratively samples the\n most distant point with regard to the rest... |
def request_endpoint(audio, speech_config, output_directory, lexical):
'Request the speech service endpoint\n Args:\n audio: Input data frame\n speech_config: Choice between scoring and \n output_folder: LUIS app ID\n case: LUIS subscription key\n lexical: Minimum confidence sc... | 6,316,814,903,347,039,000 | Request the speech service endpoint
Args:
audio: Input data frame
speech_config: Choice between scoring and
output_folder: LUIS app ID
case: LUIS subscription key
lexical: Minimum confidence score for LUIS result, between 0.00 and 1.00
Returns:
df: Scoring data frame with predicted intents and ... | src/stt.py | request_endpoint | microsoft/SpeechServices | python | def request_endpoint(audio, speech_config, output_directory, lexical):
'Request the speech service endpoint\n Args:\n audio: Input data frame\n speech_config: Choice between scoring and \n output_folder: LUIS app ID\n case: LUIS subscription key\n lexical: Minimum confidence sc... |
def process_recognition(result, filename, output_directory, lexical):
'Process recognition received from the speech service\n Args:\n result: Result object returned by STT-service\n filename: Filename for output file\n output_directory: Output directory for the file\n lexical: Boolean... | 7,902,317,877,991,636,000 | Process recognition received from the speech service
Args:
result: Result object returned by STT-service
filename: Filename for output file
output_directory: Output directory for the file
lexical: Boolean to enable extended lexical version of STT-result
Returns:
text: Processed recognition as string | src/stt.py | process_recognition | microsoft/SpeechServices | python | def process_recognition(result, filename, output_directory, lexical):
'Process recognition received from the speech service\n Args:\n result: Result object returned by STT-service\n filename: Filename for output file\n output_directory: Output directory for the file\n lexical: Boolean... |
def write_transcription(output_directory, text):
'Write transcription to file\n Args:\n text: Processed recognition as string\n output_directory: Output directory for the file\n Returns:\n Writes output to file\n '
if (not os.path.exists(f'{output_directory}/transcriptions.txt')):
... | 5,400,919,661,310,008,000 | Write transcription to file
Args:
text: Processed recognition as string
output_directory: Output directory for the file
Returns:
Writes output to file | src/stt.py | write_transcription | microsoft/SpeechServices | python | def write_transcription(output_directory, text):
'Write transcription to file\n Args:\n text: Processed recognition as string\n output_directory: Output directory for the file\n Returns:\n Writes output to file\n '
if (not os.path.exists(f'{output_directory}/transcriptions.txt')):
... |
def main(speech_files, output_directory, lexical=False, enable_proxy=False, *argv):
'Main function for STT-functionality\n Args:\n speech_files: Directory of audio files to be transcribed\n output_directory: Output directory for the file\n lexical: Boolean to enable extended lexical version ... | 4,704,541,752,918,082,000 | Main function for STT-functionality
Args:
speech_files: Directory of audio files to be transcribed
output_directory: Output directory for the file
lexical: Boolean to enable extended lexical version of STT-result
enable_proxy: Boolean to enable proxy function in case you need it
*argv: Proxy informa... | src/stt.py | main | microsoft/SpeechServices | python | def main(speech_files, output_directory, lexical=False, enable_proxy=False, *argv):
'Main function for STT-functionality\n Args:\n speech_files: Directory of audio files to be transcribed\n output_directory: Output directory for the file\n lexical: Boolean to enable extended lexical version ... |
def get_params() -> AttributeDict:
'Return a dict containing training parameters.\n\n All training related parameters that are not passed from the commandline\n is saved in the variable `params`.\n\n Commandline options are merged into `params` after they are parsed, so\n you can also access them via `p... | -8,064,291,771,705,554,000 | Return a dict containing training parameters.
All training related parameters that are not passed from the commandline
is saved in the variable `params`.
Commandline options are merged into `params` after they are parsed, so
you can also access them via `params`.
Explanation of options saved in `params`:
- exp_... | egs/librispeech/ASR/tdnn_lstm_ctc/train.py | get_params | aarora8/icefall | python | def get_params() -> AttributeDict:
'Return a dict containing training parameters.\n\n All training related parameters that are not passed from the commandline\n is saved in the variable `params`.\n\n Commandline options are merged into `params` after they are parsed, so\n you can also access them via `p... |
def load_checkpoint_if_available(params: AttributeDict, model: nn.Module, optimizer: Optional[torch.optim.Optimizer]=None, scheduler: Optional[torch.optim.lr_scheduler._LRScheduler]=None) -> None:
'Load checkpoint from file.\n\n If params.start_epoch is positive, it will load the checkpoint from\n `params.sta... | -8,657,515,020,424,165,000 | Load checkpoint from file.
If params.start_epoch is positive, it will load the checkpoint from
`params.start_epoch - 1`. Otherwise, this function does nothing.
Apart from loading state dict for `model`, `optimizer` and `scheduler`,
it also updates `best_train_epoch`, `best_train_loss`, `best_valid_epoch`,
and `best_v... | egs/librispeech/ASR/tdnn_lstm_ctc/train.py | load_checkpoint_if_available | aarora8/icefall | python | def load_checkpoint_if_available(params: AttributeDict, model: nn.Module, optimizer: Optional[torch.optim.Optimizer]=None, scheduler: Optional[torch.optim.lr_scheduler._LRScheduler]=None) -> None:
'Load checkpoint from file.\n\n If params.start_epoch is positive, it will load the checkpoint from\n `params.sta... |
def save_checkpoint(params: AttributeDict, model: nn.Module, optimizer: torch.optim.Optimizer, scheduler: torch.optim.lr_scheduler._LRScheduler, rank: int=0) -> None:
'Save model, optimizer, scheduler and training stats to file.\n\n Args:\n params:\n It is returned by :func:`get_params`.\n model... | 4,040,369,752,138,356,000 | Save model, optimizer, scheduler and training stats to file.
Args:
params:
It is returned by :func:`get_params`.
model:
The training model. | egs/librispeech/ASR/tdnn_lstm_ctc/train.py | save_checkpoint | aarora8/icefall | python | def save_checkpoint(params: AttributeDict, model: nn.Module, optimizer: torch.optim.Optimizer, scheduler: torch.optim.lr_scheduler._LRScheduler, rank: int=0) -> None:
'Save model, optimizer, scheduler and training stats to file.\n\n Args:\n params:\n It is returned by :func:`get_params`.\n model... |
def compute_loss(params: AttributeDict, model: nn.Module, batch: dict, graph_compiler: CtcTrainingGraphCompiler, is_training: bool) -> Tuple[(Tensor, MetricsTracker)]:
'\n Compute CTC loss given the model and its inputs.\n\n Args:\n params:\n Parameters for training. See :func:`get_params`.\n ... | -85,694,476,020,040,110 | Compute CTC loss given the model and its inputs.
Args:
params:
Parameters for training. See :func:`get_params`.
model:
The model for training. It is an instance of TdnnLstm in our case.
batch:
A batch of data. See `lhotse.dataset.K2SpeechRecognitionDataset()`
for the content in it.
graph_compil... | egs/librispeech/ASR/tdnn_lstm_ctc/train.py | compute_loss | aarora8/icefall | python | def compute_loss(params: AttributeDict, model: nn.Module, batch: dict, graph_compiler: CtcTrainingGraphCompiler, is_training: bool) -> Tuple[(Tensor, MetricsTracker)]:
'\n Compute CTC loss given the model and its inputs.\n\n Args:\n params:\n Parameters for training. See :func:`get_params`.\n ... |
def compute_validation_loss(params: AttributeDict, model: nn.Module, graph_compiler: CtcTrainingGraphCompiler, valid_dl: torch.utils.data.DataLoader, world_size: int=1) -> MetricsTracker:
'Run the validation process. The validation loss\n is saved in `params.valid_loss`.\n '
model.eval()
tot_loss = Me... | 5,921,257,761,175,695,000 | Run the validation process. The validation loss
is saved in `params.valid_loss`. | egs/librispeech/ASR/tdnn_lstm_ctc/train.py | compute_validation_loss | aarora8/icefall | python | def compute_validation_loss(params: AttributeDict, model: nn.Module, graph_compiler: CtcTrainingGraphCompiler, valid_dl: torch.utils.data.DataLoader, world_size: int=1) -> MetricsTracker:
'Run the validation process. The validation loss\n is saved in `params.valid_loss`.\n '
model.eval()
tot_loss = Me... |
def train_one_epoch(params: AttributeDict, model: nn.Module, optimizer: torch.optim.Optimizer, graph_compiler: CtcTrainingGraphCompiler, train_dl: torch.utils.data.DataLoader, valid_dl: torch.utils.data.DataLoader, tb_writer: Optional[SummaryWriter]=None, world_size: int=1) -> None:
'Train the model for one epoch.\... | 6,525,583,933,407,344,000 | Train the model for one epoch.
The training loss from the mean of all frames is saved in
`params.train_loss`. It runs the validation process every
`params.valid_interval` batches.
Args:
params:
It is returned by :func:`get_params`.
model:
The model for training.
optimizer:
The optimizer we are using... | egs/librispeech/ASR/tdnn_lstm_ctc/train.py | train_one_epoch | aarora8/icefall | python | def train_one_epoch(params: AttributeDict, model: nn.Module, optimizer: torch.optim.Optimizer, graph_compiler: CtcTrainingGraphCompiler, train_dl: torch.utils.data.DataLoader, valid_dl: torch.utils.data.DataLoader, tb_writer: Optional[SummaryWriter]=None, world_size: int=1) -> None:
'Train the model for one epoch.\... |
def run(rank, world_size, args):
'\n Args:\n rank:\n It is a value between 0 and `world_size-1`, which is\n passed automatically by `mp.spawn()` in :func:`main`.\n The node with rank 0 is responsible for saving checkpoint.\n world_size:\n Number of GPUs for DDP training.\n ... | 1,424,617,520,821,463,800 | Args:
rank:
It is a value between 0 and `world_size-1`, which is
passed automatically by `mp.spawn()` in :func:`main`.
The node with rank 0 is responsible for saving checkpoint.
world_size:
Number of GPUs for DDP training.
args:
The return value of get_parser().parse_args() | egs/librispeech/ASR/tdnn_lstm_ctc/train.py | run | aarora8/icefall | python | def run(rank, world_size, args):
'\n Args:\n rank:\n It is a value between 0 and `world_size-1`, which is\n passed automatically by `mp.spawn()` in :func:`main`.\n The node with rank 0 is responsible for saving checkpoint.\n world_size:\n Number of GPUs for DDP training.\n ... |
def __init__(self):
'\n The schema generator generates a GraphQL schema.\n The purpose is to provide a schema to which resolvers are then\n attached, which is then given to Ariadne, and for resolvers to\n have information about expected types.\n\n For RPSL queries and types, this ... | -6,691,687,725,778,370,000 | The schema generator generates a GraphQL schema.
The purpose is to provide a schema to which resolvers are then
attached, which is then given to Ariadne, and for resolvers to
have information about expected types.
For RPSL queries and types, this is dynamically generated based on
the RPSL objects from irrd.rpsl. Other... | irrd/server/graphql/schema_generator.py | __init__ | morrowc/irrd | python | def __init__(self):
'\n The schema generator generates a GraphQL schema.\n The purpose is to provide a schema to which resolvers are then\n attached, which is then given to Ariadne, and for resolvers to\n have information about expected types.\n\n For RPSL queries and types, this ... |
def _set_rpsl_query_fields(self):
'\n Create a sub-schema for the fields that can be queried for RPSL objects.\n This includes all fields from all objects, along with a few\n special fields.\n '
string_list_fields = {'rpsl_pk', 'sources', 'object_class'}.union(lookup_field_names())
... | -5,529,400,313,608,977,000 | Create a sub-schema for the fields that can be queried for RPSL objects.
This includes all fields from all objects, along with a few
special fields. | irrd/server/graphql/schema_generator.py | _set_rpsl_query_fields | morrowc/irrd | python | def _set_rpsl_query_fields(self):
'\n Create a sub-schema for the fields that can be queried for RPSL objects.\n This includes all fields from all objects, along with a few\n special fields.\n '
string_list_fields = {'rpsl_pk', 'sources', 'object_class'}.union(lookup_field_names())
... |
def _set_enums(self):
'\n Create the schema for enums, current RPKI and scope filter status.\n '
self.enums = ''
for enum in [RPKIStatus, ScopeFilterStatus]:
self.enums += f'''enum {enum.__name__} {{
'''
for value in enum:
self.enums += f''' {value.name}
'''
... | 3,618,900,208,178,960,000 | Create the schema for enums, current RPKI and scope filter status. | irrd/server/graphql/schema_generator.py | _set_enums | morrowc/irrd | python | def _set_enums(self):
'\n \n '
self.enums =
for enum in [RPKIStatus, ScopeFilterStatus]:
self.enums += f'enum {enum.__name__} {{
'
for value in enum:
self.enums += f' {value.name}
'
self.enums += '}\n\n' |
def _set_rpsl_object_interface_schema(self):
'\n Create the schema for RPSLObject, which contains only fields that\n are common to every known RPSL object, along with meta\n '
common_fields = None
for rpsl_object_class in OBJECT_CLASS_MAPPING.values():
if (common_fields is None)... | -4,808,421,517,788,839,000 | Create the schema for RPSLObject, which contains only fields that
are common to every known RPSL object, along with meta | irrd/server/graphql/schema_generator.py | _set_rpsl_object_interface_schema | morrowc/irrd | python | def _set_rpsl_object_interface_schema(self):
'\n Create the schema for RPSLObject, which contains only fields that\n are common to every known RPSL object, along with meta\n '
common_fields = None
for rpsl_object_class in OBJECT_CLASS_MAPPING.values():
if (common_fields is None)... |
def _set_rpsl_contact_schema(self):
'\n Create the schema for RPSLContact. This contains shared fields between\n RPSLPerson and RPSLRole, as they are so similar.\n '
common_fields = set(RPSLPerson.fields.keys()).intersection(set(RPSLRole.fields.keys()))
common_fields = common_fields.uni... | 6,149,839,233,699,117,000 | Create the schema for RPSLContact. This contains shared fields between
RPSLPerson and RPSLRole, as they are so similar. | irrd/server/graphql/schema_generator.py | _set_rpsl_contact_schema | morrowc/irrd | python | def _set_rpsl_contact_schema(self):
'\n Create the schema for RPSLContact. This contains shared fields between\n RPSLPerson and RPSLRole, as they are so similar.\n '
common_fields = set(RPSLPerson.fields.keys()).intersection(set(RPSLRole.fields.keys()))
common_fields = common_fields.uni... |
def _set_rpsl_object_schemas(self):
'\n Create the schemas for each specific RPSL object class.\n Each of these implements RPSLObject, and RPSLPerson/RPSLRole\n implement RPSLContact as well.\n '
self.graphql_types = defaultdict(dict)
schemas = OrderedDict()
for (object_class... | -2,214,039,650,622,902,500 | Create the schemas for each specific RPSL object class.
Each of these implements RPSLObject, and RPSLPerson/RPSLRole
implement RPSLContact as well. | irrd/server/graphql/schema_generator.py | _set_rpsl_object_schemas | morrowc/irrd | python | def _set_rpsl_object_schemas(self):
'\n Create the schemas for each specific RPSL object class.\n Each of these implements RPSLObject, and RPSLPerson/RPSLRole\n implement RPSLContact as well.\n '
self.graphql_types = defaultdict(dict)
schemas = OrderedDict()
for (object_class... |
def _graphql_type_for_rpsl_field(self, field: RPSLTextField) -> str:
'\n Return the GraphQL type for a regular RPSL field.\n This is always a list of strings if the field is a list and/or\n can occur multiple times.\n '
if ((RPSLFieldListMixin in field.__class__.__bases__) or field.m... | -4,626,389,360,198,936,000 | Return the GraphQL type for a regular RPSL field.
This is always a list of strings if the field is a list and/or
can occur multiple times. | irrd/server/graphql/schema_generator.py | _graphql_type_for_rpsl_field | morrowc/irrd | python | def _graphql_type_for_rpsl_field(self, field: RPSLTextField) -> str:
'\n Return the GraphQL type for a regular RPSL field.\n This is always a list of strings if the field is a list and/or\n can occur multiple times.\n '
if ((RPSLFieldListMixin in field.__class__.__bases__) or field.m... |
def _grapql_type_for_reference_field(self, field_name: str, rpsl_field: RPSLTextField) -> Tuple[(Optional[str], Optional[str])]:
'\n Return the GraphQL name and type for a reference field.\n For example, for a field "admin-c" that refers to person/role,\n returns (\'adminC\', \'[RPSLContactUnio... | -6,929,608,980,858,813,000 | Return the GraphQL name and type for a reference field.
For example, for a field "admin-c" that refers to person/role,
returns ('adminC', '[RPSLContactUnion!]').
Some fields are excluded because they are syntactical references,
not real references. | irrd/server/graphql/schema_generator.py | _grapql_type_for_reference_field | morrowc/irrd | python | def _grapql_type_for_reference_field(self, field_name: str, rpsl_field: RPSLTextField) -> Tuple[(Optional[str], Optional[str])]:
'\n Return the GraphQL name and type for a reference field.\n For example, for a field "admin-c" that refers to person/role,\n returns (\'adminC\', \'[RPSLContactUnio... |
def _generate_schema_str(self, name: str, graphql_type: str, fields: Dict[(str, str)], implements: Optional[str]=None) -> str:
'\n Generate a schema string for a given name, object type and dict of fields.\n '
schema = f'{graphql_type} {name} '
if implements:
schema += f'implements {im... | -3,460,663,556,156,464,000 | Generate a schema string for a given name, object type and dict of fields. | irrd/server/graphql/schema_generator.py | _generate_schema_str | morrowc/irrd | python | def _generate_schema_str(self, name: str, graphql_type: str, fields: Dict[(str, str)], implements: Optional[str]=None) -> str:
'\n \n '
schema = f'{graphql_type} {name} '
if implements:
schema += f'implements {implements} '
schema += '{\n'
for (field, field_type) in fields.item... |
@cached_property
def additional_properties_type():
'\n This must be a method because a model may have properties that are\n of type self, this must run after the class is loaded\n '
lazy_import()
return (bool, date, datetime, dict, float, int, list, str, none_type) | 1,702,168,743,392,494,600 | This must be a method because a model may have properties that are
of type self, this must run after the class is loaded | sdks/python/client/argo_workflows/model/lifecycle_handler.py | additional_properties_type | AnuragThePathak/argo-workflows | python | @cached_property
def additional_properties_type():
'\n This must be a method because a model may have properties that are\n of type self, this must run after the class is loaded\n '
lazy_import()
return (bool, date, datetime, dict, float, int, list, str, none_type) |
@cached_property
def openapi_types():
'\n This must be a method because a model may have properties that are\n of type self, this must run after the class is loaded\n\n Returns\n openapi_types (dict): The key is attribute name\n and the value is attribute type.\n ... | -7,129,412,387,319,964,000 | This must be a method because a model may have properties that are
of type self, this must run after the class is loaded
Returns
openapi_types (dict): The key is attribute name
and the value is attribute type. | sdks/python/client/argo_workflows/model/lifecycle_handler.py | openapi_types | AnuragThePathak/argo-workflows | python | @cached_property
def openapi_types():
'\n This must be a method because a model may have properties that are\n of type self, this must run after the class is loaded\n\n Returns\n openapi_types (dict): The key is attribute name\n and the value is attribute type.\n ... |
@classmethod
@convert_js_args_to_python_args
def _from_openapi_data(cls, *args, **kwargs):
'LifecycleHandler - a model defined in OpenAPI\n\n Keyword Args:\n _check_type (bool): if True, values for parameters in openapi_types\n will be type checked and a TypeError wi... | 7,650,468,161,058,138,000 | LifecycleHandler - a model defined in OpenAPI
Keyword Args:
_check_type (bool): if True, values for parameters in openapi_types
will be type checked and a TypeError will be
raised if the wrong type is input.
Defaults to True
_path_to_item ... | sdks/python/client/argo_workflows/model/lifecycle_handler.py | _from_openapi_data | AnuragThePathak/argo-workflows | python | @classmethod
@convert_js_args_to_python_args
def _from_openapi_data(cls, *args, **kwargs):
'LifecycleHandler - a model defined in OpenAPI\n\n Keyword Args:\n _check_type (bool): if True, values for parameters in openapi_types\n will be type checked and a TypeError wi... |
@convert_js_args_to_python_args
def __init__(self, *args, **kwargs):
'LifecycleHandler - a model defined in OpenAPI\n\n Keyword Args:\n _check_type (bool): if True, values for parameters in openapi_types\n will be type checked and a TypeError will be\n ... | -4,069,975,532,339,985,000 | LifecycleHandler - a model defined in OpenAPI
Keyword Args:
_check_type (bool): if True, values for parameters in openapi_types
will be type checked and a TypeError will be
raised if the wrong type is input.
Defaults to True
_path_to_item ... | sdks/python/client/argo_workflows/model/lifecycle_handler.py | __init__ | AnuragThePathak/argo-workflows | python | @convert_js_args_to_python_args
def __init__(self, *args, **kwargs):
'LifecycleHandler - a model defined in OpenAPI\n\n Keyword Args:\n _check_type (bool): if True, values for parameters in openapi_types\n will be type checked and a TypeError will be\n ... |
def addTableEntry(self, tableEntry=None):
'\n Add a table entry to the switch\n '
response = self.stub.AddEntry(tableEntry)
if (response.code == 0):
Log.error('Error for entry:', tableEntry, 'on switch', self.name) | -3,067,755,745,714,994,000 | Add a table entry to the switch | Controller-Implementation/libs/core/SwitchConnection.py | addTableEntry | qcz994/p4-bier | python | def addTableEntry(self, tableEntry=None):
'\n \n '
response = self.stub.AddEntry(tableEntry)
if (response.code == 0):
Log.error('Error for entry:', tableEntry, 'on switch', self.name) |
def removeTableEntry(self, tableEntry=None):
'\n Remove a table entry from the switch\n '
response = self.stub.RemoveEntry(tableEntry)
if (response.code == 0):
Log.error('Error while removing entry:', tableEntry, 'on switch', self.name) | 4,002,122,230,061,831,000 | Remove a table entry from the switch | Controller-Implementation/libs/core/SwitchConnection.py | removeTableEntry | qcz994/p4-bier | python | def removeTableEntry(self, tableEntry=None):
'\n \n '
response = self.stub.RemoveEntry(tableEntry)
if (response.code == 0):
Log.error('Error while removing entry:', tableEntry, 'on switch', self.name) |
@read_session
def get_bad_replicas_summary(rse_expression=None, from_date=None, to_date=None, filter=None, session=None):
"\n List the bad file replicas summary. Method used by the rucio-ui.\n :param rse_expression: The RSE expression.\n :param from_date: The start date.\n :param to_date: The end date.\... | 6,065,724,123,909,250,000 | List the bad file replicas summary. Method used by the rucio-ui.
:param rse_expression: The RSE expression.
:param from_date: The start date.
:param to_date: The end date.
:param filter: Dictionary of attributes by which the RSE results should be filtered. e.g.: {'availability_write': True}
:param session: The database... | lib/rucio/core/replica.py | get_bad_replicas_summary | bari12/rucio | python | @read_session
def get_bad_replicas_summary(rse_expression=None, from_date=None, to_date=None, filter=None, session=None):
"\n List the bad file replicas summary. Method used by the rucio-ui.\n :param rse_expression: The RSE expression.\n :param from_date: The start date.\n :param to_date: The end date.\... |
@read_session
def __exists_replicas(rse_id, scope=None, name=None, path=None, session=None):
'\n Internal method to check if a replica exists at a given site.\n :param rse_id: The RSE id.\n :param scope: The scope of the file.\n :param name: The name of the file.\n :param path: The path of the replic... | 2,500,845,477,038,364,000 | Internal method to check if a replica exists at a given site.
:param rse_id: The RSE id.
:param scope: The scope of the file.
:param name: The name of the file.
:param path: The path of the replica.
:param session: The database session in use. | lib/rucio/core/replica.py | __exists_replicas | bari12/rucio | python | @read_session
def __exists_replicas(rse_id, scope=None, name=None, path=None, session=None):
'\n Internal method to check if a replica exists at a given site.\n :param rse_id: The RSE id.\n :param scope: The scope of the file.\n :param name: The name of the file.\n :param path: The path of the replic... |
@read_session
def list_bad_replicas_status(state=BadFilesStatus.BAD, rse_id=None, younger_than=None, older_than=None, limit=None, list_pfns=False, vo='def', session=None):
'\n List the bad file replicas history states. Method used by the rucio-ui.\n :param state: The state of the file (SUSPICIOUS or BAD).\n ... | -5,288,423,726,230,488,000 | List the bad file replicas history states. Method used by the rucio-ui.
:param state: The state of the file (SUSPICIOUS or BAD).
:param rse_id: The RSE id.
:param younger_than: datetime object to select bad replicas younger than this date.
:param older_than: datetime object to select bad replicas older than this date.... | lib/rucio/core/replica.py | list_bad_replicas_status | bari12/rucio | python | @read_session
def list_bad_replicas_status(state=BadFilesStatus.BAD, rse_id=None, younger_than=None, older_than=None, limit=None, list_pfns=False, vo='def', session=None):
'\n List the bad file replicas history states. Method used by the rucio-ui.\n :param state: The state of the file (SUSPICIOUS or BAD).\n ... |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.