body stringlengths 26 98.2k | body_hash int64 -9,222,864,604,528,158,000 9,221,803,474B | docstring stringlengths 1 16.8k | path stringlengths 5 230 | name stringlengths 1 96 | repository_name stringlengths 7 89 | lang stringclasses 1
value | body_without_docstring stringlengths 20 98.2k |
|---|---|---|---|---|---|---|---|
@property
def n(self):
'Returns the number of elements in the set covered by this cover.'
return self._n | 2,698,383,659,266,666,500 | Returns the number of elements in the set covered by this cover. | igraph/clustering.py | n | tuandnvn/ecat_learning | python | @property
def n(self):
return self._n |
def size(self, idx):
'Returns the size of a given cluster.\n\n @param idx: the cluster in which we are interested.\n '
return len(self[idx]) | -2,611,264,052,909,075,500 | Returns the size of a given cluster.
@param idx: the cluster in which we are interested. | igraph/clustering.py | size | tuandnvn/ecat_learning | python | def size(self, idx):
'Returns the size of a given cluster.\n\n @param idx: the cluster in which we are interested.\n '
return len(self[idx]) |
def sizes(self, *args):
'Returns the size of given clusters.\n\n The indices are given as positional arguments. If there are no\n positional arguments, the function will return the sizes of all clusters.\n '
if args:
return [len(self._clusters[idx]) for idx in args]
return [len(... | -449,967,433,735,330,750 | Returns the size of given clusters.
The indices are given as positional arguments. If there are no
positional arguments, the function will return the sizes of all clusters. | igraph/clustering.py | sizes | tuandnvn/ecat_learning | python | def sizes(self, *args):
'Returns the size of given clusters.\n\n The indices are given as positional arguments. If there are no\n positional arguments, the function will return the sizes of all clusters.\n '
if args:
return [len(self._clusters[idx]) for idx in args]
return [len(... |
def size_histogram(self, bin_width=1):
'Returns the histogram of cluster sizes.\n\n @param bin_width: the bin width of the histogram\n @return: a L{Histogram} object\n '
return Histogram(bin_width, self.sizes()) | -2,461,763,455,575,568,000 | Returns the histogram of cluster sizes.
@param bin_width: the bin width of the histogram
@return: a L{Histogram} object | igraph/clustering.py | size_histogram | tuandnvn/ecat_learning | python | def size_histogram(self, bin_width=1):
'Returns the histogram of cluster sizes.\n\n @param bin_width: the bin width of the histogram\n @return: a L{Histogram} object\n '
return Histogram(bin_width, self.sizes()) |
def summary(self, verbosity=0, width=None):
'Returns the summary of the cover.\n\n The summary includes the number of items and clusters, and also the\n list of members for each of the clusters if the verbosity is nonzero.\n\n @param verbosity: determines whether the cluster members should be\n... | -6,739,493,703,869,613,000 | Returns the summary of the cover.
The summary includes the number of items and clusters, and also the
list of members for each of the clusters if the verbosity is nonzero.
@param verbosity: determines whether the cluster members should be
printed. Zero verbosity prints the number of items and clusters only.
@return... | igraph/clustering.py | summary | tuandnvn/ecat_learning | python | def summary(self, verbosity=0, width=None):
'Returns the summary of the cover.\n\n The summary includes the number of items and clusters, and also the\n list of members for each of the clusters if the verbosity is nonzero.\n\n @param verbosity: determines whether the cluster members should be\n... |
def _formatted_cluster_iterator(self):
'Iterates over the clusters and formats them into a string to be\n presented in the summary.'
for cluster in self:
(yield ', '.join((str(member) for member in cluster))) | 810,895,616,325,758,500 | Iterates over the clusters and formats them into a string to be
presented in the summary. | igraph/clustering.py | _formatted_cluster_iterator | tuandnvn/ecat_learning | python | def _formatted_cluster_iterator(self):
'Iterates over the clusters and formats them into a string to be\n presented in the summary.'
for cluster in self:
(yield ', '.join((str(member) for member in cluster))) |
def __init__(self, graph, clusters=None):
'Creates a cover object for a given graph.\n\n @param graph: the graph that will be associated to the cover\n @param clusters: the list of clusters. If C{None}, it is assumed\n that there is only a single cluster that covers the whole graph.\n ... | 3,603,294,436,514,720,300 | Creates a cover object for a given graph.
@param graph: the graph that will be associated to the cover
@param clusters: the list of clusters. If C{None}, it is assumed
that there is only a single cluster that covers the whole graph. | igraph/clustering.py | __init__ | tuandnvn/ecat_learning | python | def __init__(self, graph, clusters=None):
'Creates a cover object for a given graph.\n\n @param graph: the graph that will be associated to the cover\n @param clusters: the list of clusters. If C{None}, it is assumed\n that there is only a single cluster that covers the whole graph.\n ... |
def crossing(self):
'Returns a boolean vector where element M{i} is C{True} iff edge\n M{i} lies between clusters, C{False} otherwise.'
membership = [frozenset(cluster) for cluster in self.membership]
return [membership[v1].isdisjoint(membership[v2]) for (v1, v2) in self.graph.get_edgelist()] | 3,467,809,426,251,922,400 | Returns a boolean vector where element M{i} is C{True} iff edge
M{i} lies between clusters, C{False} otherwise. | igraph/clustering.py | crossing | tuandnvn/ecat_learning | python | def crossing(self):
'Returns a boolean vector where element M{i} is C{True} iff edge\n M{i} lies between clusters, C{False} otherwise.'
membership = [frozenset(cluster) for cluster in self.membership]
return [membership[v1].isdisjoint(membership[v2]) for (v1, v2) in self.graph.get_edgelist()] |
@property
def graph(self):
'Returns the graph belonging to this object'
return self._graph | -6,013,293,917,706,169,000 | Returns the graph belonging to this object | igraph/clustering.py | graph | tuandnvn/ecat_learning | python | @property
def graph(self):
return self._graph |
def subgraph(self, idx):
"Get the subgraph belonging to a given cluster.\n\n @param idx: the cluster index\n @return: a copy of the subgraph\n @precondition: the vertex set of the graph hasn't been modified since\n the moment the cover was constructed.\n "
return self._graph... | -3,335,315,451,606,916,600 | Get the subgraph belonging to a given cluster.
@param idx: the cluster index
@return: a copy of the subgraph
@precondition: the vertex set of the graph hasn't been modified since
the moment the cover was constructed. | igraph/clustering.py | subgraph | tuandnvn/ecat_learning | python | def subgraph(self, idx):
"Get the subgraph belonging to a given cluster.\n\n @param idx: the cluster index\n @return: a copy of the subgraph\n @precondition: the vertex set of the graph hasn't been modified since\n the moment the cover was constructed.\n "
return self._graph... |
def subgraphs(self):
"Gets all the subgraphs belonging to each of the clusters.\n\n @return: a list containing copies of the subgraphs\n @precondition: the vertex set of the graph hasn't been modified since\n the moment the cover was constructed.\n "
return [self._graph.subgraph(cl... | -6,612,431,575,192,130,000 | Gets all the subgraphs belonging to each of the clusters.
@return: a list containing copies of the subgraphs
@precondition: the vertex set of the graph hasn't been modified since
the moment the cover was constructed. | igraph/clustering.py | subgraphs | tuandnvn/ecat_learning | python | def subgraphs(self):
"Gets all the subgraphs belonging to each of the clusters.\n\n @return: a list containing copies of the subgraphs\n @precondition: the vertex set of the graph hasn't been modified since\n the moment the cover was constructed.\n "
return [self._graph.subgraph(cl... |
def __plot__(self, context, bbox, palette, *args, **kwds):
'Plots the cover to the given Cairo context in the given\n bounding box.\n\n This is done by calling L{Graph.__plot__()} with the same arguments, but\n drawing nice colored blobs around the vertex groups.\n\n This method understa... | -8,213,005,111,044,524,000 | Plots the cover to the given Cairo context in the given
bounding box.
This is done by calling L{Graph.__plot__()} with the same arguments, but
drawing nice colored blobs around the vertex groups.
This method understands all the positional and keyword arguments that
are understood by L{Graph.__plot__()}, only the diff... | igraph/clustering.py | __plot__ | tuandnvn/ecat_learning | python | def __plot__(self, context, bbox, palette, *args, **kwds):
'Plots the cover to the given Cairo context in the given\n bounding box.\n\n This is done by calling L{Graph.__plot__()} with the same arguments, but\n drawing nice colored blobs around the vertex groups.\n\n This method understa... |
def _formatted_cluster_iterator(self):
'Iterates over the clusters and formats them into a string to be\n presented in the summary.'
if self._graph.is_named():
names = self._graph.vs['name']
for cluster in self:
(yield ', '.join((str(names[member]) for member in cluster)))
... | 6,838,424,363,819,696,000 | Iterates over the clusters and formats them into a string to be
presented in the summary. | igraph/clustering.py | _formatted_cluster_iterator | tuandnvn/ecat_learning | python | def _formatted_cluster_iterator(self):
'Iterates over the clusters and formats them into a string to be\n presented in the summary.'
if self._graph.is_named():
names = self._graph.vs['name']
for cluster in self:
(yield ', '.join((str(names[member]) for member in cluster)))
... |
def __init__(self, graph, blocks=None, cohesion=None, parent=None):
'Constructs a new cohesive block structure for the given graph.\n\n If any of I{blocks}, I{cohesion} or I{parent} is C{None}, all the\n arguments will be ignored and L{Graph.cohesive_blocks()} will be\n called to calculate the ... | -2,177,125,745,029,691,600 | Constructs a new cohesive block structure for the given graph.
If any of I{blocks}, I{cohesion} or I{parent} is C{None}, all the
arguments will be ignored and L{Graph.cohesive_blocks()} will be
called to calculate the cohesive blocks. Otherwise, these three
variables should describe the *result* of a cohesive block st... | igraph/clustering.py | __init__ | tuandnvn/ecat_learning | python | def __init__(self, graph, blocks=None, cohesion=None, parent=None):
'Constructs a new cohesive block structure for the given graph.\n\n If any of I{blocks}, I{cohesion} or I{parent} is C{None}, all the\n arguments will be ignored and L{Graph.cohesive_blocks()} will be\n called to calculate the ... |
def cohesion(self, idx):
'Returns the cohesion of the group with the given index.'
return self._cohesion[idx] | -3,294,174,005,412,983,000 | Returns the cohesion of the group with the given index. | igraph/clustering.py | cohesion | tuandnvn/ecat_learning | python | def cohesion(self, idx):
return self._cohesion[idx] |
def cohesions(self):
'Returns the list of cohesion values for each group.'
return self._cohesion[:] | -9,078,904,226,652,061,000 | Returns the list of cohesion values for each group. | igraph/clustering.py | cohesions | tuandnvn/ecat_learning | python | def cohesions(self):
return self._cohesion[:] |
def hierarchy(self):
'Returns a new graph that describes the hierarchical relationships\n between the groups.\n\n The new graph will be a directed tree; an edge will point from\n vertex M{i} to vertex M{j} if group M{i} is a superset of group M{j}.\n In other words, the edges point downw... | 5,892,941,643,755,646,000 | Returns a new graph that describes the hierarchical relationships
between the groups.
The new graph will be a directed tree; an edge will point from
vertex M{i} to vertex M{j} if group M{i} is a superset of group M{j}.
In other words, the edges point downwards. | igraph/clustering.py | hierarchy | tuandnvn/ecat_learning | python | def hierarchy(self):
'Returns a new graph that describes the hierarchical relationships\n between the groups.\n\n The new graph will be a directed tree; an edge will point from\n vertex M{i} to vertex M{j} if group M{i} is a superset of group M{j}.\n In other words, the edges point downw... |
def max_cohesion(self, idx):
'Finds the maximum cohesion score among all the groups that contain\n the given vertex.'
result = 0
for (cohesion, cluster) in izip(self._cohesion, self._clusters):
if (idx in cluster):
result = max(result, cohesion)
return result | 4,111,417,056,152,360,000 | Finds the maximum cohesion score among all the groups that contain
the given vertex. | igraph/clustering.py | max_cohesion | tuandnvn/ecat_learning | python | def max_cohesion(self, idx):
'Finds the maximum cohesion score among all the groups that contain\n the given vertex.'
result = 0
for (cohesion, cluster) in izip(self._cohesion, self._clusters):
if (idx in cluster):
result = max(result, cohesion)
return result |
def max_cohesions(self):
'For each vertex in the graph, returns the maximum cohesion score\n among all the groups that contain the vertex.'
result = ([0] * self._graph.vcount())
for (cohesion, cluster) in izip(self._cohesion, self._clusters):
for idx in cluster:
result[idx] = max(... | -9,051,087,033,062,930,000 | For each vertex in the graph, returns the maximum cohesion score
among all the groups that contain the vertex. | igraph/clustering.py | max_cohesions | tuandnvn/ecat_learning | python | def max_cohesions(self):
'For each vertex in the graph, returns the maximum cohesion score\n among all the groups that contain the vertex.'
result = ([0] * self._graph.vcount())
for (cohesion, cluster) in izip(self._cohesion, self._clusters):
for idx in cluster:
result[idx] = max(... |
def parent(self, idx):
'Returns the parent group index of the group with the given index\n or C{None} if the given group is the root.'
return self._parent[idx] | -2,531,504,258,032,719,000 | Returns the parent group index of the group with the given index
or C{None} if the given group is the root. | igraph/clustering.py | parent | tuandnvn/ecat_learning | python | def parent(self, idx):
'Returns the parent group index of the group with the given index\n or C{None} if the given group is the root.'
return self._parent[idx] |
def parents(self):
'Returns the list of parent group indices for each group or C{None}\n if the given group is the root.'
return self._parent[:] | 4,112,353,736,681,686,500 | Returns the list of parent group indices for each group or C{None}
if the given group is the root. | igraph/clustering.py | parents | tuandnvn/ecat_learning | python | def parents(self):
'Returns the list of parent group indices for each group or C{None}\n if the given group is the root.'
return self._parent[:] |
def __plot__(self, context, bbox, palette, *args, **kwds):
'Plots the cohesive block structure to the given Cairo context in\n the given bounding box.\n\n Since a L{CohesiveBlocks} instance is also a L{VertexCover}, keyword\n arguments accepted by L{VertexCover.__plot__()} are also accepted her... | 3,229,657,450,994,385,400 | Plots the cohesive block structure to the given Cairo context in
the given bounding box.
Since a L{CohesiveBlocks} instance is also a L{VertexCover}, keyword
arguments accepted by L{VertexCover.__plot__()} are also accepted here.
The only difference is that the vertices are colored according to their
maximal cohesions... | igraph/clustering.py | __plot__ | tuandnvn/ecat_learning | python | def __plot__(self, context, bbox, palette, *args, **kwds):
'Plots the cohesive block structure to the given Cairo context in\n the given bounding box.\n\n Since a L{CohesiveBlocks} instance is also a L{VertexCover}, keyword\n arguments accepted by L{VertexCover.__plot__()} are also accepted her... |
def safeintdiv(x, y):
'Safe integer division that handles None gracefully'
if (x is None):
return None
return int((x / y)) | 5,398,877,169,241,703,000 | Safe integer division that handles None gracefully | igraph/clustering.py | safeintdiv | tuandnvn/ecat_learning | python | def safeintdiv(x, y):
if (x is None):
return None
return int((x / y)) |
def safebisect(intervals, x):
'Safe list bisection that handles None gracefully'
if (x is None):
return None
return bisect(intervals, x) | 8,481,049,750,746,316,000 | Safe list bisection that handles None gracefully | igraph/clustering.py | safebisect | tuandnvn/ecat_learning | python | def safebisect(intervals, x):
if (x is None):
return None
return bisect(intervals, x) |
@pytest.fixture
def pressure_sensor() -> mmr920C04.PressureSensor:
'Fixture for pressure sensor driver.'
return mmr920C04.PressureSensor() | 876,683,139,357,676,000 | Fixture for pressure sensor driver. | hardware/tests/opentrons_hardware/sensors/test_sensor_drivers.py | pressure_sensor | Opentrons/protocol_framework | python | @pytest.fixture
def pressure_sensor() -> mmr920C04.PressureSensor:
return mmr920C04.PressureSensor() |
@pytest.fixture
def capacitive_sensor() -> fdc1004.CapacitiveSensor:
'Fixture for capacitive sensor driver.'
return fdc1004.CapacitiveSensor() | -5,125,160,098,761,263,000 | Fixture for capacitive sensor driver. | hardware/tests/opentrons_hardware/sensors/test_sensor_drivers.py | capacitive_sensor | Opentrons/protocol_framework | python | @pytest.fixture
def capacitive_sensor() -> fdc1004.CapacitiveSensor:
return fdc1004.CapacitiveSensor() |
@pytest.fixture
def temperature_sensor() -> hdc2080.EnvironmentSensor:
'Fixture for temperature sensor driver.'
return hdc2080.EnvironmentSensor(SensorType.temperature) | -6,278,661,759,028,029,000 | Fixture for temperature sensor driver. | hardware/tests/opentrons_hardware/sensors/test_sensor_drivers.py | temperature_sensor | Opentrons/protocol_framework | python | @pytest.fixture
def temperature_sensor() -> hdc2080.EnvironmentSensor:
return hdc2080.EnvironmentSensor(SensorType.temperature) |
@pytest.fixture
def humidity_sensor() -> hdc2080.EnvironmentSensor:
'Fixture for humidity sensor driver.'
return hdc2080.EnvironmentSensor(SensorType.humidity) | -2,967,201,093,913,927,000 | Fixture for humidity sensor driver. | hardware/tests/opentrons_hardware/sensors/test_sensor_drivers.py | humidity_sensor | Opentrons/protocol_framework | python | @pytest.fixture
def humidity_sensor() -> hdc2080.EnvironmentSensor:
return hdc2080.EnvironmentSensor(SensorType.humidity) |
@pytest.mark.parametrize(argnames=['sensor', 'node', 'message'], argvalues=[[lazy_fixture('pressure_sensor'), NodeId.pipette_left, BaselineSensorRequest(payload=BaselineSensorRequestPayload(sensor=SensorTypeField(SensorType.pressure), sensor_id=SensorIdField(SensorId.S0), sample_rate=UInt16Field(10)))], [lazy_fixture('... | -4,447,089,151,653,519,400 | Test that a polling function sends the expected message. | hardware/tests/opentrons_hardware/sensors/test_sensor_drivers.py | test_polling | Opentrons/protocol_framework | python | @pytest.mark.parametrize(argnames=['sensor', 'node', 'message'], argvalues=[[lazy_fixture('pressure_sensor'), NodeId.pipette_left, BaselineSensorRequest(payload=BaselineSensorRequestPayload(sensor=SensorTypeField(SensorType.pressure), sensor_id=SensorIdField(SensorId.S0), sample_rate=UInt16Field(10)))], [lazy_fixture('... |
@pytest.mark.parametrize(argnames=['sensor'], argvalues=[[lazy_fixture('pressure_sensor')], [lazy_fixture('capacitive_sensor')]])
async def test_receive_data_polling(sensor: sensor_abc.AbstractAdvancedSensor, mock_messenger: mock.AsyncMock, can_message_notifier: MockCanMessageNotifier) -> None:
'Test that data is r... | 2,886,839,213,322,297,000 | Test that data is received from the polling function. | hardware/tests/opentrons_hardware/sensors/test_sensor_drivers.py | test_receive_data_polling | Opentrons/protocol_framework | python | @pytest.mark.parametrize(argnames=['sensor'], argvalues=[[lazy_fixture('pressure_sensor')], [lazy_fixture('capacitive_sensor')]])
async def test_receive_data_polling(sensor: sensor_abc.AbstractAdvancedSensor, mock_messenger: mock.AsyncMock, can_message_notifier: MockCanMessageNotifier) -> None:
def responder(... |
@pytest.mark.parametrize(argnames=['sensor', 'node', 'message'], argvalues=[[lazy_fixture('pressure_sensor'), NodeId.pipette_left, WriteToSensorRequest(payload=WriteToSensorRequestPayload(sensor=SensorTypeField(SensorType.pressure), sensor_id=SensorIdField(SensorId.S0), data=UInt32Field(SensorDataType.build([2, 2, 0, 0... | 9,097,184,948,181,648,000 | Check that writing sensor data is successful. | hardware/tests/opentrons_hardware/sensors/test_sensor_drivers.py | test_write | Opentrons/protocol_framework | python | @pytest.mark.parametrize(argnames=['sensor', 'node', 'message'], argvalues=[[lazy_fixture('pressure_sensor'), NodeId.pipette_left, WriteToSensorRequest(payload=WriteToSensorRequestPayload(sensor=SensorTypeField(SensorType.pressure), sensor_id=SensorIdField(SensorId.S0), data=UInt32Field(SensorDataType.build([2, 2, 0, 0... |
@pytest.mark.parametrize(argnames=['sensor', 'node', 'message'], argvalues=[[lazy_fixture('pressure_sensor'), NodeId.pipette_left, ReadFromSensorRequest(payload=ReadFromSensorRequestPayload(sensor=SensorTypeField(SensorType.pressure), sensor_id=SensorIdField(SensorId.S0), offset_reading=UInt8Field(False)))], [lazy_fixt... | -3,859,175,399,935,780,000 | Test that a read function sends the expected message. | hardware/tests/opentrons_hardware/sensors/test_sensor_drivers.py | test_read | Opentrons/protocol_framework | python | @pytest.mark.parametrize(argnames=['sensor', 'node', 'message'], argvalues=[[lazy_fixture('pressure_sensor'), NodeId.pipette_left, ReadFromSensorRequest(payload=ReadFromSensorRequestPayload(sensor=SensorTypeField(SensorType.pressure), sensor_id=SensorIdField(SensorId.S0), offset_reading=UInt8Field(False)))], [lazy_fixt... |
@pytest.mark.parametrize(argnames=['sensor'], argvalues=[[lazy_fixture('pressure_sensor')], [lazy_fixture('capacitive_sensor')], [lazy_fixture('temperature_sensor')], [lazy_fixture('humidity_sensor')]])
async def test_receive_data_read(sensor: sensor_abc.AbstractAdvancedSensor, mock_messenger: mock.AsyncMock, can_messa... | -7,847,721,743,718,292,000 | Test that data is received from the read function. | hardware/tests/opentrons_hardware/sensors/test_sensor_drivers.py | test_receive_data_read | Opentrons/protocol_framework | python | @pytest.mark.parametrize(argnames=['sensor'], argvalues=[[lazy_fixture('pressure_sensor')], [lazy_fixture('capacitive_sensor')], [lazy_fixture('temperature_sensor')], [lazy_fixture('humidity_sensor')]])
async def test_receive_data_read(sensor: sensor_abc.AbstractAdvancedSensor, mock_messenger: mock.AsyncMock, can_messa... |
@pytest.mark.parametrize(argnames=['sensor'], argvalues=[[lazy_fixture('pressure_sensor')], [lazy_fixture('capacitive_sensor')]])
async def test_threshold(sensor: sensor_abc.AbstractAdvancedSensor, mock_messenger: mock.AsyncMock, can_message_notifier: MockCanMessageNotifier) -> None:
'Test that data is received fro... | -666,913,451,662,296,800 | Test that data is received from the threshold function. | hardware/tests/opentrons_hardware/sensors/test_sensor_drivers.py | test_threshold | Opentrons/protocol_framework | python | @pytest.mark.parametrize(argnames=['sensor'], argvalues=[[lazy_fixture('pressure_sensor')], [lazy_fixture('capacitive_sensor')]])
async def test_threshold(sensor: sensor_abc.AbstractAdvancedSensor, mock_messenger: mock.AsyncMock, can_message_notifier: MockCanMessageNotifier) -> None:
def responder(node_id: No... |
@pytest.mark.parametrize(argnames=['node_id', 'timeout', 'sensor'], argvalues=[[NodeId.pipette_left, 1, lazy_fixture('pressure_sensor')], [NodeId.pipette_left, 5, lazy_fixture('capacitive_sensor')]])
async def test_bind_to_sync(mock_messenger: mock.AsyncMock, can_message_notifier: MockCanMessageNotifier, sensor: sensor... | -859,439,439,960,516,100 | Test for bind_to_sync.
Tests that bind_to_sync does in fact
send out a BindSensorOutputRequest. | hardware/tests/opentrons_hardware/sensors/test_sensor_drivers.py | test_bind_to_sync | Opentrons/protocol_framework | python | @pytest.mark.parametrize(argnames=['node_id', 'timeout', 'sensor'], argvalues=[[NodeId.pipette_left, 1, lazy_fixture('pressure_sensor')], [NodeId.pipette_left, 5, lazy_fixture('capacitive_sensor')]])
async def test_bind_to_sync(mock_messenger: mock.AsyncMock, can_message_notifier: MockCanMessageNotifier, sensor: sensor... |
@pytest.mark.parametrize(argnames=['sensor', 'node_id', 'timeout'], argvalues=[[lazy_fixture('capacitive_sensor'), NodeId.pipette_right, 10], [lazy_fixture('pressure_sensor'), NodeId.pipette_left, 2]])
async def test_get_baseline(mock_messenger: mock.AsyncMock, can_message_notifier: MockCanMessageNotifier, sensor: sens... | -1,316,188,757,608,001,800 | Test for get_baseline.
Tests that a BaselineSensorRequest gets sent,
and reads ReadFromSensorResponse message containing the
correct information. | hardware/tests/opentrons_hardware/sensors/test_sensor_drivers.py | test_get_baseline | Opentrons/protocol_framework | python | @pytest.mark.parametrize(argnames=['sensor', 'node_id', 'timeout'], argvalues=[[lazy_fixture('capacitive_sensor'), NodeId.pipette_right, 10], [lazy_fixture('pressure_sensor'), NodeId.pipette_left, 2]])
async def test_get_baseline(mock_messenger: mock.AsyncMock, can_message_notifier: MockCanMessageNotifier, sensor: sens... |
@pytest.mark.parametrize(argnames=['sensor', 'node_id', 'timeout'], argvalues=[[lazy_fixture('capacitive_sensor'), NodeId.pipette_left, 2], [lazy_fixture('pressure_sensor'), NodeId.pipette_right, 3]])
async def test_debug_poll(mock_messenger: mock.AsyncMock, sensor: sensor_abc.AbstractAdvancedSensor, node_id: NodeId, t... | -8,902,417,442,430,636,000 | Test for debug poll. | hardware/tests/opentrons_hardware/sensors/test_sensor_drivers.py | test_debug_poll | Opentrons/protocol_framework | python | @pytest.mark.parametrize(argnames=['sensor', 'node_id', 'timeout'], argvalues=[[lazy_fixture('capacitive_sensor'), NodeId.pipette_left, 2], [lazy_fixture('pressure_sensor'), NodeId.pipette_right, 3]])
async def test_debug_poll(mock_messenger: mock.AsyncMock, sensor: sensor_abc.AbstractAdvancedSensor, node_id: NodeId, t... |
@pytest.mark.parametrize(argnames=['sensor', 'node_id', 'timeout'], argvalues=[[lazy_fixture('capacitive_sensor'), NodeId.pipette_left, 2], [lazy_fixture('pressure_sensor'), NodeId.pipette_right, 3], [lazy_fixture('temperature_sensor'), NodeId.pipette_left, 2], [lazy_fixture('humidity_sensor'), NodeId.pipette_right, 2]... | 3,318,708,624,848,140,300 | Test for getting peripheral device status. | hardware/tests/opentrons_hardware/sensors/test_sensor_drivers.py | test_peripheral_status | Opentrons/protocol_framework | python | @pytest.mark.parametrize(argnames=['sensor', 'node_id', 'timeout'], argvalues=[[lazy_fixture('capacitive_sensor'), NodeId.pipette_left, 2], [lazy_fixture('pressure_sensor'), NodeId.pipette_right, 3], [lazy_fixture('temperature_sensor'), NodeId.pipette_left, 2], [lazy_fixture('humidity_sensor'), NodeId.pipette_right, 2]... |
def responder(node_id: NodeId, message: MessageDefinition) -> None:
'Message responder.'
can_message_notifier.notify(ReadFromSensorResponse(payload=ReadFromSensorResponsePayload(sensor_data=Int32Field(256), sensor_id=SensorIdField(SensorId.S0), sensor=SensorTypeField(sensor._sensor_type))), ArbitrationId(parts=... | -330,064,312,598,272,060 | Message responder. | hardware/tests/opentrons_hardware/sensors/test_sensor_drivers.py | responder | Opentrons/protocol_framework | python | def responder(node_id: NodeId, message: MessageDefinition) -> None:
can_message_notifier.notify(ReadFromSensorResponse(payload=ReadFromSensorResponsePayload(sensor_data=Int32Field(256), sensor_id=SensorIdField(SensorId.S0), sensor=SensorTypeField(sensor._sensor_type))), ArbitrationId(parts=ArbitrationIdParts(m... |
def responder(node_id: NodeId, message: MessageDefinition) -> None:
'Message responder.'
can_message_notifier.notify(ReadFromSensorResponse(payload=ReadFromSensorResponsePayload(sensor_data=Int32Field(256), sensor_id=SensorIdField(SensorId.S0), sensor=SensorTypeField(sensor._sensor_type))), ArbitrationId(parts=... | -330,064,312,598,272,060 | Message responder. | hardware/tests/opentrons_hardware/sensors/test_sensor_drivers.py | responder | Opentrons/protocol_framework | python | def responder(node_id: NodeId, message: MessageDefinition) -> None:
can_message_notifier.notify(ReadFromSensorResponse(payload=ReadFromSensorResponsePayload(sensor_data=Int32Field(256), sensor_id=SensorIdField(SensorId.S0), sensor=SensorTypeField(sensor._sensor_type))), ArbitrationId(parts=ArbitrationIdParts(m... |
def responder(node_id: NodeId, message: MessageDefinition) -> None:
'Message responder.'
if isinstance(message, SetSensorThresholdRequest):
can_message_notifier.notify(SensorThresholdResponse(payload=SensorThresholdResponsePayload(threshold=message.payload.threshold, sensor=SensorTypeField(sensor._senso... | 7,150,935,569,391,111,000 | Message responder. | hardware/tests/opentrons_hardware/sensors/test_sensor_drivers.py | responder | Opentrons/protocol_framework | python | def responder(node_id: NodeId, message: MessageDefinition) -> None:
if isinstance(message, SetSensorThresholdRequest):
can_message_notifier.notify(SensorThresholdResponse(payload=SensorThresholdResponsePayload(threshold=message.payload.threshold, sensor=SensorTypeField(sensor._sensor_type), sensor_id=S... |
def responder(node_id: NodeId, message: MessageDefinition) -> None:
'Message responder.'
if isinstance(message, BaselineSensorRequest):
can_message_notifier.notify(ReadFromSensorResponse(payload=ReadFromSensorResponsePayload(sensor=SensorTypeField(sensor._sensor_type), sensor_id=SensorIdField(SensorId.S... | -778,312,438,194,596,100 | Message responder. | hardware/tests/opentrons_hardware/sensors/test_sensor_drivers.py | responder | Opentrons/protocol_framework | python | def responder(node_id: NodeId, message: MessageDefinition) -> None:
if isinstance(message, BaselineSensorRequest):
can_message_notifier.notify(ReadFromSensorResponse(payload=ReadFromSensorResponsePayload(sensor=SensorTypeField(sensor._sensor_type), sensor_id=SensorIdField(SensorId.S0), sensor_data=Int3... |
def responder(node_id: NodeId, message: MessageDefinition) -> None:
'Message responder.'
if isinstance(message, PeripheralStatusRequest):
can_message_notifier.notify(PeripheralStatusResponse(payload=PeripheralStatusResponsePayload(sensor=SensorTypeField(sensor._sensor_type), sensor_id=SensorIdField(Sens... | -2,441,369,327,342,985,000 | Message responder. | hardware/tests/opentrons_hardware/sensors/test_sensor_drivers.py | responder | Opentrons/protocol_framework | python | def responder(node_id: NodeId, message: MessageDefinition) -> None:
if isinstance(message, PeripheralStatusRequest):
can_message_notifier.notify(PeripheralStatusResponse(payload=PeripheralStatusResponsePayload(sensor=SensorTypeField(sensor._sensor_type), sensor_id=SensorIdField(SensorId.S0), status=UIn... |
def has_mongo():
'Determine if MongoDB is up and usable'
if os.environ.get('MP_FAKEMONGO'):
mongo = False
else:
try:
pymongo.MongoClient()
mongo = True
except:
mongo = False
return mongo | 2,335,455,765,548,238,300 | Determine if MongoDB is up and usable | pymatgen/db/tests/common.py | has_mongo | chc273/pymatgen-db | python | def has_mongo():
if os.environ.get('MP_FAKEMONGO'):
mongo = False
else:
try:
pymongo.MongoClient()
mongo = True
except:
mongo = False
return mongo |
def connect(self, clear=False):
'Connect to Mongo DB\n\n :return: pymongo Database\n '
c = pymongo.MongoClient()
db = c[self.DB]
if clear:
for coll in (self.SRC, self.DST):
db[coll].remove()
tcoll = ((coll + '.') + CollectionTracker.TRACKING_NAME)
... | 8,094,381,745,248,001,000 | Connect to Mongo DB
:return: pymongo Database | pymatgen/db/tests/common.py | connect | chc273/pymatgen-db | python | def connect(self, clear=False):
'Connect to Mongo DB\n\n :return: pymongo Database\n '
c = pymongo.MongoClient()
db = c[self.DB]
if clear:
for coll in (self.SRC, self.DST):
db[coll].remove()
tcoll = ((coll + '.') + CollectionTracker.TRACKING_NAME)
... |
def run_command(self, args, options):
'Run the command-line given by the list\n in `args`, adding the dictionary given by\n options as long-form --{key}=value pairs.\n '
for (key, value) in options:
args.append('--{}'.format(key))
if value:
args.append(value)
... | 6,931,213,289,327,645,000 | Run the command-line given by the list
in `args`, adding the dictionary given by
options as long-form --{key}=value pairs. | pymatgen/db/tests/common.py | run_command | chc273/pymatgen-db | python | def run_command(self, args, options):
'Run the command-line given by the list\n in `args`, adding the dictionary given by\n options as long-form --{key}=value pairs.\n '
for (key, value) in options:
args.append('--{}'.format(key))
if value:
args.append(value)
... |
def myfunc(self, x, slope, intercept):
'\n input, slope, intercept\n '
return ((slope * x) + intercept) | -5,121,193,954,077,837,000 | input, slope, intercept | helloword/ml.py | myfunc | badpaybad/mypython | python | def myfunc(self, x, slope, intercept):
'\n \n '
return ((slope * x) + intercept) |
async def begin_create_or_update(self, resource_group_name: str, gallery_name: str, gallery_image_name: str, gallery_image: '_models.GalleryImage', **kwargs) -> AsyncLROPoller['_models.GalleryImage']:
"Create or update a gallery Image Definition.\n\n :param resource_group_name: The name of the resource group... | -4,321,324,683,486,622,700 | Create or update a gallery Image Definition.
:param resource_group_name: The name of the resource group.
:type resource_group_name: str
:param gallery_name: The name of the Shared Image Gallery in which the Image Definition is to
be created.
:type gallery_name: str
:param gallery_image_name: The name of the gallery I... | sdk/compute/azure-mgmt-compute/azure/mgmt/compute/v2019_03_01/aio/operations/_gallery_images_operations.py | begin_create_or_update | Darkstar1t/azure-sdk-for-python | python | async def begin_create_or_update(self, resource_group_name: str, gallery_name: str, gallery_image_name: str, gallery_image: '_models.GalleryImage', **kwargs) -> AsyncLROPoller['_models.GalleryImage']:
"Create or update a gallery Image Definition.\n\n :param resource_group_name: The name of the resource group... |
async def get(self, resource_group_name: str, gallery_name: str, gallery_image_name: str, **kwargs) -> '_models.GalleryImage':
'Retrieves information about a gallery Image Definition.\n\n :param resource_group_name: The name of the resource group.\n :type resource_group_name: str\n :param galle... | 5,591,217,699,664,159,000 | Retrieves information about a gallery Image Definition.
:param resource_group_name: The name of the resource group.
:type resource_group_name: str
:param gallery_name: The name of the Shared Image Gallery from which the Image Definitions are
to be retrieved.
:type gallery_name: str
:param gallery_image_name: The name... | sdk/compute/azure-mgmt-compute/azure/mgmt/compute/v2019_03_01/aio/operations/_gallery_images_operations.py | get | Darkstar1t/azure-sdk-for-python | python | async def get(self, resource_group_name: str, gallery_name: str, gallery_image_name: str, **kwargs) -> '_models.GalleryImage':
'Retrieves information about a gallery Image Definition.\n\n :param resource_group_name: The name of the resource group.\n :type resource_group_name: str\n :param galle... |
async def begin_delete(self, resource_group_name: str, gallery_name: str, gallery_image_name: str, **kwargs) -> AsyncLROPoller[None]:
"Delete a gallery image.\n\n :param resource_group_name: The name of the resource group.\n :type resource_group_name: str\n :param gallery_name: The name of the ... | 7,484,213,028,850,654,000 | Delete a gallery image.
:param resource_group_name: The name of the resource group.
:type resource_group_name: str
:param gallery_name: The name of the Shared Image Gallery in which the Image Definition is to
be deleted.
:type gallery_name: str
:param gallery_image_name: The name of the gallery Image Definition to be... | sdk/compute/azure-mgmt-compute/azure/mgmt/compute/v2019_03_01/aio/operations/_gallery_images_operations.py | begin_delete | Darkstar1t/azure-sdk-for-python | python | async def begin_delete(self, resource_group_name: str, gallery_name: str, gallery_image_name: str, **kwargs) -> AsyncLROPoller[None]:
"Delete a gallery image.\n\n :param resource_group_name: The name of the resource group.\n :type resource_group_name: str\n :param gallery_name: The name of the ... |
def list_by_gallery(self, resource_group_name: str, gallery_name: str, **kwargs) -> AsyncIterable['_models.GalleryImageList']:
'List gallery Image Definitions in a gallery.\n\n :param resource_group_name: The name of the resource group.\n :type resource_group_name: str\n :param gallery_name: Th... | -6,598,339,211,277,231,000 | List gallery Image Definitions in a gallery.
:param resource_group_name: The name of the resource group.
:type resource_group_name: str
:param gallery_name: The name of the Shared Image Gallery from which Image Definitions are to
be listed.
:type gallery_name: str
:keyword callable cls: A custom type or function that... | sdk/compute/azure-mgmt-compute/azure/mgmt/compute/v2019_03_01/aio/operations/_gallery_images_operations.py | list_by_gallery | Darkstar1t/azure-sdk-for-python | python | def list_by_gallery(self, resource_group_name: str, gallery_name: str, **kwargs) -> AsyncIterable['_models.GalleryImageList']:
'List gallery Image Definitions in a gallery.\n\n :param resource_group_name: The name of the resource group.\n :type resource_group_name: str\n :param gallery_name: Th... |
def _InternalConstructMessage(full_name):
'Constructs a nested message.'
from google.protobuf import symbol_database
return symbol_database.Default().GetSymbol(full_name)() | -4,080,538,055,633,935,000 | Constructs a nested message. | Plugins/UnrealEnginePython/Binaries/Win64/Lib/site-packages/google/protobuf/message.py | _InternalConstructMessage | JustinACoder/H22-GR3-UnrealAI | python | def _InternalConstructMessage(full_name):
from google.protobuf import symbol_database
return symbol_database.Default().GetSymbol(full_name)() |
def __eq__(self, other_msg):
'Recursively compares two messages by value and structure.'
raise NotImplementedError | 3,488,240,450,900,297,000 | Recursively compares two messages by value and structure. | Plugins/UnrealEnginePython/Binaries/Win64/Lib/site-packages/google/protobuf/message.py | __eq__ | JustinACoder/H22-GR3-UnrealAI | python | def __eq__(self, other_msg):
raise NotImplementedError |
def __str__(self):
'Outputs a human-readable representation of the message.'
raise NotImplementedError | -275,227,890,260,641,820 | Outputs a human-readable representation of the message. | Plugins/UnrealEnginePython/Binaries/Win64/Lib/site-packages/google/protobuf/message.py | __str__ | JustinACoder/H22-GR3-UnrealAI | python | def __str__(self):
raise NotImplementedError |
def __unicode__(self):
'Outputs a human-readable representation of the message.'
raise NotImplementedError | -7,489,656,935,393,816,000 | Outputs a human-readable representation of the message. | Plugins/UnrealEnginePython/Binaries/Win64/Lib/site-packages/google/protobuf/message.py | __unicode__ | JustinACoder/H22-GR3-UnrealAI | python | def __unicode__(self):
raise NotImplementedError |
def MergeFrom(self, other_msg):
'Merges the contents of the specified message into current message.\n\n This method merges the contents of the specified message into the current\n message. Singular fields that are set in the specified message overwrite\n the corresponding fields in the current message. Rep... | -5,259,389,814,106,578,000 | Merges the contents of the specified message into current message.
This method merges the contents of the specified message into the current
message. Singular fields that are set in the specified message overwrite
the corresponding fields in the current message. Repeated fields are
appended. Singular sub-messages and ... | Plugins/UnrealEnginePython/Binaries/Win64/Lib/site-packages/google/protobuf/message.py | MergeFrom | JustinACoder/H22-GR3-UnrealAI | python | def MergeFrom(self, other_msg):
'Merges the contents of the specified message into current message.\n\n This method merges the contents of the specified message into the current\n message. Singular fields that are set in the specified message overwrite\n the corresponding fields in the current message. Rep... |
def CopyFrom(self, other_msg):
'Copies the content of the specified message into the current message.\n\n The method clears the current message and then merges the specified\n message using MergeFrom.\n\n Args:\n other_msg (Message): A message to copy into the current one.\n '
if (self is other... | 3,342,220,506,585,992,000 | Copies the content of the specified message into the current message.
The method clears the current message and then merges the specified
message using MergeFrom.
Args:
other_msg (Message): A message to copy into the current one. | Plugins/UnrealEnginePython/Binaries/Win64/Lib/site-packages/google/protobuf/message.py | CopyFrom | JustinACoder/H22-GR3-UnrealAI | python | def CopyFrom(self, other_msg):
'Copies the content of the specified message into the current message.\n\n The method clears the current message and then merges the specified\n message using MergeFrom.\n\n Args:\n other_msg (Message): A message to copy into the current one.\n '
if (self is other... |
def Clear(self):
'Clears all data that was set in the message.'
raise NotImplementedError | -7,482,799,018,333,739,000 | Clears all data that was set in the message. | Plugins/UnrealEnginePython/Binaries/Win64/Lib/site-packages/google/protobuf/message.py | Clear | JustinACoder/H22-GR3-UnrealAI | python | def Clear(self):
raise NotImplementedError |
def SetInParent(self):
'Mark this as present in the parent.\n\n This normally happens automatically when you assign a field of a\n sub-message, but sometimes you want to make the sub-message\n present while keeping it empty. If you find yourself using this,\n you may want to reconsider your design.\n ... | 1,214,768,362,767,601,200 | Mark this as present in the parent.
This normally happens automatically when you assign a field of a
sub-message, but sometimes you want to make the sub-message
present while keeping it empty. If you find yourself using this,
you may want to reconsider your design. | Plugins/UnrealEnginePython/Binaries/Win64/Lib/site-packages/google/protobuf/message.py | SetInParent | JustinACoder/H22-GR3-UnrealAI | python | def SetInParent(self):
'Mark this as present in the parent.\n\n This normally happens automatically when you assign a field of a\n sub-message, but sometimes you want to make the sub-message\n present while keeping it empty. If you find yourself using this,\n you may want to reconsider your design.\n ... |
def IsInitialized(self):
'Checks if the message is initialized.\n\n Returns:\n bool: The method returns True if the message is initialized (i.e. all of\n its required fields are set).\n '
raise NotImplementedError | -6,005,976,788,694,109,000 | Checks if the message is initialized.
Returns:
bool: The method returns True if the message is initialized (i.e. all of
its required fields are set). | Plugins/UnrealEnginePython/Binaries/Win64/Lib/site-packages/google/protobuf/message.py | IsInitialized | JustinACoder/H22-GR3-UnrealAI | python | def IsInitialized(self):
'Checks if the message is initialized.\n\n Returns:\n bool: The method returns True if the message is initialized (i.e. all of\n its required fields are set).\n '
raise NotImplementedError |
def MergeFromString(self, serialized):
'Merges serialized protocol buffer data into this message.\n\n When we find a field in `serialized` that is already present\n in this message:\n\n - If it\'s a "repeated" field, we append to the end of our list.\n - Else, if it\'s a scalar, we overwrite our fie... | 185,471,447,270,742,600 | Merges serialized protocol buffer data into this message.
When we find a field in `serialized` that is already present
in this message:
- If it's a "repeated" field, we append to the end of our list.
- Else, if it's a scalar, we overwrite our field.
- Else, (it's a nonrepeated composite), we recursively merge
... | Plugins/UnrealEnginePython/Binaries/Win64/Lib/site-packages/google/protobuf/message.py | MergeFromString | JustinACoder/H22-GR3-UnrealAI | python | def MergeFromString(self, serialized):
'Merges serialized protocol buffer data into this message.\n\n When we find a field in `serialized` that is already present\n in this message:\n\n - If it\'s a "repeated" field, we append to the end of our list.\n - Else, if it\'s a scalar, we overwrite our fie... |
def ParseFromString(self, serialized):
'Parse serialized protocol buffer data into this message.\n\n Like :func:`MergeFromString()`, except we clear the object first.\n '
self.Clear()
return self.MergeFromString(serialized) | -5,109,456,638,149,453,000 | Parse serialized protocol buffer data into this message.
Like :func:`MergeFromString()`, except we clear the object first. | Plugins/UnrealEnginePython/Binaries/Win64/Lib/site-packages/google/protobuf/message.py | ParseFromString | JustinACoder/H22-GR3-UnrealAI | python | def ParseFromString(self, serialized):
'Parse serialized protocol buffer data into this message.\n\n Like :func:`MergeFromString()`, except we clear the object first.\n '
self.Clear()
return self.MergeFromString(serialized) |
def SerializeToString(self, **kwargs):
"Serializes the protocol message to a binary string.\n\n Keyword Args:\n deterministic (bool): If true, requests deterministic serialization\n of the protobuf, with predictable ordering of map keys.\n\n Returns:\n A binary string representation of the me... | 7,587,470,124,439,645,000 | Serializes the protocol message to a binary string.
Keyword Args:
deterministic (bool): If true, requests deterministic serialization
of the protobuf, with predictable ordering of map keys.
Returns:
A binary string representation of the message if all of the required
fields in the message are set (i.e. the ... | Plugins/UnrealEnginePython/Binaries/Win64/Lib/site-packages/google/protobuf/message.py | SerializeToString | JustinACoder/H22-GR3-UnrealAI | python | def SerializeToString(self, **kwargs):
"Serializes the protocol message to a binary string.\n\n Keyword Args:\n deterministic (bool): If true, requests deterministic serialization\n of the protobuf, with predictable ordering of map keys.\n\n Returns:\n A binary string representation of the me... |
def SerializePartialToString(self, **kwargs):
"Serializes the protocol message to a binary string.\n\n This method is similar to SerializeToString but doesn't check if the\n message is initialized.\n\n Keyword Args:\n deterministic (bool): If true, requests deterministic serialization\n of the ... | -5,496,072,132,301,760,000 | Serializes the protocol message to a binary string.
This method is similar to SerializeToString but doesn't check if the
message is initialized.
Keyword Args:
deterministic (bool): If true, requests deterministic serialization
of the protobuf, with predictable ordering of map keys.
Returns:
bytes: A serializ... | Plugins/UnrealEnginePython/Binaries/Win64/Lib/site-packages/google/protobuf/message.py | SerializePartialToString | JustinACoder/H22-GR3-UnrealAI | python | def SerializePartialToString(self, **kwargs):
"Serializes the protocol message to a binary string.\n\n This method is similar to SerializeToString but doesn't check if the\n message is initialized.\n\n Keyword Args:\n deterministic (bool): If true, requests deterministic serialization\n of the ... |
def ListFields(self):
'Returns a list of (FieldDescriptor, value) tuples for present fields.\n\n A message field is non-empty if HasField() would return true. A singular\n primitive field is non-empty if HasField() would return true in proto2 or it\n is non zero in proto3. A repeated field is non-empty if ... | 7,323,349,097,671,602,000 | Returns a list of (FieldDescriptor, value) tuples for present fields.
A message field is non-empty if HasField() would return true. A singular
primitive field is non-empty if HasField() would return true in proto2 or it
is non zero in proto3. A repeated field is non-empty if it contains at least
one element. The field... | Plugins/UnrealEnginePython/Binaries/Win64/Lib/site-packages/google/protobuf/message.py | ListFields | JustinACoder/H22-GR3-UnrealAI | python | def ListFields(self):
'Returns a list of (FieldDescriptor, value) tuples for present fields.\n\n A message field is non-empty if HasField() would return true. A singular\n primitive field is non-empty if HasField() would return true in proto2 or it\n is non zero in proto3. A repeated field is non-empty if ... |
def HasField(self, field_name):
'Checks if a certain field is set for the message.\n\n For a oneof group, checks if any field inside is set. Note that if the\n field_name is not defined in the message descriptor, :exc:`ValueError` will\n be raised.\n\n Args:\n field_name (str): The name of the fiel... | 221,908,891,018,159,100 | Checks if a certain field is set for the message.
For a oneof group, checks if any field inside is set. Note that if the
field_name is not defined in the message descriptor, :exc:`ValueError` will
be raised.
Args:
field_name (str): The name of the field to check for presence.
Returns:
bool: Whether a value has b... | Plugins/UnrealEnginePython/Binaries/Win64/Lib/site-packages/google/protobuf/message.py | HasField | JustinACoder/H22-GR3-UnrealAI | python | def HasField(self, field_name):
'Checks if a certain field is set for the message.\n\n For a oneof group, checks if any field inside is set. Note that if the\n field_name is not defined in the message descriptor, :exc:`ValueError` will\n be raised.\n\n Args:\n field_name (str): The name of the fiel... |
def ClearField(self, field_name):
'Clears the contents of a given field.\n\n Inside a oneof group, clears the field set. If the name neither refers to a\n defined field or oneof group, :exc:`ValueError` is raised.\n\n Args:\n field_name (str): The name of the field to check for presence.\n\n Raises... | -4,180,667,636,435,284,500 | Clears the contents of a given field.
Inside a oneof group, clears the field set. If the name neither refers to a
defined field or oneof group, :exc:`ValueError` is raised.
Args:
field_name (str): The name of the field to check for presence.
Raises:
ValueError: if the `field_name` is not a member of this message... | Plugins/UnrealEnginePython/Binaries/Win64/Lib/site-packages/google/protobuf/message.py | ClearField | JustinACoder/H22-GR3-UnrealAI | python | def ClearField(self, field_name):
'Clears the contents of a given field.\n\n Inside a oneof group, clears the field set. If the name neither refers to a\n defined field or oneof group, :exc:`ValueError` is raised.\n\n Args:\n field_name (str): The name of the field to check for presence.\n\n Raises... |
def WhichOneof(self, oneof_group):
'Returns the name of the field that is set inside a oneof group.\n\n If no field is set, returns None.\n\n Args:\n oneof_group (str): the name of the oneof group to check.\n\n Returns:\n str or None: The name of the group that is set, or None.\n\n Raises:\n ... | 7,706,930,664,240,627,000 | Returns the name of the field that is set inside a oneof group.
If no field is set, returns None.
Args:
oneof_group (str): the name of the oneof group to check.
Returns:
str or None: The name of the group that is set, or None.
Raises:
ValueError: no group with the given name exists | Plugins/UnrealEnginePython/Binaries/Win64/Lib/site-packages/google/protobuf/message.py | WhichOneof | JustinACoder/H22-GR3-UnrealAI | python | def WhichOneof(self, oneof_group):
'Returns the name of the field that is set inside a oneof group.\n\n If no field is set, returns None.\n\n Args:\n oneof_group (str): the name of the oneof group to check.\n\n Returns:\n str or None: The name of the group that is set, or None.\n\n Raises:\n ... |
def HasExtension(self, extension_handle):
'Checks if a certain extension is present for this message.\n\n Extensions are retrieved using the :attr:`Extensions` mapping (if present).\n\n Args:\n extension_handle: The handle for the extension to check.\n\n Returns:\n bool: Whether the extension is ... | 4,907,608,339,312,964,000 | Checks if a certain extension is present for this message.
Extensions are retrieved using the :attr:`Extensions` mapping (if present).
Args:
extension_handle: The handle for the extension to check.
Returns:
bool: Whether the extension is present for this message.
Raises:
KeyError: if the extension is repeated... | Plugins/UnrealEnginePython/Binaries/Win64/Lib/site-packages/google/protobuf/message.py | HasExtension | JustinACoder/H22-GR3-UnrealAI | python | def HasExtension(self, extension_handle):
'Checks if a certain extension is present for this message.\n\n Extensions are retrieved using the :attr:`Extensions` mapping (if present).\n\n Args:\n extension_handle: The handle for the extension to check.\n\n Returns:\n bool: Whether the extension is ... |
def ClearExtension(self, extension_handle):
'Clears the contents of a given extension.\n\n Args:\n extension_handle: The handle for the extension to clear.\n '
raise NotImplementedError | -8,967,653,825,512,342,000 | Clears the contents of a given extension.
Args:
extension_handle: The handle for the extension to clear. | Plugins/UnrealEnginePython/Binaries/Win64/Lib/site-packages/google/protobuf/message.py | ClearExtension | JustinACoder/H22-GR3-UnrealAI | python | def ClearExtension(self, extension_handle):
'Clears the contents of a given extension.\n\n Args:\n extension_handle: The handle for the extension to clear.\n '
raise NotImplementedError |
def UnknownFields(self):
'Returns the UnknownFieldSet.\n\n Returns:\n UnknownFieldSet: The unknown fields stored in this message.\n '
raise NotImplementedError | 5,739,973,485,065,129,000 | Returns the UnknownFieldSet.
Returns:
UnknownFieldSet: The unknown fields stored in this message. | Plugins/UnrealEnginePython/Binaries/Win64/Lib/site-packages/google/protobuf/message.py | UnknownFields | JustinACoder/H22-GR3-UnrealAI | python | def UnknownFields(self):
'Returns the UnknownFieldSet.\n\n Returns:\n UnknownFieldSet: The unknown fields stored in this message.\n '
raise NotImplementedError |
def DiscardUnknownFields(self):
'Clears all fields in the :class:`UnknownFieldSet`.\n\n This operation is recursive for nested message.\n '
raise NotImplementedError | 2,413,369,931,560,882,700 | Clears all fields in the :class:`UnknownFieldSet`.
This operation is recursive for nested message. | Plugins/UnrealEnginePython/Binaries/Win64/Lib/site-packages/google/protobuf/message.py | DiscardUnknownFields | JustinACoder/H22-GR3-UnrealAI | python | def DiscardUnknownFields(self):
'Clears all fields in the :class:`UnknownFieldSet`.\n\n This operation is recursive for nested message.\n '
raise NotImplementedError |
def ByteSize(self):
'Returns the serialized size of this message.\n\n Recursively calls ByteSize() on all contained messages.\n\n Returns:\n int: The number of bytes required to serialize this message.\n '
raise NotImplementedError | -7,189,590,014,356,224,000 | Returns the serialized size of this message.
Recursively calls ByteSize() on all contained messages.
Returns:
int: The number of bytes required to serialize this message. | Plugins/UnrealEnginePython/Binaries/Win64/Lib/site-packages/google/protobuf/message.py | ByteSize | JustinACoder/H22-GR3-UnrealAI | python | def ByteSize(self):
'Returns the serialized size of this message.\n\n Recursively calls ByteSize() on all contained messages.\n\n Returns:\n int: The number of bytes required to serialize this message.\n '
raise NotImplementedError |
def _SetListener(self, message_listener):
'Internal method used by the protocol message implementation.\n Clients should not call this directly.\n\n Sets a listener that this message will call on certain state transitions.\n\n The purpose of this method is to register back-edges from children to\n paren... | 2,148,599,048,241,910,500 | Internal method used by the protocol message implementation.
Clients should not call this directly.
Sets a listener that this message will call on certain state transitions.
The purpose of this method is to register back-edges from children to
parents at runtime, for the purpose of setting "has" bits and
byte-size-di... | Plugins/UnrealEnginePython/Binaries/Win64/Lib/site-packages/google/protobuf/message.py | _SetListener | JustinACoder/H22-GR3-UnrealAI | python | def _SetListener(self, message_listener):
'Internal method used by the protocol message implementation.\n Clients should not call this directly.\n\n Sets a listener that this message will call on certain state transitions.\n\n The purpose of this method is to register back-edges from children to\n paren... |
def __getstate__(self):
'Support the pickle protocol.'
return dict(serialized=self.SerializePartialToString()) | 5,659,051,950,600,202,000 | Support the pickle protocol. | Plugins/UnrealEnginePython/Binaries/Win64/Lib/site-packages/google/protobuf/message.py | __getstate__ | JustinACoder/H22-GR3-UnrealAI | python | def __getstate__(self):
return dict(serialized=self.SerializePartialToString()) |
def __setstate__(self, state):
'Support the pickle protocol.'
self.__init__()
serialized = state['serialized']
if (not isinstance(serialized, bytes)):
serialized = serialized.encode('latin1')
self.ParseFromString(serialized) | -2,837,338,939,880,771,000 | Support the pickle protocol. | Plugins/UnrealEnginePython/Binaries/Win64/Lib/site-packages/google/protobuf/message.py | __setstate__ | JustinACoder/H22-GR3-UnrealAI | python | def __setstate__(self, state):
self.__init__()
serialized = state['serialized']
if (not isinstance(serialized, bytes)):
serialized = serialized.encode('latin1')
self.ParseFromString(serialized) |
def train(model: DeepMoD, data: torch.Tensor, target: torch.Tensor, optimizer, sparsity_scheduler, test='mse', split: float=0.8, log_dir: Optional[str]=None, max_iterations: int=10000, write_iterations: int=25, **convergence_kwargs) -> None:
'[summary]\n\n Args:\n model (DeepMoD): [description]\n d... | -7,086,159,953,085,371,000 | [summary]
Args:
model (DeepMoD): [description]
data (torch.Tensor): [description]
target (torch.Tensor): [description]
optimizer ([type]): [description]
sparsity_scheduler ([type]): [description]
log_dir (Optional[str], optional): [description]. Defaults to None.
max_iterations (int, option... | src/multitaskpinn/training/training.py | train | GJBoth/MultiTaskPINN | python | def train(model: DeepMoD, data: torch.Tensor, target: torch.Tensor, optimizer, sparsity_scheduler, test='mse', split: float=0.8, log_dir: Optional[str]=None, max_iterations: int=10000, write_iterations: int=25, **convergence_kwargs) -> None:
'[summary]\n\n Args:\n model (DeepMoD): [description]\n d... |
def train_multitask(model: DeepMoD, data: torch.Tensor, target: torch.Tensor, optimizer, sparsity_scheduler, test='mse', split: float=0.8, log_dir: Optional[str]=None, max_iterations: int=10000, write_iterations: int=25, **convergence_kwargs) -> None:
'[summary]\n\n Args:\n model (DeepMoD): [description]\... | -735,518,995,854,886,700 | [summary]
Args:
model (DeepMoD): [description]
data (torch.Tensor): [description]
target (torch.Tensor): [description]
optimizer ([type]): [description]
sparsity_scheduler ([type]): [description]
log_dir (Optional[str], optional): [description]. Defaults to None.
max_iterations (int, option... | src/multitaskpinn/training/training.py | train_multitask | GJBoth/MultiTaskPINN | python | def train_multitask(model: DeepMoD, data: torch.Tensor, target: torch.Tensor, optimizer, sparsity_scheduler, test='mse', split: float=0.8, log_dir: Optional[str]=None, max_iterations: int=10000, write_iterations: int=25, **convergence_kwargs) -> None:
'[summary]\n\n Args:\n model (DeepMoD): [description]\... |
def train_multitask_capped(model: DeepMoD, data: torch.Tensor, target: torch.Tensor, optimizer, sparsity_scheduler, test='mse', split: float=0.8, log_dir: Optional[str]=None, max_iterations: int=10000, write_iterations: int=25, **convergence_kwargs) -> None:
'[summary]\n\n Args:\n model (DeepMoD): [descri... | 3,808,070,609,708,569,000 | [summary]
Args:
model (DeepMoD): [description]
data (torch.Tensor): [description]
target (torch.Tensor): [description]
optimizer ([type]): [description]
sparsity_scheduler ([type]): [description]
log_dir (Optional[str], optional): [description]. Defaults to None.
max_iterations (int, option... | src/multitaskpinn/training/training.py | train_multitask_capped | GJBoth/MultiTaskPINN | python | def train_multitask_capped(model: DeepMoD, data: torch.Tensor, target: torch.Tensor, optimizer, sparsity_scheduler, test='mse', split: float=0.8, log_dir: Optional[str]=None, max_iterations: int=10000, write_iterations: int=25, **convergence_kwargs) -> None:
'[summary]\n\n Args:\n model (DeepMoD): [descri... |
def train_gradnorm(model: DeepMoD, data: torch.Tensor, target: torch.Tensor, optimizer, sparsity_scheduler, alpha, test='mse', split: float=0.8, log_dir: Optional[str]=None, max_iterations: int=10000, write_iterations: int=25, **convergence_kwargs) -> None:
'[summary]\n\n Args:\n model (DeepMoD): [descrip... | 6,851,036,863,452,085,000 | [summary]
Args:
model (DeepMoD): [description]
data (torch.Tensor): [description]
target (torch.Tensor): [description]
optimizer ([type]): [description]
sparsity_scheduler ([type]): [description]
log_dir (Optional[str], optional): [description]. Defaults to None.
max_iterations (int, option... | src/multitaskpinn/training/training.py | train_gradnorm | GJBoth/MultiTaskPINN | python | def train_gradnorm(model: DeepMoD, data: torch.Tensor, target: torch.Tensor, optimizer, sparsity_scheduler, alpha, test='mse', split: float=0.8, log_dir: Optional[str]=None, max_iterations: int=10000, write_iterations: int=25, **convergence_kwargs) -> None:
'[summary]\n\n Args:\n model (DeepMoD): [descrip... |
def train_SBL(model: DeepMoD, data: torch.Tensor, target: torch.Tensor, optimizer, extra_params, sparsity_scheduler, split=0.8, exp_ID: str=None, log_dir: str=None, max_iterations: int=10000, write_iterations: int=25, **convergence_kwargs) -> None:
'Trains the DeepMoD model. This function automatically splits the d... | -1,715,964,040,550,835,700 | Trains the DeepMoD model. This function automatically splits the data set in a train and test set.
Args:
model (DeepMoD): A DeepMoD object.
data (torch.Tensor): Tensor of shape (n_samples x (n_spatial + 1)) containing the coordinates, first column should be the time coordinate.
target (torch.Tensor): Te... | src/multitaskpinn/training/training.py | train_SBL | GJBoth/MultiTaskPINN | python | def train_SBL(model: DeepMoD, data: torch.Tensor, target: torch.Tensor, optimizer, extra_params, sparsity_scheduler, split=0.8, exp_ID: str=None, log_dir: str=None, max_iterations: int=10000, write_iterations: int=25, **convergence_kwargs) -> None:
'Trains the DeepMoD model. This function automatically splits the d... |
def decoderawtransaction_asm_sighashtype(self):
'Test decoding scripts via RPC command "decoderawtransaction".\n\n This test is in with the "decodescript" tests because they are testing the same "asm" script decodes.\n '
tx = '0100000001696a20784a2c70143f634e95227dbdfdf0ecd51647052e70854512235f598... | 3,089,507,438,065,855,000 | Test decoding scripts via RPC command "decoderawtransaction".
This test is in with the "decodescript" tests because they are testing the same "asm" script decodes. | test/functional/rpc_decodescript.py | decoderawtransaction_asm_sighashtype | Black-NET/erexcoin-source | python | def decoderawtransaction_asm_sighashtype(self):
'Test decoding scripts via RPC command "decoderawtransaction".\n\n This test is in with the "decodescript" tests because they are testing the same "asm" script decodes.\n '
tx = '0100000001696a20784a2c70143f634e95227dbdfdf0ecd51647052e70854512235f598... |
def create_dynamic_ordering_constraint(index: int) -> str:
'\n Creates a valid AMPL constraint of the form:\n [LaTex]: $start\\_time_j+1 >= start\\_time_j + duration_j$, $\x0corall j \\in BATCH$\n :param index: j index where the current constraint should start\n :return: single AMPL JIT constraint as a ... | 6,747,510,227,925,415,000 | Creates a valid AMPL constraint of the form:
[LaTex]: $start\_time_j+1 >= start\_time_j + duration_j$, $orall j \in BATCH$
:param index: j index where the current constraint should start
:return: single AMPL JIT constraint as a string | ampljit/utils.py | create_dynamic_ordering_constraint | jkomyno/amplrestapi | python | def create_dynamic_ordering_constraint(index: int) -> str:
'\n Creates a valid AMPL constraint of the form:\n [LaTex]: $start\\_time_j+1 >= start\\_time_j + duration_j$, $\x0corall j \\in BATCH$\n :param index: j index where the current constraint should start\n :return: single AMPL JIT constraint as a ... |
def dict_to_list(obj: dict) -> list:
"\n Converts a dictionary to a list, extracting the values of the dictionary.\n The list is sorted according to the dict's keys ascendant order.\n The given dictionary should always have the same numeric keys as the result of create_batch_dictionary().\n :param obj: ... | 9,012,645,828,889,174,000 | Converts a dictionary to a list, extracting the values of the dictionary.
The list is sorted according to the dict's keys ascendant order.
The given dictionary should always have the same numeric keys as the result of create_batch_dictionary().
:param obj: the dictionary to convert which should have numeric keys
:retur... | ampljit/utils.py | dict_to_list | jkomyno/amplrestapi | python | def dict_to_list(obj: dict) -> list:
"\n Converts a dictionary to a list, extracting the values of the dictionary.\n The list is sorted according to the dict's keys ascendant order.\n The given dictionary should always have the same numeric keys as the result of create_batch_dictionary().\n :param obj: ... |
def strings_to_datetimes(str_date_lst: list, datetime_format: str) -> list:
'\n Converts a list of strings into a list of datetime objects\n :param str_date_lst: list of string objects compatible with the ISO8601 format\n :param datetime_format: format of the datetime\n :return: list of datetime objects... | -3,893,893,176,523,418,600 | Converts a list of strings into a list of datetime objects
:param str_date_lst: list of string objects compatible with the ISO8601 format
:param datetime_format: format of the datetime
:return: list of datetime objects equivalent to the given str_date_lst | ampljit/utils.py | strings_to_datetimes | jkomyno/amplrestapi | python | def strings_to_datetimes(str_date_lst: list, datetime_format: str) -> list:
'\n Converts a list of strings into a list of datetime objects\n :param str_date_lst: list of string objects compatible with the ISO8601 format\n :param datetime_format: format of the datetime\n :return: list of datetime objects... |
def minute_timedelta(first: datetime, second: datetime) -> int:
'\n Returns the difference expressed in minutes between 2 datetime objects\n :param first: datetime object that comes before second\n :param second: datetime object that comes after first\n :return: difference in minutes between second and ... | 5,728,973,006,739,140,000 | Returns the difference expressed in minutes between 2 datetime objects
:param first: datetime object that comes before second
:param second: datetime object that comes after first
:return: difference in minutes between second and first | ampljit/utils.py | minute_timedelta | jkomyno/amplrestapi | python | def minute_timedelta(first: datetime, second: datetime) -> int:
'\n Returns the difference expressed in minutes between 2 datetime objects\n :param first: datetime object that comes before second\n :param second: datetime object that comes after first\n :return: difference in minutes between second and ... |
def minute_timedeltas_wrt_first(datetime_lst: list) -> list:
"\n Converts a list of datetime objects into a list of minute time deltas with respect to the first item.\n For example, given the input datetime_lst:\n [\n '2019-08-22 14:32',\n '2019-08-22 14:38',\n '2019-08-22 14:42',\n ... | 4,508,611,415,876,424,000 | Converts a list of datetime objects into a list of minute time deltas with respect to the first item.
For example, given the input datetime_lst:
[
'2019-08-22 14:32',
'2019-08-22 14:38',
'2019-08-22 14:42',
'2019-08-22 14:52',
'2019-08-22 14:57'
],
the result would be:
[32, 38, 42, 52, 57]
:param d... | ampljit/utils.py | minute_timedeltas_wrt_first | jkomyno/amplrestapi | python | def minute_timedeltas_wrt_first(datetime_lst: list) -> list:
"\n Converts a list of datetime objects into a list of minute time deltas with respect to the first item.\n For example, given the input datetime_lst:\n [\n '2019-08-22 14:32',\n '2019-08-22 14:38',\n '2019-08-22 14:42',\n ... |
def set_minutes_to_datetimes(datetime_lst: list, minutes_lst: list) -> list:
'\n Given a list of minutes and datetime objects, sets each amount of minutes to each datetime object with respect\n to the list index. The two lists must have the same size.\n :param datetime_lst: list of datetime objects\n :p... | -1,701,486,862,418,520,600 | Given a list of minutes and datetime objects, sets each amount of minutes to each datetime object with respect
to the list index. The two lists must have the same size.
:param datetime_lst: list of datetime objects
:param minutes_lst: list of minutes to set to a list of datetime objects
:return: list of datetime object... | ampljit/utils.py | set_minutes_to_datetimes | jkomyno/amplrestapi | python | def set_minutes_to_datetimes(datetime_lst: list, minutes_lst: list) -> list:
'\n Given a list of minutes and datetime objects, sets each amount of minutes to each datetime object with respect\n to the list index. The two lists must have the same size.\n :param datetime_lst: list of datetime objects\n :p... |
def datetimes_to_strings(datetime_lst: list, datetime_format: str) -> list:
'\n Converts a list of datetime objects to strings, according to a certain datetime format.\n :param datetime_lst: list of datetime objects to convert to string\n :param datetime_format: format of the datetime\n :return: the lis... | 213,313,511,449,219,460 | Converts a list of datetime objects to strings, according to a certain datetime format.
:param datetime_lst: list of datetime objects to convert to string
:param datetime_format: format of the datetime
:return: the list of datetime objects converted to strings in the given datetime format | ampljit/utils.py | datetimes_to_strings | jkomyno/amplrestapi | python | def datetimes_to_strings(datetime_lst: list, datetime_format: str) -> list:
'\n Converts a list of datetime objects to strings, according to a certain datetime format.\n :param datetime_lst: list of datetime objects to convert to string\n :param datetime_format: format of the datetime\n :return: the lis... |
def test_len(self):
' test for boostrap helper'
for i in range(len(test_cases)):
self.assertEqual(len(results[i]), test_cases[i]['len'])
self.assertEqual(len(resultsInt64[i]), test_cases[i]['len']) | 7,617,248,818,454,541,000 | test for boostrap helper | neaps-api/neaps_lib/bootstrap_test.py | test_len | ExpediaGroup/neaps | python | def test_len(self):
' '
for i in range(len(test_cases)):
self.assertEqual(len(results[i]), test_cases[i]['len'])
self.assertEqual(len(resultsInt64[i]), test_cases[i]['len']) |
def test_value(self):
' docstring '
for i in range(len(test_cases)):
for j in range(len(results[i])):
self.assertEqual(results[i][j], test_cases[i]['mean'])
self.assertEqual(resultsInt64[i][j], test_cases[i]['meanInt64'])
self.assertIsInstance(results[i][j], np.float6... | -7,203,681,843,411,713,000 | docstring | neaps-api/neaps_lib/bootstrap_test.py | test_value | ExpediaGroup/neaps | python | def test_value(self):
' '
for i in range(len(test_cases)):
for j in range(len(results[i])):
self.assertEqual(results[i][j], test_cases[i]['mean'])
self.assertEqual(resultsInt64[i][j], test_cases[i]['meanInt64'])
self.assertIsInstance(results[i][j], np.float64)
... |
def test_less(self):
' docstring '
for i in range(len(test_cases)):
for j in range(len(results[i])):
self.assertLessEqual(results[i][j], max(test_cases[i]['sample'])) | -198,653,698,443,766,720 | docstring | neaps-api/neaps_lib/bootstrap_test.py | test_less | ExpediaGroup/neaps | python | def test_less(self):
' '
for i in range(len(test_cases)):
for j in range(len(results[i])):
self.assertLessEqual(results[i][j], max(test_cases[i]['sample'])) |
def test_greater(self):
' docstring '
for i in range(len(test_cases)):
for j in range(len(results[i])):
self.assertGreaterEqual(results[i][j], min(test_cases[i]['sample'])) | 9,088,840,391,431,958,000 | docstring | neaps-api/neaps_lib/bootstrap_test.py | test_greater | ExpediaGroup/neaps | python | def test_greater(self):
' '
for i in range(len(test_cases)):
for j in range(len(results[i])):
self.assertGreaterEqual(results[i][j], min(test_cases[i]['sample'])) |
def __init__(self, authority=None, content=None, default_port=None, file=None, host=None, path=None, port=None, protocol=None, query=None, ref=None, user_info=None):
'URL - a model defined in Swagger'
self._authority = None
self._content = None
self._default_port = None
self._file = None
self._h... | -4,837,812,409,801,342,000 | URL - a model defined in Swagger | tb_rest_client/models/models_pe/url.py | __init__ | CSTC-WTCB-BBRI/python_tb_rest_client | python | def __init__(self, authority=None, content=None, default_port=None, file=None, host=None, path=None, port=None, protocol=None, query=None, ref=None, user_info=None):
self._authority = None
self._content = None
self._default_port = None
self._file = None
self._host = None
self._path = None
... |
@property
def authority(self):
'Gets the authority of this URL. # noqa: E501\n\n\n :return: The authority of this URL. # noqa: E501\n :rtype: str\n '
return self._authority | 6,008,489,197,909,201,000 | Gets the authority of this URL. # noqa: E501
:return: The authority of this URL. # noqa: E501
:rtype: str | tb_rest_client/models/models_pe/url.py | authority | CSTC-WTCB-BBRI/python_tb_rest_client | python | @property
def authority(self):
'Gets the authority of this URL. # noqa: E501\n\n\n :return: The authority of this URL. # noqa: E501\n :rtype: str\n '
return self._authority |
@authority.setter
def authority(self, authority):
'Sets the authority of this URL.\n\n\n :param authority: The authority of this URL. # noqa: E501\n :type: str\n '
self._authority = authority | 5,373,869,373,672,434,000 | Sets the authority of this URL.
:param authority: The authority of this URL. # noqa: E501
:type: str | tb_rest_client/models/models_pe/url.py | authority | CSTC-WTCB-BBRI/python_tb_rest_client | python | @authority.setter
def authority(self, authority):
'Sets the authority of this URL.\n\n\n :param authority: The authority of this URL. # noqa: E501\n :type: str\n '
self._authority = authority |
@property
def content(self):
'Gets the content of this URL. # noqa: E501\n\n\n :return: The content of this URL. # noqa: E501\n :rtype: object\n '
return self._content | -5,395,410,704,641,924,000 | Gets the content of this URL. # noqa: E501
:return: The content of this URL. # noqa: E501
:rtype: object | tb_rest_client/models/models_pe/url.py | content | CSTC-WTCB-BBRI/python_tb_rest_client | python | @property
def content(self):
'Gets the content of this URL. # noqa: E501\n\n\n :return: The content of this URL. # noqa: E501\n :rtype: object\n '
return self._content |
@content.setter
def content(self, content):
'Sets the content of this URL.\n\n\n :param content: The content of this URL. # noqa: E501\n :type: object\n '
self._content = content | -3,132,387,909,026,987,500 | Sets the content of this URL.
:param content: The content of this URL. # noqa: E501
:type: object | tb_rest_client/models/models_pe/url.py | content | CSTC-WTCB-BBRI/python_tb_rest_client | python | @content.setter
def content(self, content):
'Sets the content of this URL.\n\n\n :param content: The content of this URL. # noqa: E501\n :type: object\n '
self._content = content |
@property
def default_port(self):
'Gets the default_port of this URL. # noqa: E501\n\n\n :return: The default_port of this URL. # noqa: E501\n :rtype: int\n '
return self._default_port | -2,337,253,307,847,136,000 | Gets the default_port of this URL. # noqa: E501
:return: The default_port of this URL. # noqa: E501
:rtype: int | tb_rest_client/models/models_pe/url.py | default_port | CSTC-WTCB-BBRI/python_tb_rest_client | python | @property
def default_port(self):
'Gets the default_port of this URL. # noqa: E501\n\n\n :return: The default_port of this URL. # noqa: E501\n :rtype: int\n '
return self._default_port |
@default_port.setter
def default_port(self, default_port):
'Sets the default_port of this URL.\n\n\n :param default_port: The default_port of this URL. # noqa: E501\n :type: int\n '
self._default_port = default_port | 1,449,772,015,788,586,500 | Sets the default_port of this URL.
:param default_port: The default_port of this URL. # noqa: E501
:type: int | tb_rest_client/models/models_pe/url.py | default_port | CSTC-WTCB-BBRI/python_tb_rest_client | python | @default_port.setter
def default_port(self, default_port):
'Sets the default_port of this URL.\n\n\n :param default_port: The default_port of this URL. # noqa: E501\n :type: int\n '
self._default_port = default_port |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.