Skip to content

Commit a757506

Browse files
authored
1 parent dc36ec1 commit a757506

27 files changed

+38
-38
lines changed

docs/IE_PLUGIN_DG/AsyncInferRequest.md

+1-1
Original file line numberDiff line numberDiff line change
@@ -33,7 +33,7 @@ AsyncInferRequest()
3333

3434
The main goal of the ``AsyncInferRequest`` constructor is to define a device pipeline ``m_pipeline``. The example below demonstrates ``m_pipeline`` creation with the following stages:
3535

36-
* ``infer_preprocess_and_start_pipeline`` is a CPU ligthweight task to submit tasks to a remote device.
36+
* ``infer_preprocess_and_start_pipeline`` is a CPU lightweight task to submit tasks to a remote device.
3737
* ``wait_pipeline`` is a CPU non-compute task that waits for a response from a remote device.
3838
* ``infer_postprocess`` is a CPU compute task.
3939

docs/MO_DG/IR_and_opsets.md

+2-2
Original file line numberDiff line numberDiff line change
@@ -19,7 +19,7 @@ Overview of Artificial Neural Networks Representation
1919
A deep learning network is usually represented as a directed graph describing the flow of data from the network input data to the inference results.
2020
Input data can be in the form of images, video, text, audio, or preprocessed information representing objects from the target area of interest.
2121

22-
Here is an illustration sof a small graph representing a model that consists of a single Convolutional layer and activation function:
22+
Here is an illustration of a small graph representing a model that consists of a single Convolutional layer and activation function:
2323

2424
.. image:: _static/images/small_IR_graph_demonstration.png
2525

@@ -52,7 +52,7 @@ A set consists of several groups of operations:
5252

5353
* Generic element-wise arithmetic tensor operations such as ``Add``, ``Subtract``, and ``Multiply``.
5454

55-
* Comparison operations that compare two numeric tensors and produce boolean tensors, for example, ``Less``, ``Equeal``, ``Greater``.
55+
* Comparison operations that compare two numeric tensors and produce boolean tensors, for example, ``Less``, ``Equal``, ``Greater``.
5656

5757
* Logical operations that are dealing with boolean tensors, for example, ``And``, ``Xor``, ``Not``.
5858

docs/MO_DG/prepare_model/Getting_performance_numbers.md

+1-1
Original file line numberDiff line numberDiff line change
@@ -128,7 +128,7 @@ Information about layer precision is also stored in the performance counters.
128128
resnet\_model/add\_5/fq\_input\_1 NOT\_RUN FakeQuantize undef 0 0
129129
=========================================================== ============= ============== ===================== ================= ==============
130130

131-
| The ``exeStatus`` column of the table includes the following possible values:
131+
| The ``execStatus`` column of the table includes the following possible values:
132132
| - ``EXECUTED`` - the layer was executed by standalone primitive.
133133
| - ``NOT_RUN`` - the layer was not executed by standalone primitive or was fused with another operation and executed in another layer primitive.
134134
|

docs/MO_DG/prepare_model/convert_model/pytorch_specific/Convert_YOLACT.md

+1-1
Original file line numberDiff line numberDiff line change
@@ -111,7 +111,7 @@ The patch modifies the framework code by adding a special command-line argument
111111
+ else:
112112
+ state_dict = torch.load(path, map_location=torch.device('cpu'))
113113

114-
# For backward compatability, remove these (the new variable is called layers)
114+
# For backward compatibility, remove these (the new variable is called layers)
115115
for key in list(state_dict.keys()):
116116
@@ -673,8 +679,11 @@ class Yolact(nn.Module):
117117
else:

docs/MO_DG/prepare_model/convert_model/tf_specific/Convert_GNMT_From_Tensorflow.md

+1-1
Original file line numberDiff line numberDiff line change
@@ -149,7 +149,7 @@ Converting a GNMT Model to the IR
149149

150150
**Step 1**. Clone the GitHub repository and check out the commit:
151151

152-
1. Clone the NMT reposirory:
152+
1. Clone the NMT repository:
153153

154154
.. code-block:: sh
155155

docs/OV_Runtime_UG/supported_plugins/CPU.md

+1-1
Original file line numberDiff line numberDiff line change
@@ -461,7 +461,7 @@ weights are loaded from DDR/L3 cache in the packed format this significantly dec
461461
and as a consequence improve inference performance.
462462

463463
To use this feature, the user is provided with property ``sparse_weights_decompression_rate``, which can take
464-
values from the interval \[0, 1\]. ``sparse_weights_decompression_rate`` defines sparse rate threashold: only operations
464+
values from the interval \[0, 1\]. ``sparse_weights_decompression_rate`` defines sparse rate threshold: only operations
465465
with higher sparse rate will be executed using ``sparse weights decompression feature``. The default value is ``1``,
466466
which means the option is disabled.
467467

docs/OV_Runtime_UG/supported_plugins/config_properties.md

+1-1
Original file line numberDiff line numberDiff line change
@@ -142,7 +142,7 @@ A returned value appears as follows: ``Intel(R) Core(TM) i7-8700 CPU @ 3.20GHz``
142142

143143
In order to understand a list of supported properties on ``ov::Core`` or ``ov::CompiledModel`` levels, use ``ov::supported_properties``
144144
which contains a vector of supported property names. Properties which can be changed, has ``ov::PropertyName::is_mutable``
145-
returning the ``true`` value. Most of the properites which are changable on ``ov::Core`` level, cannot be changed once the model is compiled,
145+
returning the ``true`` value. Most of the properties which are changable on ``ov::Core`` level, cannot be changed once the model is compiled,
146146
so it becomes immutable read-only property.
147147

148148
Configure a Work with a Model

docs/dev/debug_capabilities.md

+1-1
Original file line numberDiff line numberDiff line change
@@ -1,6 +1,6 @@
11
# OpenVINO Debug Capabilities
22

3-
OpenVINO components provides different debug capabilities, to get more infromation please read:
3+
OpenVINO components provides different debug capabilities, to get more information please read:
44

55
* [OpenVINO Model Debug Capabilities](https://docs.openvino.ai/2023.0/openvino_docs_OV_UG_Model_Representation.html#model-debug-capabilities)
66
* [OpenVINO Pass Manager Debug Capabilities](#todo)

docs/dev/installing.md

+2-2
Original file line numberDiff line numberDiff line change
@@ -108,7 +108,7 @@ To install OpenVINO Development Tools to work with Caffe models (OpenVINO suppor
108108
Linux and macOS:
109109

110110
```sh
111-
#setup virtual envrinment
111+
#setup virtual environment
112112
python3 -m venv openvino_env
113113
source openvino_env/bin/activate
114114
pip install pip --upgrade
@@ -119,7 +119,7 @@ pip install openvino_dev-<version>-py3-none-any.whl[caffe] --find-links=<INSTAL
119119

120120
Windows:
121121
```bat
122-
rem setup virtual envrinment
122+
rem setup virtual environment
123123
python -m venv openvino_env
124124
openvino_env\Scripts\activate.bat
125125
pip install pip --upgrade

docs/dev/test_coverage.md

+1-1
Original file line numberDiff line numberDiff line change
@@ -31,7 +31,7 @@ And build OpenVINO as usual.
3131

3232
## Generate coverage report
3333

34-
In order to generate coverage reports, first of all, the tests must be run. Depending on how many tests are run, the better covegare percentage can be achieved. E.g. for `openvino` component, `InferenceEngineUnitTests`, `ieUnitTests`, `ieFuncTests` must be run as well as plugin tests.
34+
In order to generate coverage reports, first of all, the tests must be run. Depending on how many tests are run, the better coverage percentage can be achieved. E.g. for `openvino` component, `InferenceEngineUnitTests`, `ieUnitTests`, `ieFuncTests` must be run as well as plugin tests.
3535

3636
```bash
3737
$ ctest -V

docs/gapi/gapi_face_analytics_pipeline.md

+2-2
Original file line numberDiff line numberDiff line change
@@ -287,7 +287,7 @@ The sample can also run in a serial mode for a reference and benchmarking purpos
287287
// happens on-the-fly here
288288
avg.start();
289289
} else {
290-
// Measurfe & draw FPS for all other frames
290+
// Measure & draw FPS for all other frames
291291
labels::DrawFPS(frame, frames, avg.fps(frames-1));
292292
}
293293
if (!no_show) {
@@ -296,7 +296,7 @@ The sample can also run in a serial mode for a reference and benchmarking purpos
296296
}
297297
}
298298

299-
On a test machine (Intel® Core™ i5-6600), with OpenCV built with `Intel® TBB <https://www.threadingbuildingblocks.org/intel-tbb-tutorial>`__ support, detector network assigned to CPU, and classifiers to iGPU, the pipelined sample outperformes the serial one by the factor of 1.36x (thus adding +36% in overall throughput).
299+
On a test machine (Intel® Core™ i5-6600), with OpenCV built with `Intel® TBB <https://www.threadingbuildingblocks.org/intel-tbb-tutorial>`__ support, detector network assigned to CPU, and classifiers to iGPU, the pipelined sample outperforms the serial one by the factor of 1.36x (thus adding +36% in overall throughput).
300300

301301
Conclusion
302302
###########

docs/glossary.md

+1-1
Original file line numberDiff line numberDiff line change
@@ -67,7 +67,7 @@ Glossary of terms used in OpenVINO™
6767
| *Batch*
6868
| Number of images to analyze during one call of infer. Maximum batch size is a property of the model set before its compilation. In NHWC, NCHW, and NCDHW image data layout representations, the 'N' refers to the number of images in the batch.
6969

70-
| *Device Affinitity*
70+
| *Device Affinity*
7171
| A preferred hardware device to run inference (CPU, GPU, GNA, etc.).
7272

7373
| *Extensibility mechanism, Custom layers*

docs/model_zoo.md

+2-2
Original file line numberDiff line numberDiff line change
@@ -27,10 +27,10 @@
2727
omz_model_api_ovms_adapter
2828

2929

30-
Open Model Zoo for OpenVINO™ toolkit delivers a wide variety of free, pre-trained deep learning models and demo applications that provide full application templates to help you implement deep learning in Python, C++, or OpenCV Graph API (G-API). Models and demos are avalable in the `Open Model Zoo GitHub repo <https://github.com/openvinotoolkit/open_model_zoo>`__ and licensed under Apache License Version 2.0.
30+
Open Model Zoo for OpenVINO™ toolkit delivers a wide variety of free, pre-trained deep learning models and demo applications that provide full application templates to help you implement deep learning in Python, C++, or OpenCV Graph API (G-API). Models and demos are available in the `Open Model Zoo GitHub repo <https://github.com/openvinotoolkit/open_model_zoo>`__ and licensed under Apache License Version 2.0.
3131

3232
Browse through over 200 neural network models, both :doc:`public <omz_models_group_public>` and from :doc:`Intel <omz_models_group_intel>`, and pick the right one for your solution. Types include object detection, classification, image segmentation, handwriting recognition, text to speech, pose estimation, and others. The Intel models have already been converted to work with OpenVINO™ toolkit, while public models can easily be converted using the :doc:`Model Optimizer <openvino_docs_MO_DG_Deep_Learning_Model_Optimizer_DevGuide>` utility.
3333

3434
Get started with simple :doc:`step-by-step procedures <openvino_docs_get_started_get_started_demos>` to learn how to build and run demo applications or discover the :doc:`full set of demos <omz_demos>` and adapt them for implementing specific deep learning scenarios in your applications.
3535

36-
@endsphinxdirective
36+
@endsphinxdirective

docs/nbdoc/README.md

+1-1
Original file line numberDiff line numberDiff line change
@@ -1,7 +1,7 @@
11
# Jupyter notebooks autodoc
22

33
Auto fetching documentations designed for openvino notebooks tutorials.
4-
This module is responsible for fetching artifats, in this particular example jupyter tutorial notebooks and converting them to notebook documentation.
4+
This module is responsible for fetching artifacts, in this particular example jupyter tutorial notebooks and converting them to notebook documentation.
55

66
## Step 0. Prepare venv
77

docs/ops/arithmetic/Add_1.md

+1-1
Original file line numberDiff line numberDiff line change
@@ -6,7 +6,7 @@
66

77
**Category**: *Arithmetic binary*
88

9-
**Short description**: *Add* performs element-wise addition operation with two given tensors applying broadcasting rule specified in the *auto_broacast* attribute.
9+
**Short description**: *Add* performs element-wise addition operation with two given tensors applying broadcasting rule specified in the *auto_broadcast* attribute.
1010

1111
**Detailed description**
1212
Before performing arithmetic operation, input tensors *a* and *b* are broadcasted if their shapes are different and ``auto_broadcast`` attribute is not ``none``. Broadcasting is performed according to ``auto_broadcast`` value.

docs/ops/arithmetic/Atanh_3.md

+2-2
Original file line numberDiff line numberDiff line change
@@ -16,13 +16,13 @@ Float type input:
1616

1717
a_{i} = atanh(a_{i})
1818

19-
Signed Intragral type put:
19+
Signed Integral type put:
2020

2121
.. math::
2222

2323
a_{i} = (i <= -1) ? std::numeric_limits<T>::min() : (i >= 1) ? std::numeric_limits<T>::max() : atanh(a_{i})
2424

25-
Unsigned Intragral type put:
25+
Unsigned Integral type put:
2626

2727
.. math::
2828

docs/ops/arithmetic/Divide_1.md

+1-1
Original file line numberDiff line numberDiff line change
@@ -6,7 +6,7 @@
66

77
**Category**: *Arithmetic binary*
88

9-
**Short description**: *Divide* performs element-wise division operation with two given tensors applying broadcasting rule specified in the *auto_broacast* attribute.
9+
**Short description**: *Divide* performs element-wise division operation with two given tensors applying broadcasting rule specified in the *auto_broadcast* attribute.
1010

1111
**Detailed description**
1212
Before performing arithmetic operation, input tensors *a* and *b* are broadcasted if their shapes are different and ``auto_broadcast`` attribute is not ``none``. Broadcasting is performed according to ``auto_broadcast`` value.

docs/ops/arithmetic/Multiply_1.md

+1-1
Original file line numberDiff line numberDiff line change
@@ -6,7 +6,7 @@
66

77
**Category**: *Arithmetic binary*
88

9-
**Short description**: *Multiply* performs element-wise multiplication operation with two given tensors applying broadcasting rule specified in the *auto_broacast* attribute.
9+
**Short description**: *Multiply* performs element-wise multiplication operation with two given tensors applying broadcasting rule specified in the *auto_broadcast* attribute.
1010

1111
**Detailed description**
1212
Before performing arithmetic operation, input tensors *a* and *b* are broadcasted if their shapes are different and ``auto_broadcast`` attribute is not ``none``. Broadcasting is performed according to ``auto_broadcast`` value.

docs/ops/arithmetic/Subtract_1.md

+1-1
Original file line numberDiff line numberDiff line change
@@ -6,7 +6,7 @@
66

77
**Category**: *Arithmetic binary*
88

9-
**Short description**: *Subtract* performs element-wise subtraction operation with two given tensors applying broadcasting rule specified in the *auto_broacast* attribute.
9+
**Short description**: *Subtract* performs element-wise subtraction operation with two given tensors applying broadcasting rule specified in the *auto_broadcast* attribute.
1010

1111
**Detailed description**
1212
Before performing arithmetic operation, input tensors *a* and *b* are broadcasted if their shapes are different and ``auto_broadcast`` attribute is not ``none``. Broadcasting is performed according to ``auto_broadcast`` value.

docs/ops/broadcast_rules.md

+1-1
Original file line numberDiff line numberDiff line change
@@ -27,7 +27,7 @@ Rules
2727
1. First input tensor A is of any rank, second input B has rank smaller or equal to the first input.
2828
2. Input tensor B is a continuous subsequence of input A.
2929
3. Apply broadcast B to match the shape of A, where provided *axis* is the start dimension index for broadcasting B onto A.
30-
4. If *axis* is set to default (-1) calculate new value: ``axis = rank(A) - rank(B)``. Except (-1) for default valule, no other negative values are allowed for *axis*.
30+
4. If *axis* is set to default (-1) calculate new value: ``axis = rank(A) - rank(B)``. Except (-1) for default value, no other negative values are allowed for *axis*.
3131
5. The trailing dimensions of size 1 for input B will be ignored for the consideration of subsequence, such as ``shape(B) = (3, 1) => (3)``.
3232

3333
Numpy examples

docs/ops/infrastructure/Assign_3.md

+1-1
Original file line numberDiff line numberDiff line change
@@ -18,7 +18,7 @@ declared in ``variable_id`` and returns an error otherwise.
1818

1919
* *variable_id*
2020

21-
* **Description**: identificator of the variable to be updated
21+
* **Description**: identifier of the variable to be updated
2222
* **Range of values**: any non-empty string
2323
* **Type**: string
2424
* **Required**: *yes*

docs/ops/infrastructure/ReadValue_3.md

+1-1
Original file line numberDiff line numberDiff line change
@@ -20,7 +20,7 @@ with the shape and type from the 1 input.
2020

2121
* *variable_id*
2222

23-
* **Description**: identificator of the variable to be read
23+
* **Description**: identifier of the variable to be read
2424
* **Range of values**: any non-empty string
2525
* **Type**: string
2626
* **Required**: *yes*

docs/ops/matrix/Einsum_7.md

+1-1
Original file line numberDiff line numberDiff line change
@@ -18,7 +18,7 @@ The operation supports ``equation`` in explicit and implicit modes. The formats
1818
In explicit mode, the einsum ``equation`` has the output subscript separated from the input subscripts by ``->``, and has the following format for ``n`` operands:
1919
``<subscript for input1>, <subscript for input2>, ..., <subscript for inputn> -> <subscript for output>``.
2020
Each input subscript ``<subscript for input1>`` contains a sequence of labels (alphabetic letters ``['A',...,'Z','a',...,'z']``),
21-
where each label refers to a dimension of the corresponsing operand. Labels are case sensitive and capital letters precede lowercase letters in alphabetical sort.
21+
where each label refers to a dimension of the corresponding operand. Labels are case sensitive and capital letters precede lowercase letters in alphabetical sort.
2222
Labels do not need to appear in a subscript in alphabetical order.
2323
The subscript for a scalar input is empty. The input subscripts are separated with a comma ``,``.
2424
The output subscript ``<subscript for output>`` represents a sequence of labels (alphabetic letters ``['A',...,'Z','a',...,'z']``).

docs/ops/movement/ScatterUpdate_3.md

+2-2
Original file line numberDiff line numberDiff line change
@@ -68,7 +68,7 @@ The value can be in the range ``[ -r, r - 1]``, where ``r`` is the rank of ``dat
6868
<dim>125</dim>
6969
<dim>20</dim>
7070
</port>
71-
<port id="2"> < !-- udpates -->
71+
<port id="2"> < !-- updates -->
7272
<dim>1000</dim>
7373
<dim>125</dim>
7474
<dim>20</dim>
@@ -102,7 +102,7 @@ The value can be in the range ``[ -r, r - 1]``, where ``r`` is the rank of ``dat
102102
<port id="1"> < !-- indices -->
103103
<dim>2</dim> < !-- {0, 2} -->
104104
</port>
105-
<port id="2"> < !-- udpates -->
105+
<port id="2"> < !-- updates -->
106106
<dim>3</dim> < !-- {1.0f, 1.0f} -->
107107
<dim>2</dim> < !-- {1.0f, 1.0f} -->
108108
</port> < !-- {1.0f, 2.0f} -->

docs/ops/sequence/CTCLoss_4.md

+1-1
Original file line numberDiff line numberDiff line change
@@ -12,7 +12,7 @@
1212

1313
*CTCLoss* operation is presented in `Connectionist Temporal Classification - Labeling Unsegmented Sequence Data with Recurrent Neural Networks: Graves et al., 2016 <http://www.cs.toronto.edu/~graves/icml_2006.pdf>`__
1414

15-
*CTCLoss* estimates likelihood that a target ``labels[i,:]`` can occur (or is real) for given input sequence of logits ``logits[i,:,:]``. Briefly, *CTCLoss* operation finds all sequences aligned with a target ``labels[i,:]``, computes log-probabilities of the aligned sequences using ``logits[i,:,:]`` and computes a negative sum of these log-probabilies.
15+
*CTCLoss* estimates likelihood that a target ``labels[i,:]`` can occur (or is real) for given input sequence of logits ``logits[i,:,:]``. Briefly, *CTCLoss* operation finds all sequences aligned with a target ``labels[i,:]``, computes log-probabilities of the aligned sequences using ``logits[i,:,:]`` and computes a negative sum of these log-probabilities.
1616

1717
Input sequences of logits ``logits`` can have different lengths. The length of each sequence ``logits[i,:,:]`` equals ``logit_length[i]``.
1818
A length of target sequence ``labels[i,:]`` equals ``label_length[i]``. The length of the target sequence must not be greater than the length of corresponding input sequence ``logits[i,:,:]``.

0 commit comments

Comments
 (0)