Skip to content

Commit 2911a37

Browse files
Apply suggestions from code review
Co-authored-by: Iaroslav Igoshev <Poolliver868@mail.ru>
1 parent 572afe1 commit 2911a37

File tree

2 files changed

+18
-19
lines changed

2 files changed

+18
-19
lines changed

docs/getting_started/using_modin/using_modin_cluster.rst

+5-5
Original file line numberDiff line numberDiff line change
@@ -1,9 +1,9 @@
1-
Using Modin
2-
===========
1+
Using Modin in a Cluster
2+
========================
33

4-
In this section, we show how Modin can be used to accelerate your pandas workflows on a cluster.
5-
Different Modin engines have some specifics in running and using a cluster,so you can choose
6-
one of the following instructions to suit the engine you are using.
4+
In this section, we show how Modin can be used to accelerate your pandas workflows in a cluster.
5+
Each Modin distributed engine has its own specifics regarding running and using a cluster so
6+
you can choose one of the following instructions to suit the engine you are using.
77

88
.. toctree::
99
:maxdepth: 4

docs/getting_started/using_modin/using_modin_cluster/using_modin_ray_cluster.rst

+13-14
Original file line numberDiff line numberDiff line change
@@ -1,6 +1,5 @@
1-
================================
2-
Using Modin in a AWS Ray Cluster
3-
================================
1+
Using Modin on Ray in a Cluster
2+
===============================
43

54
.. note::
65
| *Estimated Reading Time: 15 minutes*
@@ -26,8 +25,8 @@ First of all, install the necessary dependencies in your environment:
2625
2726
pip install boto3
2827
29-
The next step is to setup your AWS credentials. One can set `AWS_ACCESS_KEY_ID`,
30-
`AWS_SECRET_ACCESS_KEY` and `AWS_SESSION_TOKEN` environment variables or
28+
The next step is to setup your AWS credentials. One can set ``AWS_ACCESS_KEY_ID``,
29+
``AWS_SECRET_ACCESS_KEY`` and ``AWS_SESSION_TOKEN`` environment variables or
3130
just run the following command:
3231

3332
.. code-block:: bash
@@ -41,10 +40,10 @@ This example starts 1 head node (m5.24xlarge) and 5 worker nodes (m5.24xlarge),
4140
You can check the `Amazon EC2 pricing`_ .
4241

4342
You can manually create AWS EC2 instances and configure them or just use the `Ray autoscaler` to
44-
create and initialize a Ray cluster on Amazon Web Service (AWS) using `Modin's Ray cluster setup config`_ .
45-
You can read more about how to modify `Ray's autoscaler options`_ .
43+
create and initialize a Ray cluster on AWS using `Modin's Ray cluster setup config`_ .
44+
You can read more about how to modify the file on `Ray's autoscaler options`_ .
4645

47-
Detailed instructions can be found in `Ray's cluster docs`_.
46+
More details on how to launch a Ray cluster can be found on `Ray's cluster docs`_.
4847

4948
To start up the Ray cluster, run the following command in your terminal:
5049

@@ -64,7 +63,7 @@ To exit the ssh session and return back into your local shell session, type:
6463
6564
exit
6665
67-
Executing on a cluster environment
66+
Executing in a cluster environment
6867
----------------------------------
6968

7069
.. note::
@@ -78,15 +77,15 @@ on a cluster. In this tutorial, we will use a 12.5 GB `big_yellow.csv` file that
7877
created by concatenating a 200MB `NYC Taxi dataset`_ file 64 times. Preparing this
7978
file was provided as part of our `Modin's Ray cluster setup config`_.
8079

81-
If you want use another dataset in your own script, you should provide it to each of
82-
the cluster nodes in the same path. We recomnend doing this by customizing the
80+
If you want to use the other dataset, you should provide it to each of
81+
the cluster nodes with the same path. We recomnend doing this by customizing the
8382
`setup_commands` section of the [configuration file](https://github.com/modin-project/modin/blob/master/examples/tutorial/jupyter/execution/pandas_on_ray/cluster/modin-cluster.yaml).
8483

85-
To run any scripts on a remote cluster, you need to submit it to the ray. In this way,
84+
To run any script in a remote cluster, you need to submit it to the ray. In this way,
8685
the script file is sent to the the remote cluster head node and executed there.
8786

88-
In this tutorial, we provide the `exercise_5.py`_ script, which read the data from the
89-
CSV file and executed some pandas Dataframe function such as count, groupby and applymap.
87+
In this tutorial, we provide the `exercise_5.py`_ script, which reads the data from the
88+
CSV file and executes such pandas operations as count, groupby and applymap.
9089
As a result of the script, you will see the size of the file being read and the execution
9190
time of each function.
9291

0 commit comments

Comments
 (0)