Skip to content

Commit db20a47

Browse files
committed
fix typo, grammer
1 parent ee419f2 commit db20a47

File tree

2 files changed

+37
-20
lines changed

2 files changed

+37
-20
lines changed

README.md

Lines changed: 23 additions & 8 deletions
Original file line numberDiff line numberDiff line change
@@ -3,11 +3,12 @@
33
<img src="https://github.com/open-datastudio/datastudio/raw/master/docs/_static/open-datastudio-logo.png" width="110px"/>
44
</center>
55

6-
# Open data studio
6+
# Open data studio python client
77

8-
[Open data studio](https://open-datastudio.io) is a managed computing computing service on Staroid cloud. Run your machine learning and large scale data processing workloads without managing clusters and servers.
8+
[Open data studio](https://open-datastudio.io) is a managed computing service on Staroid. Run your machine learning and large scale data processing workloads without managing clusters and servers.
99

10-
Supported computing frameworks are
10+
This repository provides a python client library.
11+
Currently, the following computing frameworks are supported in the library.
1112

1213
- Apache Spark
1314
- Dask (coming soon)
@@ -28,28 +29,29 @@ pip install ods
2829

2930
### Initialize
3031

31-
1. Login staroid.com and get an [access token](https://staroid.com/settings/accesstokens). And set `STAROID_ACCESS_TOKEN` environment variable. See [here](https://github.com/staroids/staroid-python#configuration) for more detail.
32+
1. Login staroid.com and get an [access token](https://staroid.com/settings/accesstokens). And set the `STAROID_ACCESS_TOKEN` environment variable. See [here](https://github.com/staroids/staroid-python#configuration) for more detail.
3233
2. Login staroid.com and create a SKE (Star Kubernetes engine) cluster.
3334

3435
```python
3536
import ods
3637
# 'ske' is the name of kubernetes cluster created from staroid.com.
37-
# Alternatively, you can export 'STAROID_SKE' environment variable.
38+
# Alternatively, you can set the 'STAROID_SKE' environment variable.
3839
ods.init(ske="kube-cluster-1")
3940
```
4041

4142
## Spark
4243

4344
### Create spark session
44-
Create spark session with default configuration
45+
Create a spark session with the default configuration.
46+
You don't need to install/configure spark manually.
4547

4648
```python
4749
import ods
4850
spark = ods.spark("spark-1").session() # 'spark-1' is name of spark-serverless instance to create.
4951
df = spark.createDataFrame(....)
5052
```
5153

52-
Configurue initial number of worker nodes
54+
Configure initial number of worker nodes
5355

5456
```python
5557
import ods
@@ -75,7 +77,7 @@ spark = ods.spark(spark_conf = {
7577
}).session()
7678
```
7779

78-
Check [tests/test_spark.py](https://github.com/open-datastudio/ods/blob/master/tests/test_spark.py) as well.
80+
Check [tests/test_spark.py](https://github.com/open-datastudio/ods/blob/master/tests/test_spark.py) for complete working example.
7981

8082
## Dask
8183

@@ -97,3 +99,16 @@ Coming soon 🚛
9799
import ods
98100
ods.ray(cluster_name="")
99101
```
102+
103+
104+
## Get involved
105+
106+
Open data studio is an open source project. Please give us feedback and feel free to get involved!
107+
108+
- Feedbacks, questions - [ods issue tracker](https://github.com/open-datastudio/ods/issues)
109+
- [Staroid public dev roadmap](https://github.com/staroids/community/projects/1)
110+
111+
112+
## Commercial support
113+
114+
[Staroid](https://staroid.com) actively contributes to Open data studio and provides commercial support. Please [contact](https://staroid.com/site/contact).

notebook/open-data-studio.ipynb

Lines changed: 14 additions & 12 deletions
Original file line numberDiff line numberDiff line change
@@ -6,13 +6,13 @@
66
"source": [
77
"# Open data studio\n",
88
"\n",
9-
"[Open data studio](https://open-datastudio.io) is a managed computing computing service on [Staroid](https://staroid.com) cloud. Run your machine learning and large scale data processing workloads without managing clusters and servers.\n",
9+
"[Open data studio](https://open-datastudio.io) is a managed computing service on [Staroid](https://staroid.com). Run your machine learning and large scale data processing workloads without managing clusters and servers.\n",
1010
"\n",
11-
"Supported computing frameworks are\n",
11+
"[ods](https://github.com/open-datastudio/ods) library makes it easy to use in a Python environment. Currently, the library supports the following computing frameworks.\n",
1212
"\n",
13-
" - [Apache Spark](https://spark.apache.org)\n",
14-
" - [Dask](https://dask.org) (coming soon)\n",
15-
" - [Ray](https://ray.io) (coming soon)\n",
13+
" - Apache Spark\n",
14+
" - Desk (coming soon)\n",
15+
" - Ray (coming soon)\n",
1616
"\n",
1717
"Let's get started!"
1818
]
@@ -26,8 +26,8 @@
2626
"First, you need a SKE (Star Kubernetes Engine) cluster from [staroid.com](https://staroid.com) and access token for it. SKE provides a fully managed, serverless Kubernetes namespace on the cloud.\n",
2727
"\n",
2828
" - Sign up [staroid.com](https://staroid.com)\n",
29-
" - Click 'Kubernetes' -> 'New Kubernetes cluster' to create a new SKE cluster. And set `STAROID_SKE` environment variable.\n",
30-
" - Get access token from 'Account' -> ['Access tokens'](https://staroid.com/settings/accesstokens) menu. And set `STAROID_ACCESS_TOKEN` environment variable."
29+
" - Click 'Kubernetes' -> 'New Kubernetes cluster' to create a new SKE cluster. And set the `STAROID_SKE` environment variable.\n",
30+
" - Get access token from the 'Account' -> ['Access tokens'](https://staroid.com/settings/accesstokens) menu. And set the `STAROID_ACCESS_TOKEN` environment variable."
3131
]
3232
},
3333
{
@@ -47,12 +47,14 @@
4747
"metadata": {},
4848
"source": [
4949
"Now you're ready to go!.\n",
50-
"Let's install and initialize [ods](https://github.com/open-datastudio/ods) module."
50+
"Let's install and initialize the [ods](https://github.com/open-datastudio/ods) module."
5151
]
5252
},
5353
{
54-
"cell_type": "markdown",
54+
"cell_type": "code",
55+
"execution_count": null,
5556
"metadata": {},
57+
"outputs": [],
5658
"source": [
5759
"## Install"
5860
]
@@ -86,7 +88,7 @@
8688
"source": [
8789
"## Spark cluster\n",
8890
"\n",
89-
"Getting Spark cluster is simple. Create a spark session using ods api. The api will download Spark (3.0.0), configure it, create workers on the cloud and connect to them automatically."
91+
"Getting a Spark cluster is simple. Create a spark session using ods library. The library will download Spark (3.0.0), configure it, create workers on the cloud, and connect to them automatically."
9092
]
9193
},
9294
{
@@ -134,7 +136,7 @@
134136
"source": [
135137
"## Stop Spark session and clean up\n",
136138
"\n",
137-
"When spark is no longer needed, you can stop spark session and release executors."
139+
"When the spark is no longer needed, you can stop the spark session and release executors."
138140
]
139141
},
140142
{
@@ -165,7 +167,7 @@
165167
"source": [
166168
"## Commercial support\n",
167169
"\n",
168-
"[Staroid](https://staroid.com) actively contributes to Open data studio project and provides a commercial support. Please [contact](https://staroid.com/site/contact)."
170+
"[Staroid](https://staroid.com) actively contributes to Open data studio and provides commercial support. Please [contact](https://staroid.com/site/contact)."
169171
]
170172
},
171173
{

0 commit comments

Comments
 (0)