Skip to content

Commit 01c79ec

Browse files
committed
Improve notebook docs
1 parent c9aaf72 commit 01c79ec

File tree

1 file changed

+6
-8
lines changed

1 file changed

+6
-8
lines changed

examples/graph-analytics-serverless-spark.ipynb

Lines changed: 6 additions & 8 deletions
Original file line numberDiff line numberDiff line change
@@ -30,10 +30,8 @@
3030
"\n",
3131
"The notebook shows how to use the `graphdatascience` Python library to create, manage, and use a GDS Session.\n",
3232
"\n",
33-
"We consider a graph of people and fruits, which we're using as a simple example to show how to connect your AuraDB instance to a GDS Session, run algorithms, and eventually write back your analytical results to the AuraDB database. \n",
34-
"We will cover all management operations: creation, listing, and deletion.\n",
35-
"\n",
36-
"If you are using self managed DB, follow [this example](../graph-analytics-serverless-self-managed)."
33+
"We consider a graph of bicycle rentals, which we're using as a simple example to show how project data from Spark to a GDS Session, run algorithms, and eventually retrieving the results back to Spark.\n",
34+
"We will cover all management operations: creation, listing, and deletion."
3735
]
3836
},
3937
{
@@ -157,7 +155,7 @@
157155
"# Create a GDS session!\n",
158156
"gds = sessions.get_or_create(\n",
159157
" # we give it a representative name\n",
160-
" session_name=\"people_and_fruits\",\n",
158+
" session_name=\"bike_trips\",\n",
161159
" memory=SessionMemory.m_2GB,\n",
162160
" ttl=timedelta(minutes=30),\n",
163161
" cloud_location=CloudLocation(\"gcp\", \"europe-west1\"),\n",
@@ -216,7 +214,7 @@
216214
"1. Send an action `v2/graph.project.fromTriplets`\n",
217215
" This will initialize the import process and allows us to specify the graph name, and settings like `undirected_relationship_types`. It returns a job id, that we need to reference the import job in the following steps.\n",
218216
"2. Send the data in batches to the arrow server.\n",
219-
"3. Send another action called `v2/graph.project.fromTriples.done` to tell the import process that no more data will be send. This will trigger the final graph creation inside the session.\n",
217+
"3. Send another action called `v2/graph.project.fromTriplets.done` to tell the import process that no more data will be send. This will trigger the final graph creation inside the session.\n",
220218
"4. Wait for the import process to reach the `DONE` state.\n",
221219
"\n",
222220
"While the overall process is straight forward, we need to somehow tell Spark to"
@@ -295,8 +293,8 @@
295293
"source": [
296294
"## Sending the computation result back to Spark\n",
297295
"\n",
298-
"Once the computation is done. We might want to further use the result in Spark.\n",
299-
"We can do this in a similar to the projection, by streaming batches of data into each of the Spark workers.\n",
296+
"Once the computation is done, we might want to further use the result in Spark.\n",
297+
"We can do this in a similar way to the projection, by streaming batches of data into each of the Spark workers.\n",
300298
"Retrieving the data is a bit more complicated since we need some input data frame in order to trigger computations on the Spark workers.\n",
301299
"We use a data range equal to the size of workers we have in our cluster as our driving table.\n",
302300
"On the workers we will disregard the input and instead stream the computation data from the GDS Session."

0 commit comments

Comments
 (0)