From 13fefc2895d0603ef4e549d06afce18b4404b497 Mon Sep 17 00:00:00 2001 From: prmoore77 Date: Wed, 5 Jul 2023 10:56:03 -0400 Subject: [PATCH] Minor fix --- README.md | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/README.md b/README.md index 514e281..52c77c5 100644 --- a/README.md +++ b/README.md @@ -26,7 +26,7 @@ It currently lacks: You can choose to build the JAR locally, or use one of the archived JAR artifacts built from a [Github Actions workflow run](https://github.com/rymurr/flight-spark-source/actions/workflows/maven-build.yml). 1. Take the built JAR file named: `flight-spark-source-1.0-SNAPSHOT-shaded.jar` - and copy it to the spark master node. For the sake of this example, we will use the `/tmp` directory -2. Ensure you have a Flight server running and accessible to your Spark cluster. For example of a Python Flight RPC server - see [this link](https://arrow.apache.org/cookbook/py/flight.html#streaming-parquet-storage-service). +2. Ensure you have a Flight server running and accessible to your Spark cluster. For an example of a Python Flight RPC server - see [this link](https://arrow.apache.org/cookbook/py/flight.html#streaming-parquet-storage-service). NOTE: you will have to add a `get_schema` end-point to that example server for it to work - with signature: ```def get_schema(self, context, descriptor) -> pyarrow.flight.SchemaResult``` See this [link](https://arrow.apache.org/docs/python/generated/pyarrow.flight.FlightClient.html#pyarrow.flight.FlightClient.get_schema) for more details.