-
Notifications
You must be signed in to change notification settings - Fork 311
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Spark Issue + Could not initialize class org.apache.spark.sql.execution.datasources.parquet.ParquetSchemaConverter #2261
Comments
Hello @mokumar1202, thank you for submitting this issue! Yes, we have had a lot of trouble with conflicting Parquet (& Avro) dependencies in Spark over the years, I see you already found issue #1742 Which Spark version are you using (2.3.4 is the most recent 2.3.x)? Which ADAM version are you using? Git head builds against Parquet 1.10.1; we're blocked from upgrading to Parquet 1.11.x (see #2245) |
Yes, i was on version 2.1 and it was all well. I am an administrator and having to do this upgrade so the user community in my organisation can benefit from the new methods/functions in spark 2.3.0 Welcome to Using Scala version 2.11.8 (OpenJDK 64-Bit Server VM, Java 1.8.0_232) |
Also i should have said it before, I am not using ADAM shell. I am testing the spark query from RStudio so just a bash shell. Have i raised this issue in the wrong place? |
Thank you for the clarification! Does the Spark query in RStudio use ADAM as a dependency? If not, then you might want to try the Apache Spark user mailing list. There are a lot of helpful folks on there. See https://spark.apache.org/community.html |
No it does not use ADAM. I will reach out to the Apache Spark community. |
Ah, ok, thank you. Hope that helps! |
I have found a solution, Parquet jar files of version 1.8.3 fixed it and i am able to write data to hadoop. Thought i will post here, if it helps someone else. |
Hi All,
I would need guidance on this please. I upgraded spark to 2.3 from 2.1 and by doing that I am not able to write data to hadoop (I get the below error). I have found that version of the Parquet jar files is 1.8.2, i have also tried with v1.10.0 but no luck
Caused by: java.lang.NoSuchMethodError: org.apache.parquet.column.ParquetProperties.(ILorg/apache/parquet/column/ParquetProperties$WriterVersion;
I also get the below error.
java.lang.NoClassDefFoundError: Could not initialize class org.apache.spark.sql.execution.datasources.parquet.ParquetSchemaConverter$
The text was updated successfully, but these errors were encountered: