-
Notifications
You must be signed in to change notification settings - Fork 150
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[BUG] Cannot read excel files using the V2 API #896
Comments
Please check these potential duplicates:
|
Hi @nightscape, yes it is the same issue. I tried to install the artifact 3.4.2 as mentioned, but I still got problems with the DataBricks Runtime 13.3. I tried on the Runtime 14.3 LTS and it works. Thanks |
Hi . I am also facing the same issue, When i am trying to read a excel from azure adls storage. i tried with both 13.3 LTS and 14.3 . Environment
i have tried with 2.13:3.5.1 as well but still the same issue |
@sramesh-nlg you always need to use the version of spark-excel that best matches the Spark version: @massazan unfortunately DataBricks has a little bit of a habit of breaking API compatibility with the officially released Spark versions... |
Am I using the newest version of the library?
Is there an existing issue for this?
Current Behavior
When using the V2 API.
When using the version 0.20.4, the following error occurs: ClassCastException: scala.Some cannot be cast to [Lorg.apache.spark.sql.catalyst.InternalRow;
Error occurs when you omit the end boundary cell on the DataAddress parameter i.e "'0'!A5"
Error is occurs for Scala and PySpark
Expected Behavior
Spark DataReader should return a DataFrame with no errors
Steps To Reproduce
Error occurs when you omit the end boundary cell on the DataAddress parameter i.e "'0'!A5"
val configs = Map(
"inferSchema" -> "false",
"dataAddress" -> "'0'!A5",
"header" -> "false"
)
// Ensure you're using the spark-excel package
val df = spark.read.format("excel")
.option("header", configs("header"))
.option("inferSchema", configs("inferSchema"))
.option("dataAddress", configs("dataAddress"))
.load(s3_path)
df.show()
Environment
Anything else?
API V1 works fine.
The text was updated successfully, but these errors were encountered: