You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Currently, array types in SparkSQL are not explicitly mapped when converting SparkSQL metadata to an OpenSearch index mapping. This is mainly because OpenSearch inherently supports arrays in any field type. While this does not impact writes from Spark, it may lead to issues in the future when implementing pushdown filtering or aggregation, as OpenSearch treats arrays as part of the parent document rather than independent elements.
Since OpenSearch stores arrays within the parent document, its search phase returns entire matching documents rather than filtering individual array elements. This means that queries targeting specific values within arrays may not work as expected, potentially impacting future pushdown filtering or aggregation optimizations.
How can one reproduce the bug?
Steps to reproduce the behavior:
This is debatable and requires a design decision on how complex types like array in SparkSQL should be mapped to OpenSearch. The design should be consistent across Flint Spark and OpenSearch SQL/PPL handling to ensure compatibility and predictable query behavior.
Do you have any additional context?
N/A
The text was updated successfully, but these errors were encountered:
What is the bug?
Currently, array types in SparkSQL are not explicitly mapped when converting SparkSQL metadata to an OpenSearch index mapping. This is mainly because OpenSearch inherently supports arrays in any field type. While this does not impact writes from Spark, it may lead to issues in the future when implementing pushdown filtering or aggregation, as OpenSearch treats arrays as part of the parent document rather than independent elements.
Since OpenSearch stores arrays within the parent document, its search phase returns entire matching documents rather than filtering individual array elements. This means that queries targeting specific values within arrays may not work as expected, potentially impacting future pushdown filtering or aggregation optimizations.
How can one reproduce the bug?
Steps to reproduce the behavior:
Example:
What is the expected behavior?
This is debatable and requires a design decision on how complex types like array in SparkSQL should be mapped to OpenSearch. The design should be consistent across Flint Spark and OpenSearch SQL/PPL handling to ensure compatibility and predictable query behavior.
Do you have any additional context?
N/A
The text was updated successfully, but these errors were encountered: