Java Spark Sql,无法查询

Java Spark Sql,无法查询,java,Java,我有LinkeIn帐户的数据模式,如下所示。我需要查询数组中的技能,其中数组可能包含JAVA或JAVA或JAVA或JAVA developer或JAVA developer Dataset<Row> sqlDF = spark.sql("SELECT * FROM people" + " WHERE ARRAY_CONTAINS(skills,'Java') " + " OR ARRAY_CONTAINS(skills,'JAVA')"

我有LinkeIn帐户的数据模式,如下所示。我需要查询数组中的技能,其中数组可能包含JAVA或JAVA或JAVA或JAVA developer或JAVA developer

Dataset<Row> sqlDF = spark.sql("SELECT * FROM people"
            + " WHERE ARRAY_CONTAINS(skills,'Java') "
            + " OR ARRAY_CONTAINS(skills,'JAVA')"
            + " OR ARRAY_CONTAINS(skills,'Java developer') "
            + "AND ARRAY_CONTAINS(experience['description'],'Java developer')"  );
Dataset sqlDF=spark.sql(“从人中选择*”
+“其中数组_包含(技能,'Java')”
+“或数组_包含(技能,'JAVA')”
+“或数组_包含(技能,'Java开发人员')”
+“和数组_包含(经验['description'],'Java developer')”;
df.printschema()

root
 |-- skills: array (nullable = true)
 |    |-- element: string (containsNull = true)


df.show()

+--------------------+
|              skills|
+--------------------+
|        [Java, java]|
|[Java Developer, ...|
|               [dev]|
+--------------------+