You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Hi,
I am trying to connect to Spark 2.1.0 , but the documentation is fairly poor and not well covered yet.
I have tried via thrift server using the hive:// sql alchemy but with no success.
Do I need to setup Druid or can it be without it ?
Can you please share what is supported (if supported) and how to achieve it?
My Spark cluster is standalone, the Superset is running on the same network as Spark and thrift.
Thanks!
The text was updated successfully, but these errors were encountered:
I can connect using the hive:// sql alchemy, but it throws a "Could not locate column in row for column 'database_name'" error when I save the connector as there are no tables in the default (or any other) database - all tables are in global_temp to allow multi-session thrift usage for resource management purposes and thrift does not allow you to set global_temp as the default database since it's a system preserved database. The issue with multi session mode is that it creates a new session for each connection and will not have any tables loaded, so I'm trying to figure out how to create a (dummy) table when the connection is initiated (using metadata_params?) to allow it to find a table and save. Then I should be able to query the global_temp tables.
If you run thrift in single session mode and create some tables or views before connecting superset, it should work. I did have to install some sasl dependencies on the superset python install to get it to connect.
If you get a SASL error on connect, start thrift in NOSASL mode and specify "auth":"NOSASL" in engine_params in superset, or try the following:
Hi,
I am trying to connect to Spark 2.1.0 , but the documentation is fairly poor and not well covered yet.
I have tried via thrift server using the hive:// sql alchemy but with no success.
Do I need to setup Druid or can it be without it ?
Can you please share what is supported (if supported) and how to achieve it?
My Spark cluster is standalone, the Superset is running on the same network as Spark and thrift.
Thanks!
The text was updated successfully, but these errors were encountered: