[FEATURE REQUEST]: Support for sql-spark-connector #611
Replies: 9 comments
-
@rrekapalli this should just work if you do
Make sure to pass the jar file for the connector when you run |
Beta Was this translation helpful? Give feedback.
-
Hi @imback82 , thank you for the quick reply. I tried to write to the database with below command: spark-submit but got the following error. 20/08/05 16:26:52 ERROR Executor: Exception in task 6.0 in stage 17.0 (TID 1016) Appreciate if you could through some light. |
Beta Was this translation helpful? Give feedback.
-
Looks like you are missing dependencies. Can you try with |
Beta Was this translation helpful? Give feedback.
-
Still no luck! When I tired with "azure-sqldb-spark:1.0.1.jar" I got dependencies errors. So tried with the following: spark-submit ******* Output Error Message ***************** Ivy Default Cache set to: C:\Users\rajar.ivy2\cache :: problems summary ::
:: USE VERBOSE OR DEBUG MESSAGE LEVEL FOR MORE DETAILS |
Beta Was this translation helpful? Give feedback.
-
Also, tried with the following with no luck: spark-submit Following is the error message: ********************* output error message ******************************* Ivy Default Cache set to: C:\Users\rajar.ivy2\cache |
Beta Was this translation helpful? Give feedback.
-
My intention for using this driver is to speedup large volume of data transfer between Spark and SQL Server. Appreciate if you could suggest an alternative (DateStreamWriter or something) if this is not the correct approach (can't replace SQL Server for our use case). |
Beta Was this translation helpful? Give feedback.
-
Can you create an issue in https://github.com/Azure/azure-sqldb-spark for the dependency issue? You should be able to repro it with I am not familiar with benchmarks for different SQL connectors. @rapoth do you happen to know? Basically, you want to write the results to SQL server right? |
Beta Was this translation helpful? Give feedback.
-
Sure, I'll create an issue there. Thank you. Also, that's correct. I am trying to both read & write a large volume of data to/from SQL Server. I was able to do that with an MSSQL JDBC driver (mssql-jdbc-8.2.2.jre8.jar) and was successful. However, since data is large it was taking long time to write. I came across this connector which supports bulk operations while looking for a better alternative. @rapoth , could you please suggest what's the best alternative/approach for large data reads and writes from SQL Server using .NET for Spark? Thanks. |
Beta Was this translation helpful? Give feedback.
-
@rrekapalli, I know it has been 4 years but did you ever figure out how to use the sql-spark-connector and how to do bulk inserts with it? We also have noticed that dotnet spark using the default com.microsoft.sqlserver.jdbc.SQLServerDriver driver inserts one row at a time which is incredibly inefficient. Seems that in general dotnet spark might have lost support from Microsoft, so maybe would be best to stop using it, but I still have hope lol. |
Beta Was this translation helpful? Give feedback.
-
Can we expect support for "sql-spark-connector"(https://github.com/microsoft/sql-spark-connector)? Currently it's only available for Scala and Python. Would be a great addition if it's directly implemented in this library.
Beta Was this translation helpful? Give feedback.
All reactions