How to force close Spark Hive metastore connections to mysql after stopping context? -


i using spark querying hive followed transformations. scala app creates multiple spark applications. new spark app created after closing sparksession , spark context of previous spark app.

however, on stopping sc , spark, somehow connections hive metastore (mysql) not destroyed properly. every, spark app can see around 5 mysql connections being created (old connections being still active!). eventually, mysql starts rejecting new connections after 150 open connections. how can force spark close hive metastore connections mysql (after spark.stop() , sc.stop())?

note: have used spark 2.1.1. using spark's thriftserver instead of hiveserver. so, don't think have used hive metastore service.


Comments

Popular posts from this blog

PHP and MySQL WP -

android - InAppBilling registering BroadcastReceiver in AndroidManifest -

nginx - phpPgAdmin - log in works but I have to login again after clicking on any links -