Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Replace Class.forName with ShimLoader.loadClass #3726

Merged
merged 1 commit into from
Oct 1, 2021
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
Original file line number Diff line number Diff line change
Expand Up @@ -101,7 +101,7 @@ object ApiValidation extends Logging {
}

// TODO: Add error handling if Type is not present
val gpuTypes = classToTypeTag(Class.forName(gpu))
val gpuTypes = classToTypeTag(ShimLoader.loadClass(gpu))

val sparkToGpuExecMap = Map(
"org.apache.spark.sql.catalyst.expressions.Expression" ->
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -74,7 +74,7 @@ class Spark311CDHShims extends SparkBaseShims {
override def getGpuColumnarToRowTransition(plan: SparkPlan,
exportColumnRdd: Boolean): GpuColumnarToRowExecParent = {
val serName = plan.conf.getConf(StaticSQLConf.SPARK_CACHE_SERIALIZER)
val serClass = Class.forName(serName)
val serClass = ShimLoader.loadClass(serName)
if (serClass == classOf[ParquetCachedBatchSerializer]) {
GpuColumnarToRowTransitionExec(plan)
} else {
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -778,7 +778,7 @@ class Spark320Shims extends Spark32XShims {
override def getGpuColumnarToRowTransition(plan: SparkPlan,
exportColumnRdd: Boolean): GpuColumnarToRowExecParent = {
val serName = plan.conf.getConf(StaticSQLConf.SPARK_CACHE_SERIALIZER)
val serClass = Class.forName(serName)
val serClass = ShimLoader.loadClass(serName)
if (serClass == classOf[ParquetCachedBatchSerializer]) {
GpuColumnarToRowTransitionExec(plan)
} else {
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -663,7 +663,7 @@ abstract class SparkBaseShims extends Spark30XShims {
override def getGpuColumnarToRowTransition(plan: SparkPlan,
exportColumnRdd: Boolean): GpuColumnarToRowExecParent = {
val serName = plan.conf.getConf(StaticSQLConf.SPARK_CACHE_SERIALIZER)
val serClass = Class.forName(serName)
val serClass = ShimLoader.loadClass(serName)
if (serClass == classOf[com.nvidia.spark.ParquetCachedBatchSerializer]) {
GpuColumnarToRowTransitionExec(plan)
} else {
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -636,7 +636,7 @@ abstract class SparkBaseShims extends Spark31XShims {
override def getGpuColumnarToRowTransition(plan: SparkPlan,
exportColumnRdd: Boolean): GpuColumnarToRowExecParent = {
val serName = plan.conf.getConf(StaticSQLConf.SPARK_CACHE_SERIALIZER)
val serClass = Class.forName(serName)
val serClass = ShimLoader.loadClass(serName)
if (serClass == classOf[com.nvidia.spark.ParquetCachedBatchSerializer]) {
GpuColumnarToRowTransitionExec(plan)
} else {
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -3688,7 +3688,7 @@ case class GpuOverrides() extends Rule[SparkPlan] with Logging {
// is impacted by forcing operators onto CPU due to other rules that we have
wrap.runAfterTagRules()
val optimizer = try {
Class.forName(conf.optimizerClassName).newInstance().asInstanceOf[Optimizer]
ShimLoader.newInstanceOf[Optimizer](conf.optimizerClassName)
} catch {
case e: Exception =>
throw new RuntimeException(s"Failed to create optimizer ${conf.optimizerClassName}", e)
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -40,7 +40,7 @@ object HostColumnarToGpu extends Logging {

// use reflection to get access to a private field in a class
private def getClassFieldAccessible(className: String, fieldName: String) = {
val classObj = Class.forName(className)
val classObj = ShimLoader.loadClass(className)
val fields = classObj.getDeclaredFields.toList
val field = fields.filter( x => {
x.getName.contains(fieldName)
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -42,7 +42,7 @@ object PlanUtils {
val execNameWithoutPackage = getBaseNameFromClass(planClass.getName)
execNameWithoutPackage == fallbackCpuClass ||
plan.getClass.getName == fallbackCpuClass ||
Try(java.lang.Class.forName(fallbackCpuClass))
Try(ShimLoader.loadClass(fallbackCpuClass))
.map(_.isAssignableFrom(planClass))
.getOrElse(false)
}
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -21,7 +21,7 @@ import java.util.concurrent.ConcurrentLinkedQueue
import java.util.concurrent.atomic.AtomicInteger

import ai.rapids.cudf.{MemoryBuffer, NvtxColor, NvtxRange}
import com.nvidia.spark.rapids.RapidsConf
import com.nvidia.spark.rapids.{RapidsConf, ShimLoader}

import org.apache.spark.internal.Logging
import org.apache.spark.sql.rapids.storage.RapidsStorageUtils
Expand Down Expand Up @@ -560,7 +560,7 @@ object RapidsShuffleTransport extends Logging {
def makeTransport(shuffleServerId: BlockManagerId,
rapidsConf: RapidsConf): RapidsShuffleTransport = {
val transportClass = try {
Class.forName(rapidsConf.shuffleTransportClassName)
ShimLoader.loadClass(rapidsConf.shuffleTransportClassName)
} catch {
case classNotFoundException: ClassNotFoundException =>
logError(s"Unable to find RapidsShuffleTransport class " +
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -17,19 +17,17 @@
package org.apache.spark.sql.hive.rapids

import com.nvidia.spark.RapidsUDF
import com.nvidia.spark.rapids.{ExprChecks, ExprMeta, ExprRule, GpuExpression, GpuOverrides, RepeatingParamCheck, TypeSig}
import com.nvidia.spark.rapids.{ExprChecks, ExprMeta, ExprRule, GpuExpression, GpuOverrides, RepeatingParamCheck, ShimLoader, TypeSig}
import com.nvidia.spark.rapids.GpuUserDefinedFunction.udfTypeSig

import org.apache.spark.sql.catalyst.expressions.Expression
import org.apache.spark.sql.hive.{HiveGenericUDF, HiveSimpleUDF}

object GpuHiveOverrides {
def isSparkHiveAvailable: Boolean = {
// Using the same approach as SparkSession.hiveClassesArePresent
val loader = Thread.currentThread().getContextClassLoader
try {
Class.forName("org.apache.spark.sql.hive.HiveSessionStateBuilder", true, loader)
Class.forName("org.apache.hadoop.hive.conf.HiveConf", true, loader)
ShimLoader.loadClass("org.apache.spark.sql.hive.HiveSessionStateBuilder")
ShimLoader.loadClass("org.apache.hadoop.hive.conf.HiveConf")
true
} catch {
case _: ClassNotFoundException | _: NoClassDefFoundError => false
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -16,11 +16,11 @@

package com.nvidia.spark.udf

import Repr.UnknownCapturedArg
import java.lang.invoke.SerializedLambda

import com.nvidia.spark.rapids.ShimLoader
import javassist.{ClassClassPath, ClassPool, CtBehavior, CtClass, CtField, CtMethod}
import javassist.bytecode.{AccessFlag, CodeIterator, ConstPool,
Descriptor, MethodParametersAttribute}
import javassist.bytecode.{CodeIterator, ConstPool, Descriptor}

import org.apache.spark.SparkException
import org.apache.spark.sql.catalyst.expressions.Expression
Expand Down Expand Up @@ -152,9 +152,7 @@ object LambdaReflection {
}

def getClass(name: String): Class[_] = {
// scalastyle:off classforname
Class.forName(name, true, Thread.currentThread().getContextClassLoader)
// scalastyle:on classforname
ShimLoader.loadClass(name)
}

def parseTypeSig(sig: String): Option[DataType] = {
Expand Down