Fix for PySpark3 not being selected by default (#3554)

This commit is contained in:
Chris LaFreniere
2018-12-10 16:26:57 -08:00
committed by GitHub
parent 852ec44567
commit a92dd2d4e4

View File

@@ -149,10 +149,10 @@ export class SparkMagicContexts {
} }
} }
let profile = connectionInfo as IConnectionProfile; let profile = connectionInfo as IConnectionProfile;
if (foundSavedKernelInSpecs && specs && connectionInfo && profile.providerName === notebookConstants.hadoopKnoxProviderName) { if (specs && connectionInfo && profile.providerName === notebookConstants.hadoopKnoxProviderName) {
// set default kernel to default spark kernel if profile exists // set default kernel to default spark kernel if profile exists
// otherwise, set default to kernel info loaded from existing file // otherwise, set default to kernel info loaded from existing file
defaultKernel = !savedKernelInfo ? specs.kernels.find((spec) => spec.name === notebookConstants.defaultSparkKernel) : savedKernelInfo; defaultKernel = !foundSavedKernelInSpecs ? specs.kernels.find((spec) => spec.name === notebookConstants.defaultSparkKernel) : foundSavedKernelInSpecs;
} else { } else {
// Handle kernels // Handle kernels
if (savedKernelInfo && savedKernelInfo.name.toLowerCase().indexOf('spark') > -1) { if (savedKernelInfo && savedKernelInfo.name.toLowerCase().indexOf('spark') > -1) {