mirror of
https://github.com/ckaczor/azuredatastudio.git
synced 2026-02-17 02:51:36 -05:00
- Writing the config file in the core for now, will look to move to the extension in Feb release
This commit is contained in:
@@ -398,7 +398,7 @@ export class NotebookModel extends Disposable implements INotebookModel {
|
|||||||
if (!newConnection && this._activeContexts.defaultConnection.options['host'] === host) {
|
if (!newConnection && this._activeContexts.defaultConnection.options['host'] === host) {
|
||||||
newConnection = this._activeContexts.defaultConnection;
|
newConnection = this._activeContexts.defaultConnection;
|
||||||
}
|
}
|
||||||
SparkMagicContexts.configureContext(this.notebookOptions);
|
SparkMagicContexts.configureContext();
|
||||||
this._hadoopConnection = new NotebookConnection(newConnection);
|
this._hadoopConnection = new NotebookConnection(newConnection);
|
||||||
this.refreshConnections(newConnection);
|
this.refreshConnections(newConnection);
|
||||||
this._activeClientSession.updateConnection(this._hadoopConnection);
|
this._activeClientSession.updateConnection(this._hadoopConnection);
|
||||||
|
|||||||
@@ -8,7 +8,6 @@
|
|||||||
import * as path from 'path';
|
import * as path from 'path';
|
||||||
import { nb } from 'sqlops';
|
import { nb } from 'sqlops';
|
||||||
|
|
||||||
import * as json from 'vs/base/common/json';
|
|
||||||
import * as pfs from 'vs/base/node/pfs';
|
import * as pfs from 'vs/base/node/pfs';
|
||||||
import { localize } from 'vs/nls';
|
import { localize } from 'vs/nls';
|
||||||
import { IConnectionProfile } from 'sql/parts/connection/common/interfaces';
|
import { IConnectionProfile } from 'sql/parts/connection/common/interfaces';
|
||||||
@@ -17,11 +16,48 @@ import * as notebookUtils from '../notebookUtils';
|
|||||||
import { INotificationService } from 'vs/platform/notification/common/notification';
|
import { INotificationService } from 'vs/platform/notification/common/notification';
|
||||||
import { IConnectionManagementService } from 'sql/parts/connection/common/connectionManagement';
|
import { IConnectionManagementService } from 'sql/parts/connection/common/connectionManagement';
|
||||||
|
|
||||||
|
const configBase = {
|
||||||
|
'kernel_python_credentials': {
|
||||||
|
'url': ''
|
||||||
|
},
|
||||||
|
'kernel_scala_credentials': {
|
||||||
|
'url': ''
|
||||||
|
},
|
||||||
|
'kernel_r_credentials': {
|
||||||
|
'url': ''
|
||||||
|
},
|
||||||
|
|
||||||
|
'ignore_ssl_errors': true,
|
||||||
|
|
||||||
|
'logging_config': {
|
||||||
|
'version': 1,
|
||||||
|
'formatters': {
|
||||||
|
'magicsFormatter': {
|
||||||
|
'format': '%(asctime)s\t%(levelname)s\t%(message)s',
|
||||||
|
'datefmt': ''
|
||||||
|
}
|
||||||
|
},
|
||||||
|
'handlers': {
|
||||||
|
'magicsHandler': {
|
||||||
|
'class': 'hdijupyterutils.filehandler.MagicsFileHandler',
|
||||||
|
'formatter': 'magicsFormatter',
|
||||||
|
'home_path': ''
|
||||||
|
}
|
||||||
|
},
|
||||||
|
'loggers': {
|
||||||
|
'magicsLogger': {
|
||||||
|
'handlers': ['magicsHandler'],
|
||||||
|
'level': 'DEBUG',
|
||||||
|
'propagate': 0
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
};
|
||||||
export class SparkMagicContexts {
|
export class SparkMagicContexts {
|
||||||
|
|
||||||
public static get DefaultContext(): IDefaultConnection {
|
public static get DefaultContext(): IDefaultConnection {
|
||||||
// TODO NOTEBOOK REFACTOR fix default connection handling
|
// TODO NOTEBOOK REFACTOR fix default connection handling
|
||||||
let defaultConnection: IConnectionProfile = <any> {
|
let defaultConnection: IConnectionProfile = <any>{
|
||||||
providerName: notebookConstants.hadoopKnoxProviderName,
|
providerName: notebookConstants.hadoopKnoxProviderName,
|
||||||
id: '-1',
|
id: '-1',
|
||||||
options:
|
options:
|
||||||
@@ -95,7 +131,7 @@ export class SparkMagicContexts {
|
|||||||
defaultConnection = activeConnections[0];
|
defaultConnection = activeConnections[0];
|
||||||
} else {
|
} else {
|
||||||
// TODO NOTEBOOK REFACTOR change this so it's no longer incompatible with IConnectionProfile
|
// TODO NOTEBOOK REFACTOR change this so it's no longer incompatible with IConnectionProfile
|
||||||
defaultConnection = <IConnectionProfile> <any>{
|
defaultConnection = <IConnectionProfile><any>{
|
||||||
providerName: notebookConstants.hadoopKnoxProviderName,
|
providerName: notebookConstants.hadoopKnoxProviderName,
|
||||||
id: '-1',
|
id: '-1',
|
||||||
options:
|
options:
|
||||||
@@ -112,26 +148,23 @@ export class SparkMagicContexts {
|
|||||||
};
|
};
|
||||||
}
|
}
|
||||||
|
|
||||||
public static async configureContext(options: INotebookModelOptions): Promise<object> {
|
public static async configureContext(): Promise<object> {
|
||||||
let sparkmagicConfDir = path.join(notebookUtils.getUserHome(), '.sparkmagic');
|
let sparkmagicConfDir = path.join(notebookUtils.getUserHome(), '.sparkmagic');
|
||||||
// TODO NOTEBOOK REFACTOR re-enable this or move to extension. Requires config files to be available in order to work
|
// TODO NOTEBOOK REFACTOR re-enable this or move to extension. Requires config files to be available in order to work
|
||||||
// await notebookUtils.mkDir(sparkmagicConfDir);
|
await notebookUtils.mkDir(sparkmagicConfDir);
|
||||||
|
|
||||||
// // Default to localhost in config file.
|
// Default to localhost in config file.
|
||||||
// let creds: ICredentials = {
|
let creds: ICredentials = {
|
||||||
// 'url': 'http://localhost:8088'
|
'url': 'http://localhost:8088'
|
||||||
// };
|
};
|
||||||
|
|
||||||
// let configPath = notebookUtils.getTemplatePath(options.extensionContext.extensionPath, path.join('jupyter_config', 'sparkmagic_config.json'));
|
let config: ISparkMagicConfig = Object.assign({}, configBase);
|
||||||
// let fileBuffer: Buffer = await pfs.readFile(configPath);
|
SparkMagicContexts.updateConfig(config, creds, sparkmagicConfDir);
|
||||||
// let fileContents: string = fileBuffer.toString();
|
|
||||||
// let config: ISparkMagicConfig = json.parse(fileContents);
|
|
||||||
// SparkMagicContexts.updateConfig(config, creds, sparkmagicConfDir);
|
|
||||||
|
|
||||||
// let configFilePath = path.join(sparkmagicConfDir, 'config.json');
|
let configFilePath = path.join(sparkmagicConfDir, 'config.json');
|
||||||
// await pfs.writeFile(configFilePath, JSON.stringify(config));
|
await pfs.writeFile(configFilePath, JSON.stringify(config));
|
||||||
|
|
||||||
return {'SPARKMAGIC_CONF_DIR': sparkmagicConfDir};
|
return { 'SPARKMAGIC_CONF_DIR': sparkmagicConfDir };
|
||||||
}
|
}
|
||||||
/**
|
/**
|
||||||
*
|
*
|
||||||
@@ -186,10 +219,13 @@ interface ISparkMagicConfig {
|
|||||||
kernel_python_credentials: ICredentials;
|
kernel_python_credentials: ICredentials;
|
||||||
kernel_scala_credentials: ICredentials;
|
kernel_scala_credentials: ICredentials;
|
||||||
kernel_r_credentials: ICredentials;
|
kernel_r_credentials: ICredentials;
|
||||||
|
ignore_ssl_errors?: boolean;
|
||||||
logging_config: {
|
logging_config: {
|
||||||
handlers: {
|
handlers: {
|
||||||
magicsHandler: {
|
magicsHandler: {
|
||||||
home_path: string;
|
home_path: string;
|
||||||
|
class?: string;
|
||||||
|
formatter?: string
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
};
|
};
|
||||||
|
|||||||
Reference in New Issue
Block a user