* Add support for default permissions on directories
(cherry picked from commit 4e81cceba142c6763c3447b4d2965cd75764f8f9)
* Remove unneeded import
(cherry picked from commit ffe5f357357e75e9290966e89768c699df2e1311)
* Add recursive apply and clean up webhdfs
(cherry picked from commit ae76df14f99e599df1cdfcc74ee22d3822f11a59)
* Final set of changes
* Undo changes to azdata/sqlops and few minor fixes
* Remove cast to fix build error
* Hide defaults checkbox for files and switch checkbox order
* Add display property to ModelView components
* Update DisplayType property in sqlops as well
* More updates to HDFS Manage Access dialog
* More updates to HDFS Manage Access dialog
This is the 1st step to supporting HDFS Tiering
Changes:
Add new mounted folder icon. Will have separate commit for file icon
Disable delete/mkdir/upload for mounted files and folders
Disable delete for root HDFS folder (this was added in error)
* Add width and height properties to checkbox component
* Initial work on manage access dialog
* Add missed change
* Add comments and clean up
* Initial work on manage access dialog
* Add missed change
* Add comments and clean up
* Add return type
* Address comments and use apiWrapper
* Fix compile error
* Fix HDFS AD write support + improve error logging
- Implemented write stream piping correctly so data is paused until we're ready to push to server
- During testing, noted that things weren't awaited correctly. Tweaked some functions to improve this
* Use correct content type
* Fix HDFS with AD auth for browse, read
- HDFS now fully supports expanding nodes for all levels, including using cookie for auth
- HDFS now support reading files from HDFS
- HDFS write file is broken and will be fixed (either in PR update or separate PR)
- Removed hack to use gateway-0 instead of actual DNS name now these are supported. Needed for testing
* Fix Jupyter error using new DMV with endpoints
Partially working AD support for BDCs with some known issues
- Plumbed through kerberos support to Notebooks.
- Using "gateway-0" for service temporarily as service endpoints API doesn't yet return correct DNS name. Will update in separate PR once available
- Plumbed kerberos auth to HDFS, Spark. Only partially working as we use same token on each call
- Will fix in separate PR, as this requires a refactor of WebHDFS library. Will need to either get new token every time or set a cookie, both of which require refactors
- Fixed error when Data Service node expansion failed and blocked all OE expansion
- Support for SqlToolsService change to use new cluster endpoints DMV
- Updated API to add new endpoints field to replace IP + port
- Added logic to handle case where endpoints for Yarn, Grafana etc. are in the list
- Sort list and use expected new localized strings
- Updated SqlToolsService to include support for new DMV
- Add "gateway-0" handling in Jupyter session as workaround for lack of domain names in endpoints list
* initial commit: new modelview widget that shows books
* moved the new command action to notebook contribution and addressed the comments
* localize changes
* removed changes from src/vs file and string changes
* make directory for each contribution
* typo fix
* Prompt for password once when the sql instance password doesn't work for hdfs. If the user provides the correct password, connect and continue, else show Unauthorized node.
* Removed the hardcoded bad password
* Added check for empty folder scenarios
* Added ErrorStatusCode as property of TreeNode. Checking for the error code instead of the error string to avoid localization issues
* type fixed
* implemented hasExpansionError
* pulling max bytes of data through the webhdfs api
* Added warning message for the Users to inform the data truncation.
* Updated the warning message on the notification flyer.
* Spark features are enabled
* Fixed as PR comments
* minor change
* PR comments fixed
* minor fix
* change constant name to avoid conflicts with sqlopsextension
* sqlContext to context
* Changed tab name to SQL Server Big Data Cluster
* Added isCluster to ContextProvider to control display big data cluster dashboard tab
Ported New/open Notebook code to mssql extension and enable them in dashboard
* Fixed tslint
* Added data service context menu: file related operations. All new files are ported from SqlOpsStudio. Will remove these functionality from SqlOpsStudio.
* Used the existing constant hadoopKnoxEndpointName
* Rename nodeType name from hdfs to bdc. So we can have file context menu in both mssql and SqlOpsStudio. Need to add "Create External Table from CSV" support for bdc nodeType
* Rename bdc to mssqlcluster
* Added Unified connection support
* Use generic way to do expandNode.
Cleanup the ported code and removed unreference code. Added as needed later.
Resolved PR comments.
* Minor fixes and removed timer for all expanders for now. If any providers can't response, the tree node will spin and wait. We may improve later.
* Change handSessionClose to not thenable.
Added a node to OE to show error message instead of reject. So we could show partial expanded result if get any.
Resolve PR comments
* Minor fixes of PR comments