* Fix HDFS AD write support + improve error logging
- Implemented write stream piping correctly so data is paused until we're ready to push to server
- During testing, noted that things weren't awaited correctly. Tweaked some functions to improve this
* Use correct content type
* Fix HDFS with AD auth for browse, read
- HDFS now fully supports expanding nodes for all levels, including using cookie for auth
- HDFS now support reading files from HDFS
- HDFS write file is broken and will be fixed (either in PR update or separate PR)
- Removed hack to use gateway-0 instead of actual DNS name now these are supported. Needed for testing
* Fix Jupyter error using new DMV with endpoints
* Adds button extension for explicitly setting aria-label, in case the label is not screen-reader friendly
* Explicitly set aria-label for ellipsis buttons in SchemaCompare
* Changing localize call to double-quotes
* Removes unused getter for ariaLabel
* Adding aria-rowcounts for long scrollable tables
* fixing simple name changes
* Fixing button color and tabbing on tabs
* removing some extra lines of code
* Adding some null checks
* Updating as per PR comments
* added test for sections and navigation
* update variable names
* latest from master
* add await for creating file methods
* wait for all toc.yml to be found
* add event to signal all toc.yml files read
* add workspae folder parameter back
* remove toc filter
* use onReadAllTOCFIles event
* added timeout logic for getting toc.yml files
* latest from master
* new method for comparing book items
* Adds button extension for explicitly setting aria-label, in case the label is not screen-reader friendly
* Explicitly set aria-label for ellipsis buttons in SchemaCompare
* Changing localize call to double-quotes
* Removes unused getter for ariaLabel
Partially working AD support for BDCs with some known issues
- Plumbed through kerberos support to Notebooks.
- Using "gateway-0" for service temporarily as service endpoints API doesn't yet return correct DNS name. Will update in separate PR once available
- Plumbed kerberos auth to HDFS, Spark. Only partially working as we use same token on each call
- Will fix in separate PR, as this requires a refactor of WebHDFS library. Will need to either get new token every time or set a cookie, both of which require refactors
- Fixed error when Data Service node expansion failed and blocked all OE expansion
- Support for SqlToolsService change to use new cluster endpoints DMV
- Updated API to add new endpoints field to replace IP + port
- Added logic to handle case where endpoints for Yarn, Grafana etc. are in the list
- Sort list and use expected new localized strings
- Updated SqlToolsService to include support for new DMV
- Add "gateway-0" handling in Jupyter session as workaround for lack of domain names in endpoints list
Updated existing serialization code so it actually supports serialization. Still needs work to replace the saveAs function when a QueryProvider doesn't support save as, but want to handle in separate PR.
Removed separate MainThread/ExtHostSerializationProvider code as the DataProtocol code is the right place to put this code
Plumbed support through the gridOutputComponent to use the new serialize method in the serialization provider
Refactored the resultSerializer so majority of code can be shared between both implementations (for example file save dialog -> save -> show file on completion)
* Update to latest SQLToolsService release