* Add support for default permissions on directories
(cherry picked from commit 4e81cceba142c6763c3447b4d2965cd75764f8f9)
* Remove unneeded import
(cherry picked from commit ffe5f357357e75e9290966e89768c699df2e1311)
* Add recursive apply and clean up webhdfs
(cherry picked from commit ae76df14f99e599df1cdfcc74ee22d3822f11a59)
* Final set of changes
* Undo changes to azdata/sqlops and few minor fixes
* Remove cast to fix build error
* Hide defaults checkbox for files and switch checkbox order
* Add display property to ModelView components
* Update DisplayType property in sqlops as well
* More updates to HDFS Manage Access dialog
* More updates to HDFS Manage Access dialog
This is the 1st step to supporting HDFS Tiering
Changes:
Add new mounted folder icon. Will have separate commit for file icon
Disable delete/mkdir/upload for mounted files and folders
Disable delete for root HDFS folder (this was added in error)
* Add width and height properties to checkbox component
* Initial work on manage access dialog
* Add missed change
* Add comments and clean up
* Initial work on manage access dialog
* Add missed change
* Add comments and clean up
* Add return type
* Address comments and use apiWrapper
* Fix compile error
* Fix HDFS with AD auth for browse, read
- HDFS now fully supports expanding nodes for all levels, including using cookie for auth
- HDFS now support reading files from HDFS
- HDFS write file is broken and will be fixed (either in PR update or separate PR)
- Removed hack to use gateway-0 instead of actual DNS name now these are supported. Needed for testing
* Fix Jupyter error using new DMV with endpoints
Partially working AD support for BDCs with some known issues
- Plumbed through kerberos support to Notebooks.
- Using "gateway-0" for service temporarily as service endpoints API doesn't yet return correct DNS name. Will update in separate PR once available
- Plumbed kerberos auth to HDFS, Spark. Only partially working as we use same token on each call
- Will fix in separate PR, as this requires a refactor of WebHDFS library. Will need to either get new token every time or set a cookie, both of which require refactors
- Fixed error when Data Service node expansion failed and blocked all OE expansion
- Support for SqlToolsService change to use new cluster endpoints DMV
- Updated API to add new endpoints field to replace IP + port
- Added logic to handle case where endpoints for Yarn, Grafana etc. are in the list
- Sort list and use expected new localized strings
- Updated SqlToolsService to include support for new DMV
- Add "gateway-0" handling in Jupyter session as workaround for lack of domain names in endpoints list
* Prompt for password once when the sql instance password doesn't work for hdfs. If the user provides the correct password, connect and continue, else show Unauthorized node.
* Removed the hardcoded bad password
* Added check for empty folder scenarios
* Added ErrorStatusCode as property of TreeNode. Checking for the error code instead of the error string to avoid localization issues
* type fixed
* implemented hasExpansionError
* pulling max bytes of data through the webhdfs api
* Added warning message for the Users to inform the data truncation.
* Updated the warning message on the notification flyer.
* Added Unified connection support
* Use generic way to do expandNode.
Cleanup the ported code and removed unreference code. Added as needed later.
Resolved PR comments.
* Minor fixes and removed timer for all expanders for now. If any providers can't response, the tree node will spin and wait. We may improve later.
* Change handSessionClose to not thenable.
Added a node to OE to show error message instead of reject. So we could show partial expanded result if get any.
Resolve PR comments
* Minor fixes of PR comments