Hue Using
Hue Using
Using Hue
Date published: 2020-07-28
Date modified: 2023-05-05
https://docs.cloudera.com/
Legal Notice
© Cloudera Inc. 2024. All rights reserved.
The documentation is and contains Cloudera proprietary information protected by copyright and other intellectual property
rights. No license under copyright or any other intellectual property right is granted herein.
Unless otherwise noted, scripts and sample code are licensed under the Apache License, Version 2.0.
Copyright information for Cloudera software may be found within the documentation accompanying each component in a
particular release.
Cloudera software includes software from various open source or other third party projects, and may be released under the
Apache Software License 2.0 (“ASLv2”), the Affero General Public License version 3 (AGPLv3), or other license terms.
Other software included may be released under the terms of alternative open source licenses. Please review the license and
notice files accompanying the software for additional licensing information.
Please visit the Cloudera software product page for more information on Cloudera software. For more information on
Cloudera support services, please visit either the Support or Sales page. Feel free to contact us directly to discuss your
specific needs.
Cloudera reserves the right to change any products at any time, and without notice. Cloudera assumes no responsibility nor
liability arising from the use of products, except as expressly agreed to in writing by Cloudera.
Cloudera, Cloudera Altus, HUE, Impala, Cloudera Impala, and other Cloudera marks are registered or unregistered
trademarks in the United States and other countries. All other trademarks are the property of their respective owners.
Disclaimer: EXCEPT AS EXPRESSLY PROVIDED IN A WRITTEN AGREEMENT WITH CLOUDERA,
CLOUDERA DOES NOT MAKE NOR GIVE ANY REPRESENTATION, WARRANTY, NOR COVENANT OF
ANY KIND, WHETHER EXPRESS OR IMPLIED, IN CONNECTION WITH CLOUDERA TECHNOLOGY OR
RELATED SUPPORT PROVIDED IN CONNECTION THEREWITH. CLOUDERA DOES NOT WARRANT THAT
CLOUDERA PRODUCTS NOR SOFTWARE WILL OPERATE UNINTERRUPTED NOR THAT IT WILL BE
FREE FROM DEFECTS NOR ERRORS, THAT IT WILL PROTECT YOUR DATA FROM LOSS, CORRUPTION
NOR UNAVAILABILITY, NOR THAT IT WILL MEET ALL OF CUSTOMER’S BUSINESS REQUIREMENTS.
WITHOUT LIMITING THE FOREGOING, AND TO THE MAXIMUM EXTENT PERMITTED BY APPLICABLE
LAW, CLOUDERA EXPRESSLY DISCLAIMS ANY AND ALL IMPLIED WARRANTIES, INCLUDING, BUT NOT
LIMITED TO IMPLIED WARRANTIES OF MERCHANTABILITY, QUALITY, NON-INFRINGEMENT, TITLE, AND
FITNESS FOR A PARTICULAR PURPOSE AND ANY REPRESENTATION, WARRANTY, OR COVENANT BASED
ON COURSE OF DEALING OR USAGE IN TRADE.
Cloudera Runtime | Contents | iii
Contents
Procedure
1. Log into the CDP web interface and navigate to the Data Warehouse service.
2. In the Data Warehouse service, navigate to the Overview page.
Note: You can also launch Hue from the Virtual Warehouse page using the same steps.
4
Cloudera Runtime About using Hue
complete feature that is built in. To view the language reference, click the book icon to the right
of the query editor panel.
4. To run Hive queries:
a) On the Overview page under Virtual Warehouses, click on the Hue button.
The Hive query editor is displayed:
complete feature that is built in. To view the language reference, click the book icon to the right
of the query editor panel.
5
Cloudera Runtime About using Hue
Procedure
1. Go to the Cloudera Data Warehouse (CDW) web interface and open Hue from your Virtual Warehouse.
2. Click Jobs from the left assist panel.
The Job Browser page is displayed.
3. Click Queries.
The Hive queries that were run for the past seven days are displayed. You can select the time period for which you
want to view the historical data.
You can also filter queries by their status.
Related Information
Adding Query Store Administrator users in CDW
Procedure
1. Go to the Cloudera Data Warehouse (CDW) web interface and open Hue from your Virtual Warehouse.
2. Click Jobs from the left assist panel.
The Job Browser page is displayed.
6
Cloudera Runtime About using Hue
3. Go to the Queries tab and click on the query for which you want to view the query details.
The following image shows the Query Info tab on the Hue web interface:
Procedure
1. Go to the Cloudera Data Warehouse (CDW) web interface and open Hue from your Virtual Warehouse.
2. Click Jobs from the left assist panel.
The Job Browser page is displayed.
3. Go to the Queries tab and click on the query for which you want to view the query details.
7
Cloudera Runtime About using Hue
5.
(Optional) Click to download the query explain plan in JSON format.
Procedure
1. Go to the Cloudera Data Warehouse (CDW) web interface and open Hue from your Virtual Warehouse.
2. Click Jobs from the left assist panel.
The Job Browser page is displayed.
3. Go to the Queries tab and click on the query for which you want to view the query details.
4. Click on Timeline.
The following image shows the Timeline tab on the Hue web interface:
8
Cloudera Runtime About using Hue
Procedure
1. Go to the Cloudera Data Warehouse (CDW) web interface and open Hue from your Virtual Warehouse.
2. Click Jobs from the left assist panel.
The Job Browser page is displayed.
3. Go to the Queries tab and click on the query for which you want to view the query details.
4. Click on Query Config.
The following image shows the Query Config tab on the Hue web interface:
9
Cloudera Runtime About using Hue
Procedure
1. Go to the Cloudera Data Warehouse (CDW) web interface and open Hue from your Virtual Warehouse.
2. Click Jobs from the left assist panel.
The Job Browser page is displayed.
3. Go to the Queries tab and click on the query for which you want to view the query details.
4. Click DAG Info to see the DAG ID, DAG name, the status of the query, the time taken to execute the DAG, start
time, and end time.
The following image shows the DAG Info tab on the Hue web interface:
The following table lists and describes the status of the Tez job:
Status Description
Submitted The DAG is submitted to Tez but is not running
Running The DAG is currently running
Succeeded The DAG was completed successfully
Failed The DAG failed to complete successfully
Killed The DAG was stopped manually
Error An internal error occurred when executing the DAG
10
Cloudera Runtime About using Hue
Here, the input to vertices Map 1 and Map 2 are the tables displayed in green boxes. Next, Map 2 depends on the
result set generated by Map 1. Map 2 is the last vertex in the DAG flow and after it completes its execution, the
query output is written to a file in a filesystem such as S3 or ABFS.
There are a few options to change the layout of the DAG flow. You can hide the input and the output nodes to
view only the task vertices by clicking the Toggle source/sink visibility button. You can switch between the
horizontal and vertical orientation by clicking the Toggle orientation button.
6. Click DAG Swimlane to see the DAG of the vertices against time.
Each mapping and reducing task is a vertex. Each horizontal bar of the swimlane represents the total time taken
by the vertex to complete the execution. The vertical lines indicate the time when the vertex was initialized, the
time when the vertex started, the time when the first task started, the time when the last task was completed, and
the time when the vertex finished its execution. When you mouse over the vertical line, the bubble displays the
stage of the vertex execution and provides a timestamp. The vertical lines connecting two vertices denote the
dependency of a vertex on another vertex.
The following image shows the DAG Swimlane tab on the Hue web interface:
In this example, Map 1 depends on the results of Map 5. Map 1 will finish its execution only when Map 5 finishes
its execution successfully. Similarly, Reducer 2 depends on Map 1 to complete its execution.
The consolidated timeline shows the percentage of time each vertex took to complete executing.
11
Cloudera Runtime About using Hue
7. Click DAG Counters to see details such as the number of bytes read and written, number of tasks that initiated and
ran successfully, amount of CPU and memory consumed, and so on.
The DAG Counters tab provides a way to measure the progress or the number of operations that occur within a
generated DAG. Counters are used to gather statistics for quality control purposes or problem diagnosis.
The following image shows the DAG Counters tab on the Hue web interface:
12
Cloudera Runtime About using Hue
8. Click DAG Configurations to see the Tez configuration details for a query that has a DAG associated with it.
The following image shows the DAG Configurations tab on the Hue web interface:
Procedure
1. Go to the Cloudera Data Warehouse (CDW) web interface and open Hue from your Virtual Warehouse.
2. Click Jobs from the left assist panel.
The Job Browser page is displayed.
13
Cloudera Runtime About using Hue
3. Click Queries.
The Hive queries that were run for the past seven days are displayed. You can select the time period for which you
want to view the historical data.
You can also search using the query ID, sort queries by various parameters such as duration, peak memory, and so
on, and filter queries by their status.
Procedure
1. Go to the Cloudera Data Warehouse (CDW) web interface and open Hue from your Virtual Warehouse.
2. Click Jobs from the left assist panel.
The Job Browser page is displayed.
3. Go to the Impala Queries tab and click on the query for which you want to view the query details.
The following image shows the Query Info tab on the Hue web interface:
Procedure
1. Go to the Cloudera Data Warehouse (CDW) web interface and open Hue from your Virtual Warehouse.
2. Click Jobs from the left assist panel.
The Job Browser page is displayed.
14
Cloudera Runtime About using Hue
3. Go to the Impala Queries tab and click on the query for which you want to view the execution plan.
The following image shows the Plan tab on the Hue web interface:
Procedure
1. Go to the Cloudera Data Warehouse (CDW) web interface and open Hue from your Virtual Warehouse.
2. Click Jobs from the left assist panel.
The Job Browser page is displayed.
15
Cloudera Runtime About using Hue
3. Go to the Impala Queries tab and click on the query for which you want to view the query metrics.
The following image shows the Metrics tab on the Hue web interface:
Procedure
1. Go to the Cloudera Data Warehouse (CDW) web interface and open Hue from your Virtual Warehouse.
2. Click Jobs from the left assist panel.
The Job Browser page is displayed.
3. Go to the Queries tab.
A list of queries that were run is displayed.
4. Select the queries that you want to stop and click Kill.
16
Cloudera Runtime About using Hue
Procedure
1. Go to the Cloudera Data Warehouse (CDW) web interface and open Hue from your Virtual Warehouse.
2. Click Jobs from the left assist panel.
The Job Browser page is displayed.
3. Go to the Queries tab.
A list of queries that were run is displayed.
17
Cloudera Runtime About using Hue
4. Select the two queries you want to compare and click Compare.
Query comparison report for Hive queries:
18
Cloudera Runtime About using Hue
To run stored procedures from Hue, create a Hive Virtual Warehouse in CDW and enable the hplsql option in the
hue-safety-valve field.
The following example creates a procedure and returns records by passing a cursor:
Attention: In the hplsql mode, you must terminate the commands using the forward slash (/). The semicolon
(;) is used throughout procedure declarations and can no longer be relied upon to terminate a query in the
editor.
Note: HPL/SQL does not support all types of Hive statements, such as JOIN or EXPLAIN. Refer to the
HPL/SQL Reference for more information.
Related Information
Enabling stored procedures for Hive in Cloudera Data Warehouse
19
Cloudera Runtime About using Hue
Procedure
1. Log in to the Data Warehouse service as an administrator.
2. Go to Virtual Warehouse Edit CONFIGURATIONS Hue and select hue-safety-valve from the Configuration
files drop-down list.
3. Add the following lines in the hue-safety-valve:
[notebook]
[[interpreters]]
[[[hive]]]
name=Hive
interface=hiveserver2
[[[hplsql]]]
name=Hplsql
interface=hiveserver2
4. Click APPLY.
5. Restart the Virtual Warehouse.
Procedure
1. Log in to Hue and go to either the Hive or Impala editor.
2. Place your cursor in the editor window and then use one of the following keyboard shortcuts to open the editor
configuration panel:
• On a Mac system, use the Command key followed by a hyphen and then a comma:
Command-,
• On a Windows system, use the Ctrl key followed by a hyphen and then a comma:
Ctrl-,
Tip: Type a question mark (?) anywhere but in the active editor window to open a menu of editor
keyboard shortcuts.
20
Cloudera Runtime About using Hue
3. To enable autocompletion, check the box adjacent to Enable Autocompleter. When you check Enable
Autocompleter, Enable Live Autocompletion is automatically enabled as well. Place your cursor in the editor
window to close the configuration panel.
4. To disable autocompletion:
• Uncheck Enable Live Autocompletion but leave Enable Autocompleter checked, and then place your cursor in
the editor window to close the configuration panel. This disables live autocompletion, but if you want to use
autocompletion while building your queries in the editor, enter the following key stroke sequence to activate
autocompletion: Ctrl + Space Key
• Uncheck both Enable Autocompleter and Enable Live Autocompletion, and then click in the editor to close the
configuration panel. This disables all autocompletion functionality.
Procedure
1. Go to Query Editor Impala or Hive.
21
Cloudera Runtime About using Hue
2. To locate the tags or classifications in Apache Atlas, in the metadata search box located just to the right of the
Query drop-down menu, type a tag: or classification: facet followed by its name. For example, type classification:
wine as shown in the following image:
After you type the search facet and the tag or classification name in the search box, the <database>.<table> where
the tag or classification is found is returned. Click the <database>.<table> to view the tags and classifications that
have been defined for it.
In Cloudera Data Warehouse (CDW), you can browse Amazon S3 buckets from Hue in the following ways:
• With Ranger Authorization Service (RAZ)
• Without Ranger Authorization Service (RAZ)
If you have registered your CDP Public Cloud environment using RAZ, policies attached to the Ranger RAZ
Service role at the Data Lake-level can control access to external S3 buckets. If your Virtual Warehouse predates the
capability to use RAZ configurations and policies, then you must manually enable RAZ for CDW and then configure
Hue to access S3 buckets. Your Data Lake must be RAZ-enabled to manually enable RAZ exclusively for CDW.
Related Information
Registering a RAZ-enabled AWS environment
Enabling RAZ manually in CDW Public Cloud
22
Cloudera Runtime About using Hue
Procedure
1. Follow the instructions listed in Introduction to RAZ on AWS environments to register an AWS environment with
the Enable Ranger authorization for AWS S3 option enabled. You can use the CDP web interface or the CDP CLI
to complete this task.
Note: You must enable RAZ while registering your environment with CDP.
2. Log in to the CDP Management Console as a DWAdmin or DWUser and go to the Cloudera Data Warehouse
service.
3.
Click Open Ranger on your Database Catalog.
4. Create the following Ranger policies:
a) Hadoop SQL policy (all - database, table, column, all - url).
You must grant permissions to individual users or groups in these Ranger policies. To grant permissions to all
users, you can specify {USER} in the Permission section.
b) S3 (cm_S3) policy (Default: User Home)
You must grant permissions to the following users in the Permissions section for the user home directory (/
user/{USER}): {USER}.
Specify the bucket name in the S3 Bucket field and the directory path in the Path field of the cm_S3 Ranger
policy.
c) S3 (cm_S3) policy (Default: user)
You must grant permissions to the following users in the Permissions section for the root directory (/user/):
hive, impala.
5. You must also grant appropriate permissions to the users in CDP User Management Service (UMS). For example,
EnvironmentUser.
Procedure
1. Sign in to Cloudera Data Warehouse DWAdmin or DWUser.
2.
Go to the Virtual Warehouse from which you want to access the S3 buckets and click .
3. Go to CONFIGURATIONS Hue and select hue-safety-valve from the Configuration files drop-down menu.
4. Add the path to your S3 bucket under the [filebrowser] section as follows:
[filebrowser]
remote_storage_home=s3a://[***S3-BUCKET-NAME***]/user
(Optional) Per-user home directories are created by default. To disable automatic user directory creation, you can
add the following lines in the hue-safety-valve as follows:
[desktop]
[[raz]]
autocreate_user_dir=false
5. Click APPLY.
You should be able to view the icon for the S3 File Browser on the left assist panel on the Hue web interface.
23
Cloudera Runtime About using Hue
Only Hue superusers can view and access the S3 File Browser.
Procedure
1. Sign in to the CDP Management Console as an administrator.
2. Go to Data Warehouse service Environments and click the More… menu.
3. Search and locate the environment in which you want to add the S3 bucket and click the edit icon.
The Environment Details page is displayed.
4. Specify the name of the S3 bucket you want to configure access to in the Add External S3 Bucket text box.
If the bucket belongs to another AWS account, then select the Bucket belongs to different AWS account option.
5. Select the access mode.
Read-only access is sufficient to import data in Hue.
6. Click Add Bucket to save the configuration.
A success message is displayed.
7. Click APPLY to update the CDW environment.
Tip: If you configure read only access to an external S3 bucket, there is no need to restart Virtual
Warehouses. However, if you configure read/write access to an external S3 bucket, you must restart
Virtual Warehouses by suspending them and starting them again.
Related Information
Adding Cloudera Data Warehouse cluster access to external S3 buckets in the same AWS account
Adding Cloudera Data Warehouse cluster access to external S3 buckets in a different AWS account
AWS instance profiles
24
Cloudera Runtime About using Hue
Procedure
1. Sign in to Cloudera Data Warehouse.
2.
Click the Open Ranger option on your Database Catalog.
Tip: To grant permissions to all users, you can specify {USER} in the Select User column.
Procedure
1. Sign in to Cloudera Data Warehouse.
2.
Go to the Virtual Warehouse from which you want to access the S3 buckets and click .
25
Cloudera Runtime About using Hue
3. On the Virtual Warehouses detail page, click the Hue tab and select hue-safety-valve from the drop-down menu.
4. Add the following configuration for Hive or Impala Virtual Warehouse in the space provided:
For the Hive Virtual Warehouse:
[desktop]
# Remove the file browser from the blocked list of apps.
# Tweak the app_blacklist property to suit your app configuration.
app_blacklist=oozie,search,hbase,security,pig,sqoop,spark,impala
[aws]
[[aws_accounts]]
[[[default]]]
access_key_id=[***AWS-ACCESS-KEY***]
secret_access_key=[***SECRET-ACCESS-KEY***]
region=[***AWS-REGION***]
[filebrowser]
# (Optional) To set a specific home directory path:
remote_storage_home=s3a://[***S3-BUCKET-NAME***]
[desktop]
# Remove the file browser from the blocked list of apps.
# Tweak the app_blacklist property to suit your app configuration.
app_blacklist=spark,zookeeper,hive,hbase,search,oozie,jobsub,pig,sqoop,sec
urity
[aws]
[[aws_accounts]]
[[[default]]]
access_key_id=[***AWS-ACCESS-KEY***]
secret_access_key=[***SECRET-ACCESS-KEY***]
region=[***AWS-REGION***]
[filebrowser]
# (Optional) To set a specific home directory path:
remote_storage_home=s3a://[***S3-BUCKET-NAME***]
5. Click APPLY.
The S3 File Browser icon appears on the left Assist panel on the Hue web interface after the Virtual Warehouse
restarts.
Creating tables by importing CSV files from AWS S3 in Cloudera Data Warehouse
You can create tables in Hue by importing CSV files stored in S3 buckets. Hue automatically detects the schema and
the column types, thus helping you to create tables without using the CREATE TABLE syntax.
Procedure
1. Sign in to the Cloudera Data Warehouse service.
2. On the Overview page, select the Virtual Warehouse in which you want to create the table and click on Hue.
3. From the left assist panel, click on Importer.
26
Cloudera Runtime About using Hue
If the file is present on your computer, then you can upload it to S3 by clicking Upload a file. To do this, you must
have enabled read/write access to the S3 bucket from the CDW environment.
27
Cloudera Runtime About using Hue
6. Select the CSV file that you want to import into Hue.
Hue displays the preview of the table along with the format:
Hue automatically detects the field separator, record separator, and the quote character from the CSV file. If you
want to override a specific setting, then you can change it by selecting a different value from the drop-down menu.
28
Cloudera Runtime About using Hue
7. Click Next.
On this page, you can set the table destination, partitions, and change the column data types.
29
Cloudera Runtime About using Hue
Hue displays the logs and opens the Table Browser from which you can view the newly created table when the
operation completes successfully.
30
Cloudera Runtime About using Hue
Procedure
1. Follow the instructions listed in Introduction to RAZ on Azure environments to register an Azure environment
with the Enable Ranger authorization for ADLS Gen2 option enabled. You can use the CDP web interface or the
CDP CLI to complete this task.
Note: You must enable RAZ while registering your environment with CDP.
2. Log in to the CDP Management Console as a DWAdmin or DWUser and go to the Cloudera Data Warehouse
service.
3.
Click Open Ranger on your Database Catalog.
4. Create the following Ranger policies:
a) Hadoop SQL policy (all - database, table, column, all - url).
You must grant permissions to individual users or groups in these Ranger policies. To grant permissions to all
users, you can specify {USER} in the Permission section.
b) ABFS (cm_ADLS) policy (Default: User Home)
You must grant permissions to the following users in the Permissions section for the user home directory:
{USER}.
c) ABFS (cm_ADLS) policy (Default: user)
You must grant permissions to the following users in the Permissions section for the root directory (/user/):
hive, impala.
5. You must also grant appropriate permissions to the users in CDP User Management Service (UMS). For example,
EnvironmentUser.
6. Specify the storage account name in the Storage Account field and the directory path of the container and its sub-
directories in the Storage Account Container field of the cm_ADLS Ranger policy.
Enabling the ABFS File Browser for Hue in CDW with RAZ
The ABFS File Browser in Hue is enabled by default. However, you must set the path to your ADLS Gen2 container
in the hue-safety-valve field to avoid a 403 error when you click on the ABFS File Browser.
Procedure
1. Sign in to Cloudera Data Warehouse DWAdmin or DWUser.
2.
Go to the Virtual Warehouse from which you want to access the ADLS Gen2 containers and click .
3. Go to CONFIGURATIONS Hue and select hue-safety-valve from the Configuration files drop-down menu.
4. Add the path to your ADLS Gen2 container under the [filebrowser] section as follows:
[filebrowser]
remote_storage_home==abfs://[***CONTAINER-FOR-DATA-ACCESS***]/user
(Optional) Per-user home directories are created by default. To disable automatic user directory creation, you can
add the following lines in the hue-safety-valve as follows:
[desktop]
[[raz]]
autocreate_user_dir=false
5. Click APPLY.
You should be able to view the icon for the ABFS File Browser on the left assist panel on the Hue web interface.
31
Cloudera Runtime About using Hue
Only Hue superusers can view and access the ABFS File Browser.
Procedure
1. Sign in to the Microsoft Azure portal as an administrator.
2. On the Create storage account Advanced page, enable Data Lake Storage Gen2 so that the objects and files
within your account can be organized into a hierarchy of directories and nested subdirectories in the same way that
the file system on your computer is organized.
abfs://storage-fs@[***AZURE-STORAGE-ACCOUNT-NAME***].dfs.core.windows.net
Procedure
1. Sign in to Cloudera Data Warehouse.
2.
Go to the Virtual Warehouse from which you want to access the ADLS Gen2 containers and click .
3. On the Virtual Warehouses detail page, click the Hue tab and select hue-safety-valve from the drop-down menu.
4. Add the following configuration for Hive or Impala Virtual Warehouse in the space provided:
For the Hive Virtual Warehouse:
[desktop]
# Remove the file browser from the blocked list of apps.
# Tweak the app_blacklist property to suit your app configuration.
app_blacklist=oozie,search,hbase,security,pig,sqoop,spark,impala
[azure]
[[azure_accounts]]
[[[default]]]
client_id=[***AZURE-ACCOUNT-CLIENT-ID***]
client_secret=[***AZURE-ACCOUNT-CLIENT-SECRET***]
tenant_id=[***AZURE-ACCOUNT-TENANT-ID***]
[[abfs_clusters]]
[[[default]]]
fs_defaultfs=abfs://[***CONTAINER-NAME***]@[***AZURE-STORAGE-ACCOUNT-N
AME***]>.dfs.core.windows.net
32
Cloudera Runtime About using Hue
webhdfs_url=https://[***AZURE-STORAGE-ACCOUNT-NAME***].dfs.c
ore.windows.net/
[desktop]
# Remove the file browser from the blocked list of apps.
# Tweak the app_blacklist property to suit your app configuration.
app_blacklist=spark,zookeeper,hive,hbase,search,oozie,jobsub,pig,sqoop,sec
urity
[azure]
[[azure_accounts]]
[[[default]]]
client_id=[***AZURE-ACCOUNT-CLIENT-ID***]
client_secret=[***AZURE-ACCOUNT-CLIENT-SECRET***]
tenant_id=[***AZURE-ACCOUNT-TENANT-ID***]
[[abfs_clusters]]
[[[default]]]
fs_defaultfs=abfs://[***CONTAINER-NAME***]@[***AZURE-STORAGE-ACCOUNT-NA
ME***]>.dfs.core.windows.net
webhdfs_url=https://[***AZURE-STORAGE-ACCOUNT-NAME***].df
s.core.windows.net/
Make sure that the container name and the Azure storage account name that you specify under the abfs_clusters
section is same as what you specified under Data Access Storage Location Base while activating the Azure
environment, so that Hive or Impala has permission to access the uploaded files.
5. Click APPLY.
The ABFS File Browser icon appears on the left Assist panel on the Hue web interface after the Virtual
Warehouse restarts.
Procedure
1. In the CDW service Overview page, select the Virtual Warehouse in which you want to create the table, click the
options menu in the upper right corner and click Open Hue.
2. From the left assist panel, click on Importer.
33
Cloudera Runtime About using Hue
34
Cloudera Runtime About using Hue
Procedure
1. Open Hue from Cloudera Data Warehouse Virtual Warehouse as an EnvironmentAdmin.
2. Go to admin Manage Users Groups .
3. Click on the group to whom you want to grant the filebrowser application permissions.
4. On the Edit [***GROUP-NAME***] page, select the required permission under the permission section and click
Update group.
Uploading files using Hue with the task server enabled in CDW
The task server is enabled by default. When task server is in the enabled state, the files are uploaded using an
asynchronous task queue or job queue. This improves performance and allows you to upload multiple files as large as
5 GB each in parallel.
Procedure
1. Log in to the Hue web interface as a normal user.
2. Go to the Hue File Browser.
3. Click Schedule Upload.
4. Click Select files to browse the files from your local system, and then click Upload.
Alternatively, you can drag and drop the file into the file upload dialogue box.
Note: 5 GB is the maximum supported upload file size per file.
~ Tilde
@ Ampersat
# Hash
$ Dollar sign
& Ampersand
35
Cloudera Runtime About using Hue
( Left paranthesis
) Right paranthesis
* Asterisk
! Exclamation mark
+ Plus
= Equal
: Colon
Not supported with Knox.
; Semicolon
, Comma
. Period
? Question mark
Not supported with Knox.
/ Forward slash
Not supported with Knox.
\ Backslash
36
Cloudera Runtime About using Hue
• External RDBMS in the query editor is not supported out of the box by default. Cloudera support will assist on a
best-effort basis. Cloudera recommends that you raise issues in the open-source github community.
• Impala queries stay in the “executing” state so that Hue can display results when users are ready
• We need to limit the amount of data available to download from Hive/Impala because massive downloads cause
performance degradation. Multiple simultaneous downloads of result sets could also degrade performance.
• Upstream features and connectors may not function properly in CDP. Cloudera recommends that you raise issues
in the open-source github community.
37