Articles & News

15 September 2020

SQL Querying Improvements: Phoenix, Flink, SparkSql, ERD Table...

Hi Data Crunchers, Are you looking at executing your SQL queries more easily? Here is a series of various querying improvements coming in the next release of Hue! New Databases Hue is getting a more polished experience with Apache Phoenix, Apacke Flink SQL and Apache Spark SQL (via Apache Livy). Apache Phoenix Apache Phoenix makes it easy to query the Apache HBase database via SQL. Now the integration is fully working out of the box and several corner cases (e.…

4 minutes read - Version 4 / Phoenix / Flink SQL / ksqlDB / Spark SQL

14 September 2020

REST API for sending SQL queries and Browsing files

Hi Data App Builders, Are you looking at executing some SQL queries of Browsing S3 files programatically? (so that it can be automated for example) The Hue development flow continues to mature (Docker Quick Start, improved CI, shareable Web Components…) and is now getting more help on how to reuse its API. Concept The REST API is not properly public yet and can (will) be simplified in the current work in progress HUE-1450.…

7 minutes read - Version 4 / Development

22 August 2020

Hue in Kubernetes

Let’s see how to automate the deployment further and run this into a container orchestration like Kubernetes.  We previously documented the Hue Docker image that provides the latest Hue web server into a “box”. Hue ships with a Helm chart to make it really easy to get started. Helm is a package manager for Kubernetes. More advanced, directly use the yaml config files. In a Shell, after installing Helm, just execute these three instructions to boot a live Hue with its database:…

1 minute read - Version 4

19 August 2020

Quickstart Hue in Docker and query any of your Database

Query any Data Warehouse in minutes! In this tutorial we configure Hue to use an existing MySQL database and point to an Apache Hive data warehouse that we want to query. Note If you are looking another warehouse than Hive, check out all the other connectors. There are two important concepts: Persistence: Hue needs an existing database with transactions like MySQL to support concurrent requests and also not lose the registered users, saved queries, sharing permissions… when the server gets stopped.…

3 minutes read - Version 4

15 August 2020

Automatically checking the Python style and title format of Git commits

Hi Query Engineers, The investment in Continuous Integration (CI) and automation continues in order to help scale the resource and quality of the Hue project. This past year saw a lot of improvements with: An integrated commit flow automatically run by Circle CI and websites link checking JavaScript styling and artifacts licensing check Here is the latest about how to automatically check that the coding convention of the Python API is followed by everybody as well as the format of the git commit titles.…

4 minutes read - Version 4 / Development

23 June 2020

Monitoring Hue activity with Grafana Dashboards

Besides Hue Active Users Metric Improvements, we’ve also added a new Prometheus metrics number of queries, which shows how many queries were executed in the last ten minutes. Now both of them are available in Hue’s Grafana dashboard. Let’s go to the Grafana dashboard list, you will find Hue’s Home dashboard in the Hue folder. After opening the Home dashboard, you will see four panels which list CPU, memory, active users, and number of queries.…

1 minute read - Version 4 / Development

22 June 2020

Automated checks for JavaScript modules compatible licenses and non absolute paths with Continuous Integration

Hi Interface Builders, The investment in Continuous Integration (CI) and automation continues in order to help scale the resource and quality of the Hue project. This past year saw a lot of improvements with an integrated commit flow and adding series of link checking automatically run by Circle CI. Here is the latest about how to automatically check that JavaScript artifacts have the correct licensing and do not contain absolute paths.…

2 minutes read - Version 4 / Development

19 May 2020

How to grant Ranger permissions for a new user on a Secure Cluster

Hello, Hue administrators, The Apache Ranger™is a framework to enable, monitor and manage comprehensive data security across the Hadoop platform. Hue integrates with Ranger since Hue4.6, on any secure cluster with Ranger installed, the user has to have proper permissions to operate on the data to avoid following permission warning. Here let's show the detailed steps to grant permission for any new user through CM UI. Steps On any CM managed cluster, you can navigate to Ranger service, note down the ‘Ranger Usersync’ host on “Instances” tab, then open ‘Ranger Admin Web UI’ open a terminal and ssh to your ranger usersync host which you noted down at Step 1;…

2 minutes read - Version 4

06 May 2020

SQL Editor for Apache Flink SQL

Flink SQL Editor This is the very first version of the SQL Editor for Flink. The goal is to demo how to execute Flink SQL queries. We use the new Flink SQL gateway project and point to a Flink cluster with live data in a docker container. Hue is used as the SQL Editor for querying Flink tables. Feel free to read more about Flink SQL and continuous queries. Setup Any Flink 1.…

4 minutes read - Version 4 / Flink SQL

05 May 2020

How to Configure Hue to authenticate with Apache Knox SSO on a Secure Cluster

Hello, Hue administrators, The Apache Knox™ Gateway is an Application Gateway for interacting with the REST APIs and UIs of Apache Hadoop deployments. Hue supports KnoxSpnegoDjango since Hue4.6, we can turn on Hue’s KnoxSpnegoDjango auth by updating Hue configurations through CM UI or hue.ini. On any cluster with Knox service installed, update hue.ini as following and restart Hue: [desktop] [[auth]] backend=desktop.auth.backend.KnoxSpnegoDjangoBackend [[knox]] knox_principal=knox, Or on any CM managed cluster, Hue can be configured with KnoxSpnegoDjango backend through CM UI: Fill knox_proxyhosts field with accurate knox proxy hostname, you can get the hosts by navigating to Clusters->KNOX, and click on ‘Instances’ tab: For Knox HA cluster, you can fill in all the hosts by clicking on “+” icon: Click ‘Save Changes’, you will see a warning about role missing kerberos keytab.…

2 minutes read - Version 4

More recent stories

30 June 2021
Azure Storage sharing by leveraging SAS tokens so that your users don’t need credentials
Read More
10 June 2021
Hue 4.10 and its new SQL Editor component, REST API, small File Importer and Slack App are out!
Read More
29 May 2021
Build your own SQL Editor (BYOE) in 5 minutes via Sql Scratchpad component and public REST API.
Read More