Custom Reports

After generating sources and transferring files to Hadoop, xenn.lake gives you the opportunity to filter and analyze the transferred data.

Xenn.lake discovery module supports the most common file formats which are CSV, JSON and XML. According to your monitoring and filtering needs, PLAIN TEXT files can also be processed.

To filter and process the data, you need to login to xenn.lake web and click on the <DISCOVERY> Menu from the top menu bar. This will lead you to the Discovery Module of the xenn.lake.

For the first login after a fresh installation, there will be no reports defined. To create your first custom report, click on <CUSTOM> button on the left menu. This will lead you new custom report create wizard.

Enter the necessary information for the report definition;

Report Name: The name of the report you want to create

Report Description: The report description, which helps you to find your report after the creation.

Source Definition: Here you define the source, which your processing will work on. Your source can be the direct source of the data which is transferred to Hadoop, or you can define another report as the source for this report definition.

Source Name: The name of the source you have defined in the INTEGRATION module.

Source File: The name of the file defined in the source definition.

Click NEXT to go to the next step.

Second step shows a data preview to define the format of the file. In this example our data is in csv format. So we select CSV as theFile Format. Xenn.lake supports all of the most used file formats including CSV, XML and JSON formats. As we select CSV for the format of the file, we should also provide the Delimiter. In this example, the delimiter is a comma, so we specify <,> as the delimiter.

Click NEXT to go to the next step.

Second preview window appears with the formatted data as you have specified from the previous step. Now you can define your filter conditions and aggregation functions. Here we defined COUNT function on the grouping sets which are the category and subcategory of the product. You can add more conditions by clicking on the <Add> button or delete the conditions by clicking <Delete> button. Once you complete your filtering and aggregation definitions, click NEXT to go to the next step.

Lastly, you should define the repository for the processed data. There are two options;

HADOOP: Processed data will be stored in the Hadoop filesystem.

FILESYSTEM: Processed data will be stored in the OS filesystem.

You should also select the option for the visualisation layer. By default, your processed insight will be shown in the BARCHART format. You can also select a different visual for the presentation layer.

Click FINISH to save your report which will lead you back to the DISCOVERY Module where your saved report exists.

As you can see, the status if the report is DEFINED and it did not run for the time being. To examine the insight, your report should be run by clicking on the <RUN> button. Now you can see that your report status changed to RUNNING.

After the execution completed and the report status updates as COMPLETED, now you can examine your insight. Please also aware thatLast Run:value of the report is also filled by the last run path inside the Hadoop filesystem.

Now you can click on the report name to see the insight. You can also change the way you examine the insight by changing the view type. To change the view type, use the buttons on the left panel of the presentation layer.

Last updated