How to create a Stacked Bar Chart with my data set created? - spagobi

I have just now deployed a SpagoBI server and after setting up my data source: MS SQL Server.
I am able to successfully create a data set. Now using my data set I want to create a stacked bar chart. So I select Charts->Stacked but to my surprise that perhaps a default stackbar of "Sales, costs and revenues over the year Monthly detail - Year: 2012" always comes up. I was not able to see charts for my data set.
However I still see my data set there. I tried many ways to use the output of my data set for stack bar but un-fortunately always the default chart comes up?
How can I create a Stacked Bar Chart with my data set created?

1.First, write xml related to your query in database.
2.then , log in to SpagoBI server then create datasource-> dataset-> choose document.
3.Fill all the informaion, and there is an Template option browse the xml file in it.
4.And create Analytical driver management and lov management related to your query columns thatyou wants to choose.
5.And run the report document.

If you creating dashboard in SpagoBI Studio 5.1 then
1.Click on project and Create under SpagoBI.
2.Click on window - others perspective - choose - report design.
3.In report design - create Report - under Business Analysis folder finish.
4.Right Click on data source create - new data source and choose your db as per your requirement.
5.then , data set.
6.then create your dashboard as you want to create.
7.When you deploy your report in SpagoBI server again click on windows - other perspective - choose SpagoBI
8.Expand your project and you see a Server name folder then right click on server provide SpagoBI url and password
9.Right click on report provide name all the credential. and deploy.
Try this Prakash.

Related

Power Automate: Export to Data Lake Business Events Alerts

Essentially, we are trying to use get a notification whenever a table from export to data lake changes state (i.e., deactivated) as described in this article: https://learn.microsoft.com/en-us/dynamics365/fin-ops-core/dev-itpro/data-entities/azure-data-lake-generates-biz-events
We have installed the Dynamics 365 Export to Data Lake solution in the Power Platform admin center and can see the Catalog and Category as intended. However, we are getting the error message: “Could not retrieve values. Did not find any entities for catalog 'msdynedl_catalog' and category 'msdynedl_dataexportcategory'.” when trying to finalize the setup (Data Lake is also set up and running): (https://i.stack.imgur.com/C32m8.png)
I have gone to Data Management > Framework parameters > Refresh entity list and subsequently ‘Rebuilt business event catalog’ without success.
If I add (none) into the field as per Microsoft's documentation, I get the following error a step later: (https://i.stack.imgur.com/oUFbJ.png)
Or filling in all steps:
"Flow save failed with code 'DynamicOperationRequestClientFailure' and message 'The dynamic operation request to API 'commondataserviceforapps' operation 'GetMetadataForActionInputAndResponseForWhenAnActionIsPerformedTrigger' failed with status code 'NotFound'. This may indicate invalid input parameters. Error response: {"error":{"code":"EntityNotFound","message":"Could not find table with name '(none)'."}}'."
I have gone to Data Management > Framework parameters > Refresh entity list and subsequently ‘Rebuilt business event catalog’ without success. The desired business events are not showing up even after enabling the D365 Export to Data Lake solution in the Power Platform admin center as per the documentation.
I cannot figure out how to resolve this issue. Any advice is appreciated!
Solution: The issue was that the listed environment was not actually the correct one but just a user created one. An admin had to go in and activate the Export to Data Lake Add-in in the Power Platform Admin center and then set up the connection themselves in Power Automate, as the environment would only be visible to the admin user. The rest populated as expected after this change.

Oracle Apex: Is it possible to change the default schema at runtime?

We have an Apex application (version 20.1) and our users must be able to change the database schema at runtime via button click (preferably without logging in again).
Currently we are solving this by installing our application multiple times, once per schema.
We recently discovered the function apex_export.get_application. We intend to use this function to bring our frontend under version control (finally!). We would like to deploy our application directly from the exported files. Having a single application, we would not have to mess with the internal component ids from the exported files.
Is it possible to install the application once and change the default schema via Pl/SQL code? Thank you!
I don't think this can be done, but perhaps the following is a reasonable compromise
add all the schemas you need to support to the workspace schema list
Any SQL (and I do mean any) in your app would be prefixed with an application item, eg
Before: select * from my_table
After: select * from &my_schema..my_table
At login time (or when a user selects it) you modify the MY_SCHEMA application item
(I've not tried this...so test/tread carefully)

Exporting Sonarqube reports into Excel - based on major, minor and critical categories

Is there a way to export Sonarqube reports into Excel - based on major, minor and critical categories?
You can use the REST API, to query the data into JSON text and then export that JSON to a CSV file.
I used the command below to get a JSON response:
http://xxxxx.xx.xxxx.com:9000/api/issues/search?componentRoots=test_xxx_xx&statuses=OPEN,REOPENED&pageSize=500&pageIndex=1
Where componentRoots is the your sonar project name.
It gave all the issues in JSON and then I converted it in to a CSV.
The only way is to use the api/issues/search web service
Yes you can use the REST APIs provided with SonarQube to query. The documentation of APIs is also embedded into every Sonar instance as different versions expose different APIs.
We use Python for similar work as response will be in JSON and it will be easier to manipulate. Once you have arrived extracted issues of your liking, write them into .CSV or excel.
Link to web services will be in footer of Sonar Instance.
PS: Expanded answer to offset short-sighted answering. What is provided here is only abstract and not complete answer with query details.
Not recommended by community
With DB:- If you have been using Sonar for long and if you won't upgrade SQ too often you can choose to study table structure and understand how data is organised. We have done this too, but it gets messier with every passing upgrade (more inner queries). Cost of querying will be saved on bulk process of data as ES is not involved
Please try the following command to get all issues in JSON format. Then you may consider to parse output by using jason parser programs.
Replace "XXX:XXX" with Sonar Key defined in "sonar.projectKey" variable on "sonar-project.properties.txt" file
http://localhost:9000/api/issues/search?componentKeys=XXX:XXX
I used F12 developer tools to create the report. Here is step by step solution:
Before opening Sonar report, press F12 to open developer tools in the browser tab (Refer screenshot). Click on Network tab. Initially it'll be empty:
Open Sonar report in the same browser tab.
Now you have to select a request in URL column (Refer screenshot in point # 4). You've to select that request whose URL is of the format - https://sonar:8443/api/issues/search?. Base URL might be different depending upon the name of the server where you have hosted Sonar website.
Click on Response tab (Refer screenshot):
Copy the entire JSON data to clipboard. First press Ctrl + A and then Ctrl + C to copy the contents.
JSON data that you got is in below format:
Data of all the issues is present inside issues key. It is of array type as it contains a list of issues. Paste the entire JSON data into any notepad editor and get the value of issues node. A sample issues node data is shown below:
[
{
"key":"AYMBNpviy48WWZHdsU1Z"
,"rule":"OWASP:UsingComponentWithKnownVulnerability"
,"severity":"MAJOR"
,"component":"B91661CE-50F8-45CB-8F54-29CD044EF32D"
,"project":"B91661CE-50F8-45CB-8F54-29CD044EF32D"
,"flows":[]
,"status":"OPEN"
,"message":"Filename: jquery-ui-1.10.3.custom.js | Reference: CVE-2021-41184 | CVSS Score: 6.1 | Category: CWE-79 | jQuery-UI is the official jQuery user interface library. Prior to version 1.13.0, accepting the value of the `of` option of the `.position()` util from untrusted sources may execute untrusted code. The issue is fixed in jQuery UI 1.13.0. Any string value passed to the `of` option is now treated as a CSS selector. A workaround is to not accept the value of the `of` option from untrusted sources."
,"author":""
,"tags":["cve","cwe","cwe-937","owasp-a9","vulnerability"]
,"transitions":["confirm"]
,"actions":["set_tags","comment","assign"]
,"comments":[]
,"creationDate":"2022-08-18T06:35:49+0100"
,"updateDate":"2022-08-18T06:35:49+0100"
,"type":"VULNERABILITY"
,"branch":"sonarqubepr"
,"scope":"MAIN"
,"quickFixAvailable":false
},
{
"key":"AYMBNpviy48WWZHdsU1_"
,"rule":"OWASP:UsingComponentWithKnownVulnerability"
,"severity":"MAJOR"
,"component":"B91661CE-50F8-45CB-8F54-29CD044EF32D"
,"project":"B91661CE-50F8-45CB-8F54-29CD044EF32D"
,"flows":[]
,"status":"OPEN"
,"message":"Filename: jquery.ui.datepicker-ml.min.js | Reference: CVE-2021-41183 | CVSS Score: 6.1 | Category: CWE-79 | jQuery-UI is the official jQuery user interface library. Prior to version 1.13.0, accepting the value of various `*Text` options of the Datepicker widget from untrusted sources may execute untrusted code. The issue is fixed in jQuery UI 1.13.0. The values passed to various `*Text` options are now always treated as pure text, not HTML. A workaround is to not accept the value of the `*Text` options from untrusted sources."
,"author":""
,"tags":["cve","cwe","cwe-937","owasp-a9","vulnerability"]
,"transitions":["confirm"]
,"actions":["set_tags","comment","assign"]
,"comments":[]
,"creationDate":"2022-08-18T06:35:49+0100"
,"updateDate":"2022-08-18T06:35:49+0100"
,"type":"VULNERABILITY"
,"branch":"sonarqubepr"
,"scope":"MAIN"
,"quickFixAvailable":false
}
]
The JSON array data obtained in previous step is of real use. Convert this JSON array data into an EXCEL file using any online converter. I used the following website - https://www.convertcsv.com/json-to-csv.htm
What if my report has more than one page?
Sonar loads 100 records per page. To get the next page you've to click on Show More button at the bottom:
Every time you click on the Show More button, a new request goes to the Sonar server which will show up in the F12 developer tool's Network tab. So for all subsequent pages, repeat same set of steps to create a new Excel sheet per page. Then merge all the Excel sheets manually to create single master report.
Note: While merging the Excel files, remember that every Excel file will have a header row at the top. While merging the files, you'll have to keep the header row from from the first Excel sheet and ignore from the rest.

Crystal Reports Data Source in Oracle

I inherited a legacy system using Crystal Reports in VS.NET pulling info from an Oracle11g DB.
The database fields are in "REPORT_DATA", I updated the stored procedure which the code is querying to add a field, however I cannot see that new field in the field explorer.
When I click on "Verify Database" and manage to log into the DB I get a "Remove Table" prompt - "The database table "REPORT_DATA" cannot be found. Proceed to remove this table from the report?"
Of course, there is no REPORT_DATA table and there never was.
When I preview the report or put it on the server it works, so there is some behind-the-scenes magic which I don't understand.
Figured it out.
I had to highlight the server name in the “Current Data Source” area and the name of the server in the “Replace With” area on the Set Datasource Location window. Then click the “Update” button.
Since the report wasn't built on my machine it had to be reconfigured - can't figure out why it still connected to the DB when previewing though.

Reporting Services dynamic image source URL problems

I have a report in SQL Reporting Services 2008 R2 that includes an image whose source is given by an expression. The image won't render though. This report was moved from a 2005 machine where it works just fine.
The image source is set to:
="http://example.com/SiteMap/Directory/EmployeePhotos/" + rtrim(Fields!employee_id.Value) + ".jpg"
Which should render out to this source in the actual report:
http://example.com/SiteMap/Directory/EmployeePhotos/275.jpg
Instead I just get an empty image spacer; when I look at the properties in IE I get a bunch of "Not Available" values.
I set up a row in my report with the exact expression above and it gives the expected value.
Again, the image works just fine on the 2005 box and it looks fine when it's rendered in VS 2010. Any ideas? Thanks as always for your help.
The reporting server grabs the image, not the browser. This means the service account, not just the user, must have access to the external resources.
The good news is you don't need to change the actual account under which SSRS runs. Reporting services allows you to assign an execution account which SSRS will use to get external resources. Set up an account that has access to the resources, then set that account as the execution account in SSRS configuration manager. Easy.

Resources