Gephi: display modularity in data laboratory - modularity

I use Gephi 0.9.1 and 0.9.2. I run modularity and have a number of class. When right-clicking on a node, I choose select in data laboratory. In a youtube tutorial, one got an extra column with modularity which allowed identification of modularity and export to csv files for further processing, lists, etc.
Is this possible?

Just found out that the number of displayed columns is limited. So one has to deselect some of the default ones if one wants additional columns to show up.

Related

Power BI: Do not output empty leaves in the hierarchy, Power BI / VS Report Builder, Joining two or more tables

Project Project PowerBI
I ask for your help, have already climbed a bunch of resources and can not find a suitable solution
The problem is this:
There is a heterogeneous hierarchy in the visual element (matrix), how to remove the empty leaves (levels) of the hierarchy? (project in the attachment)
As far as I saw, there are still no standard methods, but still. From the side of logic, you just need not to calculate the measure for empty elements in the line (the hierarchy goes line by line for each unique value) or replace the value with 0 by it and filter the measure so that 0 is not displayed
how to make hierarchy in columns with expand / collapse +/-, I saw and used different visual elements in the marketplace None came up, the most suitable was "Zebra BI", but it terribly slows down a million records in the fact table and adding several levels to the hierarchy columns, just at some point can’t load anything (or I'm doing something wrong, but it seems hard to make a mistake somewhere)
There is a report builder — either Power BI Report Builder or Visual Studio, and there the Reporting Services Projects project (as I understand it can be said the same thing) - we create and then upload Power BI reports to the server. It seems like he can do anything (this also applies to the second question), if you know that this is so and there are some examples, then I will be glad to study them, I just did not have time to work with him
There are different tables (matrix) on the layout (one below the other) have the same columns, but they have different rows and hierarchies by rows, they also have different measures.
2.1. How to glue two or more tables one under the other so that they can be viewed at the same time - when expanding the columns, did the columns expand in the two tables? (I understand that it is possible to do a full disclosure of the hierarchy of columns and filters to do everything synchronously, but in order to do the expansion at the top (detail), and all the tables were magic at once)
2.2 So that when expanding the hierarchies in the rows of the table, a vertical slider does not appear for each table separately, i.e. The matrix table was rubber and there was no collision to the table below. In Microstrategy, it happened that I created a table, cut a layout (canvas), inserted another table below the cut, or whatever, it turned out as if it were one piece and the tables were rubber and did not overlap.
In general, it is necessary - different tables, one under the other with different measures and row hierarchies, but with the same hierarchies of columns to make work as a whole. Again Report Builder to help? In the attached project, let's imagine that these are different tables and we need them to work as one.
I am waiting for your help, my knowledge is not enough and Google does not help, I don’t know how to contact him =)
Thanks!

Design a dimension with multiple data sources

I am designing a few dimensions with multiple data sources and wonder what other people have done to align the multiple business keys per data source.
My Example:
I have 2 data sources - the Ordering System and the Execution System. The Ordering system has details about payment and what should happen; the Execution System has details on what actually happened (how long it took etc, who executed on the order). Data from both systems is need to created a single fact.
In both the Ordering and Execution system they is a Location table. The business keys from both systems are mapped via an esb . There are attributes in both systems that make up the complete picture about a single location. Billing information is in the Ordering system, latitude and longitude are in the Execution system. And Location Name exists in both systems.
How do you design a SCD accomodate changes from both systems to the dimension?
We follow a fairly strict Kimball methodology - fyi, but I am open to looking at everyone's solutions.
Not necessarily an answer but here are my thoughts:
You've already covered the real options in your comment. Either:
A. Merge it beforehand
You need some merge functionality in staging which matches the two (or more) records, creates a new common merge key and uses that in the dimension. This requires some form of lookup or reference to be stored in addition to normal DW data
OR
B. Merge it in the dimension
Put both records in the dimension and allow the reporting tool to 'merge' it by, for example, grouping by location name. This means you don't need prior merging logic you just dump it in the dimension
However you have two constraints that I feel makes the choice between A & B clearer
Firstly, you need an SCD (Type 2 I assume). This means Option B could get very complicated as when there is a change in one source record you have to go find the the other record and change it as well - very unpleasant for option B. You still need some kind of pre-stored key to link them, which means option B is no longer simple
Secondly, given that you have two sources for one attribute (Location Name), you need some kind of staging logic to pick a single name when these don't match
So given these two circumstances, I suggest that option A would be best - build some pre-merging logic, as the complexity of your requirements warrants it.
You'd think this would be a common issue but I've never found a good online reference explaining how someone solved this before.
My thinking is actually very trivial. First you need to be able to conclude what is your master dataset on Geo+Location and granularity.
My method will be:
DIM loading
Say below is my target
Dim_Location = {Business_key, Longitude, Latitude, Location Name}
Dictionary
Business_key = Always maps to master record from source system (in this case it is the execution system). Imagine now the unique key from business is combined (longitude, latitude) for this table.
Location Name = Again, since we assume the "Execution system" is master for our data then it will host from Source="Execution System".
The above table is now loaded for Fact lookup.
Fact Loading
You have already integrated record between execution system and billing system. It's a straight forward lookup and load in staging since it exists with necessary combination of geo_location.
Challenging scenarios
What if execution system has a late arriving records on orders?
What if same geo_location points to multiple location names? Not possible but worth profiling the data for errors.

ServiceNow Compare Servers Attributes

How can I get a diff between two servers taking all the attributes like OS Details, Patches installed, Load Balncer associated, Software installed etc.
As far as I know there is no default option to get the diff between two records in ServiceNow. Some ideas:
Use the list view of the table, show all the columns you want to compare and do it manually.
You can create a UI Page with a simple diff functionality using Jelly or Angular, etc. You would need to get the records and compare each field in the code.
Export the records as XML files and compare them with external tools: http://text-compare.com/, notepad++, etc.
Hope it helps!
You may want to check AvailabilityGuard. It's a ServiceNow Addon that allows Server Comparison as well as compliance standard checks and out of the box Downtime and Data-Loss risk assessment.
Another option is to check Symantec's CCS. More security oriented but potentially does something similar.
Mark.

How flexible is Pentaho for dynamic transformations? (user-input based parameters)

Based on the following use case, how flexible are pentaho tools to accomplish a dynamic transformation?
The user needs to make a first choice from a catalog. (using a web interface)
Based on the previously selected item, the user has to select from another catalog (this second catalog must be filtered based on the first selection).
steps 1 and 2 may repeat in some cases, (i.e. more than two dynamic and dependent parameters).
From what the user chose in step 1 and 2, the ETL has to extract information from a database. The tables to select data from will depend on what the user chose in previous steps. Most of the tables have a similar structure but different name based on the selected item. Some tables have different structure and the user have to be able to select the fields in step 2, again based on the selection of step 1.
All the selections made by the user should be able to be saved, so the user doesn't have to repeat the selection in the future, only re-run the process to get updated information based on the pre-selected filters. However he/she must be able to make a different selection and save it for further use if he/she wants different parameters.
Is there any web-based tool to allow the user to make all this choices based? I made the whole process using kettle but not dynamically, since all the parameters need to be passed when running the process in the console. The thing is, the end user doesn't know all the parameter values unless you show them and let them chose, and some parameters depend on a previous selection. When testing I can use my test-case scenario parameters, so I have no problem, but in production there is no way to know in advance what combination the user will chose.
I found a similar question, but it doesn't seem to require user input between transformation steps.
I'd appreciate any comments about the capabilities of Pentaho tools to accomplish the aforementioned use case.
I would disagree with the other answer here. If you use CDE it is possible to build a front end that will easily do those prompts you suggest. And the beauty of CDE is that a transformation can be a native data source via the CDA data access layer. In this environment kettle is barely any slower than executing the query directly.
The key thing with PDI performance is to avoid starting the JVM again and again - when running in a web app you're already going so performance will be good.
Also; The latest release of PDI5 will have the "light jdbc" driver (EE customers) which is basically a SQL interface on PDI jobs. So that again shows that PDI is much more these days than just a "batch" etl process.
This is completely outside the realm of a Kettle use case. The response time from Kettle is far too slow for anything user facing. It's real strength is in running batch ETL processes.
See, for example, this slideshow (especially slide 11) for examples of typical Kettle use cases.

Merging two dataset

I have one "big" TOracleDataSet which I can't change 'cause it's using in many different parts of huge project. I want to add just one record to this dataset for using in another grid. The solve way I see it is create another one oracle data set which will combine wanted record and another ones from "big" dataset. In other words, "small" dataset includes "big" dataset.
Try this maybe?
TxQuery Project
This was a commercial project at one time, but the auther was convinced to release it opensource.
TxQuery component is a TDataSet descendant component that can be used to query one or more TDataSet descendant components using SQL statements. It is implemented in Delphi 100% source code, no DLL required, because it implements its own SQL syntax parser and SQL engine.
That quote was taken from the page of the current maintainer, I believe, Chau Chee Yang .
Either this or maybe TClientDataset might be your best options.

Resources