Can we pull data in servicenow ui from a external data source/api? - servicenow

Can we pull data in servicenow ui from a different data source/api?

A few options
Load script to import set
https://docs.servicenow.com/bundle/paris-platform-administration/page/administer/import-sets/reference/custom-type-data-source.html
Remote Table
https://docs.servicenow.com/bundle/rome-servicenow-platform/page/administer/remote-tables/concept/remote-tables.html

Related

Service Now master-detail extract

My users have dashboards in SNOW that shows specific types of RITM.
They need to extract the RITM data alongside with their tasks.
A master-detail view export where the master gets repeated for each detail.
Currently, they manually open one RITM, create a word file, copy RITM content and the content of each task to the file and format it.
Is there a way to export all RITMs, matching criterias, extract its contents and its tasks to an Excel or CSV file or web service calls or a PowerApps connector ? We do have a Mid Server that we could use.
Any advice on how I could automate this extraction of data and building of the word file ?

Can we copy json from exported Power Platform solution and create flows from JSON?

Can we create a Power automate flow from a json file? Something like we do for customer connector.
When you go make.powerautomate.com you can import flow/package using 2 options.
yes can export solution but not just json, In addition you can export only flow from make.powerautomate.com and then import it
Have a look at the Power Automate Web API.
It has a Create a cloud flow endpoint which includes a clientData property which uses json for objects like connectors & actions:
https://learn.microsoft.com/en-us/power-automate/web-api#create-a-cloud-flow

How to export data in CSV in kibana custom plugin using Reporting plugin

how can I export data which I rendered in a dataGrid inside of a custom plugin using data plugin
How did I render the data inside a plugin?
follow -> Git app.tsx
I want to export this data using the kibana plugin
Dashboard plugin has an example for exporting data
To export these datas, you can use the kibana's API.
To find the http request to execute, just use developper's tools in chrome to find a POST request with a URL like /s/echange/api/reporting/v1/generate/immediate/csv_searchsource.
When you find your request, try to execute it in your script : you will retrieve your csv's datas.

How to use existing dynamo db with AWS Amplify and graphql

We have existing database in dynamodb for our application. For one of our new React app, we want to use AWS Amplify and we are trying to use the existing tables.
I created a skeleton project and went to backend AWS AppSync console -> Data Sources and map the existing table to it. And I added that table definition to my schema.graphql in my react app. When I do an amplify push, I see it creating a new table in dynamodb, rather than pointing to the existing db.
I also tried to map one of the tables in AppSync Console, and did amplify pull from local project, assuming that will add the definition of the table in my local schema.graphql but that did not happen either. It did not pull down the details of the newly mapped data source from backend app sync console.
My existing db has lot of data and is shared with other applications as well. I do not want to create a new table.
Can you pls suggest how to accomplish this?
Also we have existing lamdba functions which we would want to leverage into this new Amplify project as well. Could you pls suggest pointers for this as well.
Any help is much appreciated!
Line 22 in the file below shows how the table name is constructed when the graphql.schema transformer runs; that is - the name is derived from the GraphQLAPI ID and this seems to be a one way process.
TableName: joinWithEnv('-', [SyncResourceIDs.syncTableName, Fn.GetAtt(ResourceConstants.RESOURCES.GraphQLAPILogicalID, 'ApiId')]),
https://github.com/aws-amplify/amplify-cli/blob/e1e07b245db0963c4655e646c53e7615febe2930/packages/graphql-transformer-core/src/util/syncUtils.ts
The only option then would be to try and patch the resulting CFN script.
You can use amplify import storage to import existing DynamoDB tables or S3 buckets.
more here: https://docs.amplify.aws/cli/storage/import/
Also look at
Re-use existing AppSync GraphQL API - it might help with reusing lambda functions, if they are wrapped by an API.

How i can import data from parse.com service in my own a mongodb database?

As you know, parse.com will be closed.
I made export of all collections.
Now i have the many dumps of collections.
Collections name like:
_User
AppParentCompany
It's normal tables name, but between this files i have relations, a relation contained in the next file:
_Join:parentCompanies:_User.json
Latest file name i understand like:
_Join - is relations
parentCompany - it's column
_User - a collection containing parentCompany column
Based on this, i have next questions:
How i can import this relations into a database?
What name will be for this collection of relation?
Thanks!
Take a look at this page: https://github.com/ParsePlatform/parse-server/wiki/Migrating-an-Existing-Parse-App
Prepare mongodb database connection string first.
Go to Parse your project's App Settings > General.
Find App Management > Migrate to external database. Press Migrate button.
Enter your database connection string, it will start with mongodb://...
Then press Begin the migration.
Then Parse will do the rest for you.
P.S. If you are using mongolab, go the web browser, into your database, then you can find _Join:xxx:xxx,​​​​​​ _SCHEMA and more in Collections.

Resources