Is it possible to make a parse.com database available for public download/export? - parse-platform

I'm wondering if it's possible to make a database on parse.com available publicly for download. I have a scraper running on Google App Engine that grabs data from Kickstarter and saves it to Parse, and I want to make it available for researchers, hobbyists, etc.

This is quite broad in terms of how you want the data to be "available" to the "public".
But in general, yes this is absolutely possible. Just write a parse cloud code function that returns the data you want. Then have your frontend call that function, and your frontend will do the interpretation of how to display it (or structure it for download). This method returns JSON to be handled by the frontend.
If you want to make it available for file download (e.g. in the form of a table or something), you can use the parse rest api.
https://parse.com/docs/rest/guide#files
The procedure is to upload it, and the callback returns the url where it can be accessed. The file is public to anyone that knows the url to the file.

Related

Is it possible to create a Google Document directly from a Template's JSON

Since the Docs API allows for retrieval of Google Docs via JSON, is there a way to directly inject that JSON data, with alterations, into a new Google Document? I have several Google Documents that are used as templates for contract agreements, and the data points for all the information to be entered into the templates is already pre-stored in a database.
My goal would be to store the template's JSON server-side and call that data to generate a current agreement and modified agreement by iterating through the placeholder values, inserting database / user input values at those placeholders, and then creating a new Google Doc from that JSON that can be downloaded. Skipping the step of creating google-side copies, grabbing their ID's, and calling something like Batch Update
My inspiration for this came from Federico Tomassetti's post: A template system for Google Docs: Google Drive automation and PDF generation with Google Execution API.
However, Federico creates copies of the template and then fills that copy using App Scripts. I'm currently searching for a solution that would allow for Agreements to be fully generated before being inserted to reduce overhead and simplify permissions, etc. Skipping the step of creating copies and then editing.
I'm just surprised there isn't some kind of way to directly upload without having to call a GET to the template every time, creating copies, and then editing those copies via Batch Update. Hopefully I'm missing some API call that exists, but so far I haven't found anything.

GlideAjax in UI Page with DIRECT flag enabled

I am relatively new to ServiceNow, and I am building some UI pages where I basically do not need any of the SN structure except for Glide Ajax (I need to get data from a Script Include).
The problem is that when I select "Direct" the Glide Ajax functions are not available any more in the client script.
Does anybody know if this is possible to achieve? I searched everywhere without success.
Thanks a lot!
If you check "Direct", it omits all ServiceNow specific JavaScript and CSS. "GlideAjax" is ServiceNow specific JavaScript.
If the data that you are pulling back from the Script Include is static (meaning that you can pull it when the UI page is loaded) then you can probably do it inside an "evaluate" block in the HTML section of the UI page. You could have your Script Include return a JSON object as a string and store it in a variable. Then you could have your client JavaScript parse that variable. I am not a Jelly expert, so I would be curious to know if it works.

Is there a way to programmaticaly upload a whole skill to DialogFlow?

I want to generate the whole action JSONs in code, and then upload them, instead of working through the console.
There is an option to download the whole package as a zipped JSON and theoretically you can also upload. That might supply a hint on how to create the JSONs.
However, these files have all kinds of IDs for the different building blocks, such as Intents or Entity. So, when I export from DialogFlow, of course I get IDs for these.
But, if I want to create a new google action, do I generate these action IDs myself?
Is there documentation on what the structure of these JSONs should be?
The format used by the export/import process is not documented, and while you can try to work with it, there is no guarantee that it will continue to work in the future.
Depending on your needs, it may be better to work with the Dialogflow API (the former API of API.ai). This provides an API to build and modify Intents and Entities (and do some other things). It isn't clear, however, that this provides access to the settings for various integrations.

How to add search to my static laravel site?

I'm building my first ever website using laravel 5.2. Right now, I'm only serving static content with a few API requests for things like the current weather. I've never built a website before, but I'm running my own droplet at DO, so there's no shared hosting limitations.
How would I implement a search that allows users to search my site's content from the main screen? Currently there's no interaction on a DB, it's all just Blade/HTML. I want to avoid using Google Custom Search as there should be no ads, and I want to learn along the way.
Please advise.
It depend on the amount of static contents you have in your site. Maybe you could try implementing it the following way.
Create lookup of the views pages with the most relevant keywords
When user search for keyword which match to any of my relevant keywords, I would load the respective view page.
I would store the the lookup in json format. It would be similar to contents with multiple tags on them.
Here, creating json file is going to be tedious and needs to be done manually.

Fetching data sent by lightstreamer

I'm using (not programming) an application that sends my browser, using a technology called "lightstreamer" (which i have no clue about), data every second or so (I guess using AJAX?). these are constant changing stock values.
Now... is there any program/thing I can use to automatically fetch/sniff/whatever the raw data that my browser gets, so that for example i could later paste it to Excel and create charts?
Why not just copy the data from the browser window you might ask, and the reason is that the application always shows me only the last 20 values for a given stock, and i wish to automatically get, let's say, the last 1,000 values and throw it to Excel.
Thanks :)
PS I see that the app is written in asp id it matters.
There is no way that I know of to (easily) reconstruct sniffed Lightstreamer communications into tabular data. Lightstreamer pushes updates to the client using a hidden IFRAME, but those updates are efficient, but intended for consumption only by the Lightstreamer client code.
Developers using the Lightstreamer Javascript API can easily hook into update events if they wish to.
However as an application user, you are best off raising a change request with the application owner to add some form of Excel export functionality.

Resources