Array bound of exception in Split Text Processor - hortonworks-data-platform

I am using nifi 1.1.1 package. I applied the patch files in the source code by referring the below link due to issue faced "Destination cannot be within sources" while split flowfile when make header count greater than 0. 
https://issues.apache.org/jira/browse/NIFI-3255. 
After apply patches, split text processor works fine, if the header line count given as 0 and above 1.
Those changes in Split text processor can works if we have lesser number of rows only. For example: if flowfile contains 1000 rows it could be split.
If the input file contains more than 20000 rows then it doesn't splits the data and leads `"ArrayIndexOutOfBoundsException" exception.
Here i attached image in which faced. 
Anyone please guide me way to resolve that issue.
https://i.stack.imgur.com/UNKI0.png

After some digging it seems that you have run into a problem in the 1.1 version of Nifi.
As discussed here, upgrading to Nifi 1.2 or above should resolve the issue.

Related

Add Additional Text when repeated group header present in Birt Report

Using BIRT 3.2.22
In my reports I need to conditionally display some text if the same Group Header is repeated. The repeating Group Header should show some text like 'Continued...'
I followed article mentioned below that doesn't help me to resolve the issue.
https://www.eclipse.org/forums/index.php/t/157862/
Can anyone help me on how to do this?
In addition to the old forum entry you mentioned in your question, it is probably necessary to use separate RunTask and RenderTask to make this work.
I never tried something like this myself.

STRAPI is cutting the last 3 chars

I'm using the strapi 3.0.0-beta.15
DB - mongo 4.2.2
After updating one of model fields - which is type of 'text', it is cutting the last three chars and set it to '000'. Saw this only if there are numbers at field.
Writting text (before saving
After saving
Did anyone have stacked with such problems yet?
This issue has been fixed in the last version of Strapi (at least beta.18)
I suggest you update your application using a migration guide if necessary, or, using the application update guide that is into the documentation.
Here is what I did in a short video - https://www.loom.com/share/23076126176546708f9dcaa7eac1924c

Jmeter http simple server(plugin) ADD unique line does not work

http://localhost:9191/sts/ADD?FILENAME=/tmp/newEventsFlag.csv&LINE=7ddb876ac39c485a&ADD_MODE=FIRST&UNIQUE=TRUE
If i run this multiple times and then do a GET http://localhost:9191/sts/LENGTH?FILENAME=/tmp/newEventsFlag.csv
I can see that the body length is equal to the number of posts I made.
So unique parameter does not work I guess or am I missing something here?
If it does not work is it possible to get the source code to fix this?
This is a bug in the simple table server version 2.2.
Should be fix in 2.3

SSRS error on preview : "The size necessary to buffer the XML content exceeded the buffer quota" hides original error

I understand that there is definitely something wrong with my report (e.g. columns missmatcch) and I need to correct it but what I see is the WCF error message that hides actual problem and exactly this hiding irritates me much more than original problem: columns missmatch.
I guess we need to adjust the WCF 'buffer size' and we will get original problem message. But where is the config file?
Text search of "system.serviceModel" in the C:\Program Files (x86)\Microsoft Visual Studio 10.0 doesn't bring good idea...
P.S. Since this is just preview of report I do not think that it is SSRS configuration problem. Problem localised somewhere in DevStudio process or int the DevStudio's internal web server process ...
P.P.S Please help me too improve the question. I see that responders doesn't understand what kind of help I need.
I have encountered multiple "flavors" of this bug in SSRS Preview. It seems the renderer for Preview mode is quite fragile.
There is a simple way to solve this. Ignore the error and attempt to upload the RDL file to your reporting server. The uploader will happily tell you exactly what is wrong with your file - it will tell you exactly what field has a problem and what that problem is. If there are multiple errors, you will get told each and every field and the error associated with each one.
I can create this bogus XML buffer error with any of the following:
Add a new Tablix, start to connect it to a dataset, then cancel out.
Copy/paste some text into a textbox from a MS Word document where one or more lines have a negative right indent (right column end is outside page margin).
Connect a dataset with a varchar(8000) returned value.
Please Check if any of your report items are referencing fields that are not in existing dataset scope.
This indeed worked for me.
See Below link for more information:
http://connect.microsoft.com/SQLServer/feedback/details/742913/ssdt-reporting-services-designer-error
I have seen this error when adding a new field to an existing dataset by clicking "Refresh Fields".
The dataset source was a stored procedure. The result was only a few of original fields showed up in the dataset field list and not the new field. If I tried to preview the report I get XML buffer error.
Workaround was to not refresh fields but hit add new field and type the new field name into the dataset properties field last.
Worked fine after that.
I got this error again today.
I had created a table to hold data to replace two slow queries. I changed some names to clean up the process.
I think the error actually means that there are so many problems with my report that the buffer holding the various error messages isn't large enough which leads to the error message.
The size necessary to buffer the XML content exceeded the buffer quota
Of course this should be an easy fix but Microsoft has said that they will not fix it.
https://connect.microsoft.com/SQLServer/feedback/details/742913/ssdt-reporting-services-designer-error
EDIT: I've updated my answer based on having fixed the issue.
I'm currently experiencing this problem after having changed multiple stored procedures and updating the dataset names in the SSRS report.
And when I try to run the preview I get the exact same error.
As it turns out, after investigating the issue, the problem was that I had changed the name property of my datasets.
There several places in my report where formulas or expressions use the old name properties of the datasets I renamed. After reverting the dataset names back, I managed to get the real errors like missing fields etc. atcual errors came back after I set my dataset name properties back to what they were.
I only changed the name property back to what it was, the stored procedure names were correctly referring to my renamed stored procedures.
I had this problem when after copying and pasting a tablix, it changed CDbl in a formula to Microsoft.ReportingServices.RdlObjectModel.ExpressionParser.VBFunctions.CDbl. I opened up the XML and removed all instances of "Microsoft.ReportingServices.RdlObjectModel.ExpressionParser.VBFunctions." and the report then worked.
For a working report, when I tried to add a column it gave me this error. I edited the .rdl file using notepad++. After SSRS prompt to reload the change from disk, it worked without issues.
I got this error after copying my Custom Code to Visual Studio for hightlighting the code for better readability. Well, Visual Studio added class definitions to the beginning and end of the file. After editing code, I pasted it back to report Custom Code, then got this error. Fix was just to remove class definitions (Public Class Class1 and End Class) from Custom Code. So, check your Custom Code also (if any).
I got this error after adding some new parameters to an existing report.
For some reason when I created the parameters first then modified the Dataset to use the new parameters I got the error, but when I modified the Dataset first then added the parameters second and I did not get the error.
This seemed like very strange behavior to me so I tested it by restoring the report from repository and repeating the process three times with each method, and had identical behavior every time.
I am also facing this problem. I solve this Find and replace
Microsoft.VisualBasic.Interaction.iif ==> iif
Microsoft.ReportingServices.RdlObjectModel.ExpressionParser.VBFunctions.cdbl
==> cdbl
I hope this may helps someone. Thanks
Possible root causes
Parameter name is incorrect(case/order)
Accessing non-existing property.
and many more...
Solution: To get the exact error message are
Deploy SSRS report and find the error : Suggested by "Kim Crosser" already
Remove the section(SSRS/Report content) temporarily you feel is error free to free space in buffer so that you can get actual error message. Later add sections back to the page(removed earlier).
I had the same error message and it was totally caused by my doing. It's a bit embarrassing, but if it helps someone out then great! I had accidentally copied my dataset query that included a small sub select statement within it, which I was using to check parameter/variable values.
Another solution is to open the .rdl file in Report Builder 3.0 (as opposed to Visual Studio) and try to preview it. I found this gave me the details of the error, although if more than one error is present it only shows the first.
I previously binded a TextBox to
Fields!FieldName
and fixed it with
Fields!FieldName.Value
With that said, and with the other answers posted as well, this error happens in different flavors. My issue was fixed after I had the field property "Value" included.

geokettle shapefile input transformation not generation row numbers

Does anyone know why, on GeoKettle, the 'Shapefile File Input Step' is not generating the row number?
How can I do a around? With the 'Add Sequence' step?
This is all happening on a transformation.
Tks
After some amount of digging on the inners of this, what I came out with was that this is simply a bug on the implementation of the step "Shapefile File Input". I did not find this on other inputs that generate the same row number...
The GeoKettle version is 2.5 as of this answer and in turns is based on Pentaho Kettle version 3.2...

Resources