Free CVS reporting tools - reporting

Please suggest me good and free reporting tools which works with CVS in Windows environment? The purpose of the reporting is to get the most detailed information from a CVS repository, including committing activity on a per user basis.

Checkout statCVS.
Here is a brief set of feature list from their site:
Timeline for the lines of code
Lines of code for each developer
Activity by Clock time
Authors Activity
Author activity per Module
Author Most Recent Commits with links to ViewVc
Stats per directory
File count
Average file size
Largest files
Files with most revisions
Directory Sizes
Repository Tags Number of LOC per version.
Repository tree with file count and lines of code
LOC and Churn the evolution of LOC and the amount of change per day
Repo Map the dynamic hierarchical view of your repo for the last 30 days

Related

Automating Validation of PDF Prepared Reports

Our team uses Spotfire to host online analyses and also prepare monthly reports. One pain point that we have is around validation. The reports are all prepared reports, and the process for creating them each month is as simple as 1) refresh the data (through Infolink connected to Oracle) and 2) Press button to export each report. The format of the final product is a PDF.
The issue is that there are a lot of small things that can go wrong with the reports (filter accidentally applied, wrong month selected, data didn't refresh, new department not grouped correctly, etc.) meaning that someone on our team has to manually validate each of the reports. We create almost 20 reports each month and some of them are as many as 100 pages.
We've done a great job automating the creation of the reports, but now we have this weird imbalance where it takes like 25 minutes to create all the reports but 4+ hours to validate each one.
Does anyone know of a good way to automate, or even cut down, the time we have to spend each month validating the reports? I did a brief google and all I could find was in the realm of validating reports to meet government regulation standards
It depends on 2 factors:
Do your reports have the same template (format) each time you extract them? You said that you pull them out automatically so I guess the answer is Yes.
What exactly are you trying to check/validate? You need to have a clear list on what are you validating. You mentioned month, grouping, data values (for the refresh)). But the clearer the picture you have for validation, the more likely the process can be fully automated.
There are so called RPA (robot process automation) tools that can automate complex workflows.
A "data extract" task, which is part of a workflow, can detect and collect data from documents (PDF for example).
A robot that runs on the validating machine can:
batch read all your PDF reports from specified locations on your computer (or on another computer);
based on predefined templates it can read through the documents for specific fields that you specify (through defined anchors on the templates) and collect the exact data from there;
compare the extracted data with the baseline that you set (compare the month to be correct, compare a data field to confirm proper refresh of the data, another data field to confirm grouping, etc.);
It takes a bit of time to dissect the PDF for each report template and correctly set the anchors but then it runs seamless each time.
One such tool I used is called Atomatik. It has a studio environment where you design the robot (or robots) and run the process.

30 million row monthly CSV that requires fields added on each line based on a lookup from a separate file

I have a 30 million row CSV that gets created each month, I am trying to add 2 fields that are populated based on a Lookup from a separate file and let it run unattended. I am trying to choose the technology right now - I'd rather use a scripting language that can be run from the command line (Windows) and something free ideally but open to suggestions. SQL database not really an option.
Take a look at Pentaho Data Integration. It’s Java based, multi-threaded and can cope with large CSV files at 100k+ rows per second.
You can call it from the command line in either linux or windows, and can parametrize the jobs and transformations to take command line parameters for things such as file paths, db connections, etc.
There’s a paid for Enterprise Edition but also a free, open source community version.
See community.pentaho.com.
Beware: steep learning curve. Shout if you need additional pointers.

Retrospective updates

We are planning to use the "history" interaction to support viewing of historical snapshots of a resource (for example, viewing care event details for an encounter as historical snapshots of the encounter)
For example,
GET encounter/{id}/_history/{vid}
We wanted to use the same structure to perform retrospective updates to a particular history entry using a PUT interaction
PUT encounter/{id}/_history/{vid}
However, there seems to be a restriction in doing so as mentioned here
Accordingly, there is no way to update or delete past versions of the
record, except that the metadata can be modified (mainly for access
control purposes)
Is there any other mechanism for performing retrospective updates?
There is no mechanism to adjust history. History does not represent "the history of what occurred". It represents "the set of versions that existed on this server at a particular period of time". As such, short of time travel, there's no meaningful need to change history records. If you wanted to assert multiple separate things about a resource at different times, you could create multiple instances and link them together using Linkage - e.g. A Condition that had one severity for a year, then escalated for 2 years, then went into remission for a year, then came back, then got resolved could be represented using multiple Condition records each with different effective periods. Linkage could be used to indicate that they were all talking about the same Condition. And all could be created "now" as the time when the server first became aware of that historical information.

Tracing a single Issue

Problem: I have selected few issues. Now, I want to trace an issue within the source code files starting from the moment it was first detected as an issue until it is repaid/resolved/removed/deleted/remaining in the latest repository.
So, for each unique issue (unique to an specific source file), I want a list that has N rows (N = number of analysis, e.g., SNAPSHOTS) where each row shows the existence of of the issue in a source file (preferable also with its location in the source file).
Questions: Apparently, I couldn't find an API for this. When I explored the database, I was unable to establish a connection between SNAPSHOTS and ISSUES tables that I could use to separate issues from one SNAPSHOT/analysis to another.
Do you see any way to solve the problem?
How can I separate issues from one snapshot to the others?
What is the format/encoding of the LOCATION field of the ISSUE
table? Can this be used to identify an issue location in the source
file?
Relation between issues and analysis is not persisted over time. Still each issue has a creation date, the date of last change (status, assignee, ...) and optionally the close date. That allows you to match issues with the dates of analysis.
As a side note, the database must never be accessed by plugins nor external applications. The only API to extract is provided by web services, api/issues/search and api/issues/changelog in your case.

In Visual Source Safe is there a limit to how many labels you can apply?

We have a collection of 25 or so projects in a source safe database and we want to apply daily labels to each of them. Is there a limit to the number of labels that can be applied?
We've been using one repository for 5 years so far. It stores numerous projects and something like ten daily builds run every day on that repository. Every daily build creates a new label on some project every time it runs. So far we haven't seen any problems with possible labels exhaustion.
No, there is no practical limit. We once did what you ask about - daily builds applied a label with the build number, this went on for several years.

Resources