Bug in Fuzzy Lookup in Visual Studio SSIS - visual-studio

I'm using Visual Studio Pro 2013 and want to use the Fuzzy Lookup task but there seems to be a bug that prevents the component from connecting to the reference tables.
A reference file and table is specified in the Connection tab which all seems fine but the Columns tab is also needed to create the reference links between different fields of the data however there is nothing there:
The error messages are as follows:
I've read elsewhere this was a known bug on older versions of SSIS from about 2005 - anyone know what the problem is here and how I can fix this?
I should add that the connection manager and the table seem fine as they have been used many times elsewhere in the project. I've tried recreating the data flow in a new document and even restarting my PC but this simply won't work. I should add that I've not used the Fuzzy Lookup before but have looked at several references and know that the column tab should be populated with data and not be an issue.
Many Thanks,
Kw

Not a bug, it's how the product works. From the manual, it specifies the reference table must be a table in sql server. A table in Access, therefore, is unsuitable for use in the fuzzy lookup component.
The transformation needs access to a reference data source that contains the values that are used to clean and extend the input data. The reference data source must be a table in a SQL Server database

Related

Microsoft Office 14.0 access database engine ole db provider in visual studio 2015

This is my first attempt at writing a Visual Basic app using Visual Studio. I am using an Access 2013 database as the source and everything was going fine until I decided to add some calculated fields to one of the tables. Now VS is telling me that I need the "Microsoft Office 14.0 Access Database Engine OLE DB Provider" to read the tables.
I've been searching and searching but can't find anything about version 14.0 of the engine. Am I just missing something simple, or can I not access calculated fields in the DB and have to just pull the data and do the calculations in the app?
I am working on a machine with Win7 64bit, Office 2013 Pro, and Visual Studio 2015 Community.
Thanks!
EDIT: The calculated fields look like this:
"http://travellermap.com/api/jumpmap?sector="+Left([Sector],4)+"&hex="+[Hex]+"&style=print&jump=3"
Where [Sector] and [Hex] are fields in the same database. As soon as I added the fields to the database VS2015 would throw the error. If I tired use the Dataset Wizard it would tell me that there were fields in the database that were missing or could not be read without v14 of the engine.
Then even after I removed the calculated fields it still could not read the data. I have to revert to a backup copy of the database that never had calculated fields in it to get things back to working.
My exact steps were as follows:
Create database with multiple tables
Create a view that pulls together data from all the tables that I needed.
Create dataset in VS that uses that view which I called SystemMaster
Build app around it using the XAML designer
Decide that it would be quicker to use a calculated field int he DB than write the VB code to put together a URL that is needed to display an image in the app.
Close out VS2015 and laucnh Access 2013
Create the calculated field in the primary table and edit the SQL code for the view to include those fields
Save and close Access, go back into VS2015
View the dataset to see if the new fields are there, see that they are not and launch the Wizard only to find out that the Wizard says it can't access anything in that view anymore because fields are either missing or can't be read with the current version of the Access engine (12.0) and that I need version 14.0.
Search for answer, don't find one. Remove the fields from the tables and try to get back to a working application.
Realize that even after removing the fields, VS2015 won't read the tables ro views that had them.
Revert to a version of the database that did not have the calculated views, and everything works again.
While I know that the proper way to do this would be to construct the URLs in the VB code, this was easier and quicker (in theory) to do at this point which is basically a proof-of-concept/prototype.

SSIS value does not fall within the expected range with OLE DB Datasource

I'm using Visual studio 2013 with update 3 and a collegue of mine with update 4 installed each. We are using the data dools for sql server 2014.
I've created a few DTS packages which my collegue updated so far it worked without problems.
But all of a sudden I get "value does not fall within the expected range" warning from the datasource and can't edit columns there,.. . I needed to recreate the datasource for the message to disappear again.
My question here is can it be that the appearance of additional columns in the table which the datasource accesses was the cause of this problem? (I've seen out of sync warnings for datadestinations whenever a destination table got new columns or lost columns, but this is the first time something changed for a source table).
Or can that problem have a completely different cause?
It has been a long while since I've worked on an SSIS project, but I do recall seeing this error as well.
My experience was that it was caused to the metadata of the input being out of date in a certain way, and what you describe as your suspicion fits with this.
The solution I found to avoid this was to be very specific on all my input components, selecting the exact columns I wanted rather than selecting all. I think in the end I actually changed them all to use hand written SQL queries rather than the GUI column selector.
Also I don't remember if this was the same error but a similar one: sometimes after a schema changing when trying to open a component the GUI would throw an error and not open but when I tried again it would have resolved the error.
Sorry I couldn't be more definitive in my answer but hopefully this information helps point you in the right direction.
I used a simple method and it is working fine. In the OLE DB Source Editor while I retained the same connection manager, changed Data access mode (from Table/View) to SQL command and used SQL command to select the required columns. Error message no longer appeared and I could see the column values....

ODAC/Entity Framework Model Update Issues

We are looking to introduce odac into our application but I am running into a number of issues and I can't seem to find any solutions that fix our issues.
We are using an oracle database and trying to use ODAC 12c Release 1 (12.1.0.1.0) with Oracle Developer Tools for Visual Studio.
In our model we would like to have multiple schemas so we can perform cross schema queries. The schemas which we select in the filter for the database connection appear when we are creating the model. However when we try to update our model from the database only the default schema is visible. Sometimes this can be fixed by opening the relevant part of the database in Server Explorer in Visual Studio but this doesn't always work. This fix never works after we add multiple connection strings for the same model (depending on the location of the user will depend on which database the user gets directed to).
The next issue we are having is that we can't the return types of stored procedures to be auto-generate. I have tried to retrieve the column information but it is never able to retrieve the metadata. I have seen a few suggestions of modifying the stored procedures then getting the column information and reverting the stored procedures but this is something we would like to avoid. Also the suggestions don't seem to work on oracle databases (but that could be me, I don't have much experience with databases).
The final issue (this is a minor issue) is that I cannot figure out how to get the generate code to omit underscores from the classes/methods generated. This isn't a huge issue it is purely to make it easier migrating the code we have.

Visual Studio 2010 - How can I control the order of schema updates when using schema compare?

I've been attempting to use Visual Studio 2010 schema compare to take updates from a Dev database and move it to a UAT environment.
The compare itself works fine, but the tool continually orders the update scripts incorrectly.
It will try to update a stored procedure first, then the view that the procedure depends on. If my view includes new fields that the procedure depends on, then it will fail the update.
I've attempted to force the dependency to be recognised by qualifying all references to the dependent views with the schema name (essentially dbo.view rather than view), as suggested in http://msdn.microsoft.com/en-us/library/aa833294.aspx
Is there any way to force the scripts to a particular order (tables, views then sprocs), or is there a way to tell how and why the dependencies are calculated so I can see what's going wrong?
I don't think that either of the things I was hoping to do are possible.
What I have learnt is that the refresh on the schema compare doesn't always seem to recalculate dependencies correctly.
Closing it and starting a new one worked, just refreshing the original didn't.

VS2008 DataSet Wizard doesn't match tables for updating

first question ever on this site.
I've been having a real stubborn problem using Visual Studio 2008 and I'm hoping someone has figured this out before.
I have 2 libraries and 1 project that use strongly typed datasets (MSSQL backend) that I generated using the "Configure DataSet with Wizard" option on in Data Sources. I've had them working just fine for awhile and I've written a lot of code in the non-designer file for the row classes. I've also specified a lot of custom queries using the dataset designer. This is all work I can't afford to loose.
I've recently made some changes to re-organize my libraries which included changing the names of the libraries themselves. I also changed the connection string to point to a different database which is a development copy (same exact schema).
Problem is now when I open up "Configure DataSet with Wizard" to pickup a new column I've added to one of the tables it no longer matches the tables correctly in the wizard. The wizard displays all of the tables in the database and none of them have check boxes next to them (ie: are not part of this dataset). Below those it shows all of the tables again but with red Xs and these are checked. Basically meaning that Visual Studio sees all of the tables it currently has in the DataSet and sees all of the tables in the database, but believes they are no longer the same and thus do not match!
I've had this same thing happen quite awhile back and I think I just re-built the xsd from scratch and manually copied the code over and then had to redefine all of the custom queries I built in the dataset designer. That's not a good solution.
I'm looking for 2 answers:
1. What causes this to happen and how to prevent it.
2. How do I fix this so that the wizard once again believes the tables in its xsd are the same tables that are in the database (yes, they have the exact same names still).
Thanks.
The dataset designer uses the default query (The first one with a check on it) to sync up the schema for each table. Whenever you go to edit the default query, VS will actually connect to your datasource and look for changes in the query. If new columns are added, they will show up as new columns for you to add to your table. Renamed columns show up as new, since VS doesn't have any way to know that you changed the name.
Answer 1. The XSD file contains the names of the database tables that it used to create the table originally. If you change the name of the table, the designer won't know which table to sync to.
Answer 2. You can edit the XML inside the XSD file. Do a "Find and Replace" inside the XSD file replacing the old tablename for the new tablename. Make sure you have a backup of the XSD file before you do. Be careful to only change instances of the old table name and not any other working XML.

Resources