I have Oracle 11.2.0 (x64) installed on my personal development environment. When running SQL selections that Left Join more than 10 tables I am easily to get Oracle crash with ORA-12537: Network Session: End of file. If I use Inner Join there is no such problem. For a same particular query this happens all the time, no matter how many data is in my database (in fact not much data is there, no more than several hundred rows.)
I work with around 100 people who install Oracle with same version and roughly same settings, and apparently not seen anywhere else as far as I know.
This error can be triggered no matter if I run the query from SQL Developer or using C#.Net Data Access.
Any ideas please?
My Alert/Log.Xml has these:
Exception [type: ACCESS_VIOLATION, UNABLE_TO_READ] [ADDR:0x45] [PC:0x35797B4, _kkqstcrf()+1342]
and
Errors in file c:\oracle\diag\rdbms\orcl1\orcl1\trace\orcl1_ora_1192.trc (incident=41028):
ORA-07445: exception encountered: core dump [kkqstcrf()+1342] [ACCESS_VIOLATION] [ADDR:0x45] [PC:0x35797B4] [UNABLE_TO_READ] []
Related
I know so little about Sybase that I don't even know whether database link is the correct term in Sybase. What I'm looking for is a way to communicate with Sybase over Oracle's heterogeneous services, commonly called a database link in Oracle (linked server if you speak SQL Server).
We are a primarily Oracle shop with a healthy dose of SQL Server, but zero Sybase experience. As you might guess, we have a piece of software that happens to run on Sybase, and we need to get information out of its database.
Professional services from the software vendor have not been helpful at all; frankly, we don't even know if they gave us the correct driver.
Our Oracle DBA is as sure as he can be (without knowing anything about Sybase) that he has everything set up correctly on the Oracle side.
When we try to run a query from Oracle to get data from Sybase, we get:
select * from call_request#sybase
*
ERROR at line 1:
ORA-28500: connection from ORACLE to a non-Oracle system returned this message:
[SAP][ASE ODBC Driver][Adaptive Server Enterprise]Incorrect syntax near the
keyword 'where'.
{ZZZZZ,NativeErr = 156}
ORA-02063: preceding 3 lines from SYBASE
Complaining about syntax near "where" doesn't make much sense, since there is no predicate at all. And it doesn't help to add a dummy
where 1=1
We're completely unsure where to turn. What can we even check next? Where is the documentation for setting this sort of thing up? Any ideas what is wrong?
I am connecting to oracle 12 in Oracle cloud, from PowerBI Desktop windows server 2016.
Oracle client is installed and TNS file configured.
Oracle is hosted by a vendor so my only access is to directly query the database.
In powerBI, when using an oracle connection, i get ORA-03113 errors about 50% of the time when refreshing data. There is no discernible pattern to the appearance of the error.
If i connect via a System ODBC connection set up in windows, I dont get any issues or errors, although the data load is a bit slower.
I would appreciate ideas on what may be causing this issue or what to check to help get more information.
I'm afraid your issue needs some deeper analysis as ORA-03113 might have various reasons, but typically it means that the 'oracle' executable has terminated unexpectedly once there was an existing connection. You should try to isolate the SQL command that is executing when the error occurs. It can be done either by checking the trace files on the server or by using SQL*Net trace if you don't have access to the server. If a statement can be isolated which consistently raises the ORA-3113 error, then it can be further analysed (like execution plan, triggers, etc), or maybe the best to raise an SR so Oracle Support can work on the issue. If you have access to Oracle Support you can find more information about ORA-03113 troubleshooting in MOS Doc ID 1506805.1. Let me know if I can help you any further.
I'm running into a problem while linking some tables and views present on a Oracle 11g database to a Access 2007 file.
I'm using the Oracle Client (SQORA32.DLL) version 11.02.00.03.
If the view/table returns a small amount of data, there's no problem. The problem happens when the view or the table returns a "large" amount of data. I've tried to increase the buffer size on the driver (default is 64000) to see if that happens. I've also removed the "Enable query timeout" option - otherwise I would get a "Query cancelled by user" or a "ODBC - Call Failed" error.
In order to link the tables/views, I've used the "native tool" (External Data -> ODBC Database -> Link to data source by creating a linked table).
I was wondering if I could retrieve the data from the tables/views using vba. Sometimes, I (you should read "I" as "the users") may need to update data on some tables (control tables).
Please let me know your thoughts.
EDIT: Our goal with this project was to migrate from SQL Server 2005 to Oracle 11gR2. After analyzing the behaviour of the Access files regarding the SQL Server, I've concluded that the results are showing like a "cursor" - if you scroll down on the result window, it will load more.
I think that this may be the issue because, AFAIK, Oracle (driver, maybe?) pulls everything from the DB and, only then, populates MS Access.
It's a long time after this so here goes the solution. MS access has a flag for the ODBC connection as "Treat Float as Numeric". This have made the trick.
When enabling Fast Load in Attunity Oracle Destination Components in several similar SSIS packages using Oracle 11g as target - a few packages fail and return the error below, but the rest of them work fine.
The error message I get is:
Description: Fast Load error encountered during PreLoad or Setup
phase. Text: ORA-39826: Direct path load of view or synonym (
TABLE_NAME ) could not be resolved.
If I'll disable the Fast Load, those that failed would work fine too of course.
More importantly, the failing packages work fine with Fast Load when using Oracle 10g as target.
I don't understand why it doesn't work in those that failed.
What am I missing? What should I do to make the Fast Load work at all times and not sometimes?
probably a driver issue.
The 'fast load' option internally uses a BULK INSERT statement for uploading data into the destination table instead of a simple INSERT statement for each single row. Since bulk insert is a native sql server function you should try to understand how does it work for oracle. It probably changed from 10g to 11g
By pure chance, I discovered that the target component fails in an SSIS package if its ‘TableName’ property contains spaces before or after(!) the name of the table. Once deleted it works fine.
This error didn’t occur on 10g.
I'm working on a legacy application using VB6 and Classic ASP. We're using disconnected ADO recordsets to pass data back and forth. Normally this works. But what has started happening recently is for any inner/outer join, ADO is including these in the available records to choose from. So when were specifying a column to update (in the cases it errors out, the primary key column), it in turns updates the wrong column (with the same name). I know it's normal for ADO to pull the primary keys for any joined tables, but the default for this is for ADO to hide them. In our case ADO isn't hiding them.
What I've narrowed it down to is the SQL Native Client driver is not working correctly. I can go back to the SQL Server driver (SQL 2000) and it works great, but as soon as I switch back to SQL Native Client, it exhibits the behavior above. I've checked the properties on the open connection and the properties of the recordsets themselves, they match in every instance except one (the count of how many hidden columns there are which makes sense, as SQL Native isn't hiding them).
I've tried everything from deleting the MSADC folder from IIS and re-adding it, I've uninstalled SQL Native and reinstalled it (and subsequently upgraded it to the newest version). I've recreated the ODBC connection several times as well in the process of troubleshooting it. At this point I'm at a loss.
Also one thing to add, it appears SQL Native Client works fine on our other servers and no one else is having this issue. Anyone might have an idea of what could be happening? Thanks!
Edit : Example of what's happening (this occurs on for any query (stored procedures if it matters) and with >= 1 joins of any kind)
select temp_id, temp_value on temp_test
inner join another_table on another_table.temp_id = temp_test.temp_id
inner join yet_another_table on yet_another_table.another_id = another_table.another_id
this'll produce in the ado recordset :
SQL Native Client
(0) temp_id
(1) temp_value
(2) temp_id (primary key of another_table)
(3) another_id (primary key of yet another_table)
SQL Server driver
(0) temp_id
(1) temp_value
SQL Server 2005 will show it as it should be as : temp_id, temp_value
this occurs on for any query (stored procedures if it matters)
It's not the issue described here is it? :
If a change in the connection string changes the behavior, I would suppose that you have two different schemas, and then two versions of the same stored procedure; and the one that is executed with SQL Nativ Client is the incorrect one.
I have exactly the same scenario, and have had it for over a year on our servers and servers at our client. I never found a solution and as a result we simply have to use the SQL Server driver, which is a shame as SQL Native seems to connect significantly faster.
It's nothing to do with different schemas or different versions of the same stored proceedure as suggested above. I use a file dsn and simply changing the driver name changes the behaviour to/from that mentioned above. It seems to happen to all views (probably stored proceedures too as indicated)
If anyone does find a solution I'd be keen to hear about it.
Warwick