How to check segmentation details of an existing table? I need to retrieve information that were set with SEGMENTED BY HASH(some_column) ALL NODES on table creation. I guess it is possible to query for that.
Depends on your Vertica version. Since 7.1 there is a column SEGMENTATION_EXPRESSION in V_CATALOG.PROJECTIONS.
Related
Ok, tricky question I am trying to figure out where a database schema is storing a particular pointer. I know the pointer value I just don't what table it is in or what column. I know the pointer is 123123123. How do I check all table columns to see if any of them have that value?
Thanks.
In h2 you can use fulltext search, but then you would need to add all tables in the search scope and indexing.
If you need to index only primary keys, then it might be better but you still need to come up with individual FT_CREATE_INDEX() calls for each table. You can automate this with several languages or with ETLs (like scriptella).
If you've enough disk space, you could dump a SQL from your db and use a viewer for big files like glogg.
The advantage of the first solution is no external tools but you need to work out a specific indexing script for SQL for any existing or new table. The 2nd solution is a 1 time fix.
I use SQL Search from RedGate. It's free and it helps you find any text anywhere in the database.
https://www.red-gate.com/products/?gclid=CjwKEAjwiYG9BRCkgK-G45S323oSJABnykKAE7IH_EMhnmq7OdLdXljfIkdGZrDD6OnOrT4VB0agahoCVn3w_wcB
I am using golden gate to replicate table from one DB to multiple DB's. The challenging part is that in one DB the table should be replicated full (all table columns), but in the rest of the DBs the table needs to be half replicated, meaning just a few columns, not all.
Is it possible to have columns exception at the replication level ?
I know that is it possible at the extract level, but this doesn't fit to my scenario.
COLSEXCEPT is an EXTRACT parameter only. It cannot be used in replication.
For tables with large number of columns, using COLEXCEPT can help in excluding some columns instead of entering all the columns in the extract file.
You need to solve this in the REPLICAT side by mapping the necessary columns to the target table using COLMAP. I think USEDEFAULTS will not work in this case for REPLICAT since you mentioned that you need few columns only(Does that mean the table structure is different from SOURCE to TARGET???)
I am new to accumulo.I dont know how to create a table and also columns in accumulo. can anyone suggest me with a sample how to create table and columns in accumulo. thanks in advance.
I think you really need to re-read the user manual at https://accumulo.apache.org/1.7/accumulo_user_manual.html
Accumulo doesn't have columns in the traditional sense of relational databases. Instead of having lots of different columns in a table with a schema, you have rows, which have a ColumnQualifier and ColumnFamily. This is a fairly significant difference from what you are trying to do. Again, re-read the user document and then re-ask/reformat your question.
Are you trying to convert a relational table to Accumulo? That's fairly easy using the Primary Key as your row, then use the Field Name as your Column Family, leave the Column Qualifier blank and finally use the Field Value as your value.
There is no open source gui tool that I know of to view Accumulo data.
Am working on a OWB project. My sources are tables partitioned monthly. I have a control_table that holds the last run date. I want to use OWB to fetch a whole partition (monthly) data into a landing table, enjoying the performance. Kindly assist in coming up with the way OWB treats partition as its source.
Thanks.
I don't think it's a good idea.
Just implement you logic in OWB. If partitions are available, OWB/ Oracle will use them to get any possible benefits.
If the partitions are not used, there should be a issue with the partitioning strategy. Try to correct the partitioning strategy.
Partitioning should be transparent.
There is a specific row in a table that I would like to research. I'd like to know when this row was inserted, when its individual fields were modified, the various values each field in this row might have had etc.. In short, its audit.. Is it possible ? How ?..
I'm using Oracle 11g
You can enable auditing. If this is after the fact, no I don't think there's much you can do.
You can try LogMiner after the fact, but that depends on whether you've got access to the necessary redo log files.