I am new to DB2 and I'm working with an IBM i series system. I have an object with a special column "X". I want to store data for X in another object whenever some criteria satisfied and I can get data from the original object for both special records(which data have been stored in another object) and normal data (which values stored in the original object).
How can I link this column to another object? can DB2 data link: data-link help me in this situation? if yes, how can I implement that? I can't find a complete tutorial on how to do this.
I will appreciate any help.
Related
I'm new to building SSIS packages, in fact this is my first package. I need to pull data from one DB view on Azure managed instance to an SQL on prem. I have built out the data flow and all. I'm moving data from a database view into a another database table but the destination table has a column that the source doesn't have hence my destination mapping view looks like (See attached image) How do I fix this or what are my options?
If this columns needs to stay empty in the source and you don't have it in source your best and only option is leave it like this. It basically needs to ignore it so no information will be fed. That will work.
In case you need information as current date you can add derivied column box in between your source and destination in your Data Flow where you can add current date or more columns that come from variable for example.
Its self explanatory that ignore(optional) means mapping for those columns can be ignored and if you want columns to be mapped with any calculated column you can do it by using derived column SSIS component Reference
As per your use case,try to use OLD DB component instead of ADO.NET component
to optimize performance for a relatively large data set
I'm using JScript and SQL Server 2012 to verify a grid in a UI.
Is there a way to compare SQL results to a single column? The entire grid is mapped, but nothing else. Even with the Map Object From Screen tool, I can't highlight a single row or column.
I've searched every property and mapped item I could find, but nothing get me closer to what I need.
Thanks in advance.
Separate rows, columns or cells are not UI objects. Using TestComplete, you can create a Table checkpoint and specify that only a single column should be verified. More information on table checkpoints in TestComplete can be found in the About Table Checkpoints help topic.
I am working on a transformation step for Pentaho Kettle. It selects several input columns and based on that adds two new columns during transformation. I am unable to understand (based on code from other plugins), how I can add the two new columns so that 1) steps downstream are aware of these columns and 2) i can push the transformed data into these columns.
Thanks in advance.
You might need to override meta.getStepFields() to add new ValueMetaInterface objects to the RowMetaInterface passed in. This is the standard way to add columns at runtime; however, the row's metadata (i.e. list of ValueMetaInterface objects) must be the same from row to row or else the next step in your transformation will complain.
Often when doing data-driven custom plugins, you consume as many rows as you need (using getRow()) in order to figure out what the outgoing row format/metadata will be, then you can construct a RowMetaInterface (usually using meta.getStepFields()) that will be passed into the putRow() call. If you intend to pass through the incoming fields, do something like:
RowMetaInterface outputRowMeta = getInputRowMeta().clone();
If you're creating new rows use this:
RowMetaInterface outputRowMeta = new RowMeta();
Either way when you call meta.getStepFields(outputRowMeta, ...) it should populate outputRowMeta with the appropriate fields, by adding/changing/removing ValueMetaInterface objects from outputRowMeta.
I've got a blog post using Groovy to add/replace fields in the incoming rows here:
http://funpdi.blogspot.com/2014/10/flatten-json-to-key-value-pairs-in-pdi.html
Not sure if that is similar to your use case or not. If you have more questions, feel free to find me on IRC at ##pentaho (my nick is usually mburgess_pdi)
IF i have understood your question correctly, i think you are trying to create an output file with dynamic column. So you can do this by checking on the "fast dumping" option in Text File Output Step. While doing so , donot define any column names in the "Fields" tab
Check my image below:
Hope it helps :)
I am a new user of Parse.com and do like what I see so far. I am helping a friend out. They have a locations table that houses lat and long values. However, the developer that created it used an Object data type instead of a GeoPoint data type to store lat/long info. The developer wrote an iOS app that pulls records from parse and stores them locally to display on the map. What I am trying to do is figure out how that developer pulls only the records within X number of miles... or even based on a bounding box. Anyways, my question is if I have an object column with data like what I have shown below is there any way to filter the results by lat/long values (gt or lt etc)?
{"Latitude":33.51882,"Longitude":-93.97484}
It seems like an Object data type is a loose way of storing key value pairs without having to create an entities table etc. However, I cannot find a way to query this subset.
thanks for any help!
I am looking for a way to get the data type of a column given the table that it is in and the name of the column. I have seen other answers refer to the MSysIMEXColumns table but my Access database does not have that table. I am using Access 2003 if that helps any.
I would like to get the result into my LUA code as a hash table so that before inserts or lookups I can look up the data type of column. Basically it's so that I can determine if field needs quotes around it or not.
LuaSQL is being used to access the access database.
Any help or direction is appreciated
The LuaSQL cursor object documentation here suggests that you should be able to use
cur:getcoltypes()
to return "a list (table) of column types".