top button
Flag Notify
    Connect to us
      Site Registration

Site Registration

how to call DB2 functions in Informatica ETL?

+1 vote

How to call DB2 functions in Informatica ETL?

Only used views in ETL before, have no idea how to do so for the functions, anyone got any idea?

posted Oct 17, 2014 by Sachin

Share this question
Facebook Share Button Twitter Share Button LinkedIn Share Button

1 Answer

0 votes

Did you check if the Stored Procedure transformation can be used to invoke functions? It is the same as envoke stored procedures.

answer Oct 20, 2014 by Shweta Singh
Similar Questions
+1 vote

How to Import data from db2 table using informatica power Center without powerexchange.

0 votes

I have a stored procedure being called from an informatica mapping. There are no manual commits inside the procedure and the procedure is created with 'COMMIT ON RETURN NO'

I can see the updates done inside the procedure being committed. I think this could be because of the auto commit setting in the Informatica Server. (And I assume the 'commit after end of file' option in the task properties does not apply to stored proc calls)

If this is the case,assuming I do not want to commit anything even if procedure completes without errors, where do I find the autocommit setting (BD2 Driver configuration/Informatica configuration) and change it?

+3 votes

Most of my company uses powercenter informatica for Extract-Transform-Load type data move jobs between databases.

However project I am on has a big custom Perl job with some Java thrown in for good measure to move data and trigger some other updates.

There is talk of rewriting the thing to use powercenter instead, what are people's experiences on such a project, does it make sense? Seems like you trade lot of flexibility in going to such a "off the shelf" solution, but do the ETL tools buy you much in terms of productivity for example?

+1 vote

Can we use Hadoop as a substitue for ETL tools like Informatica for ETL processes ?

+3 votes

I wan't sure how to word this question so I'll try and explain. I have a third-party database on SQL Server 2005. I have another SQL Server 2008, which I want to "publish" some of the data in the third-party database too. This database I shall then use as the back-end for a portal and reporting services - it shall be the data warehouse.

On the destination server I want store the data in different table structures to that in the third-party db. Some tables I want to denormalize and there are lots of columns that aren't necessary. I'll also need to add additional fields to some of the tables which I'll need to update based on data stored in the same rows. For example, there are varchar fields that contain info I'll want to populate other columns with. All of this should cleanse the data and make it easier to report on.

I can write the query(s) to get all the info I want in a particular destination table. However, I want to be able to keep it up-to-date with the source on the other server. It doesn't have to be updated immediately (although that would be good) but I'd like for it be updated perhaps every 10 minutes. There are 100's of thousands of rows of data but the changes to the data and addition of new rows etc. isn't huge.

I've had a look around but I'm still not sure the best way to achieve this. As far as I can tell replication won't do what I need. I could manually write the t-sql to do the updates perhaps using the Merge statement and then schedule it as a job with sql server agent. I've also been having a look at SSIS and that looks to be geared at the ETL kind of thing.

I'm just not sure what to use to achieve this and I was hoping to get some advice on how one should go about doing this kind-of thing? Any suggestions would be greatly appreciated.