top button
Flag Notify
    Connect to us
      Site Registration

Site Registration

In informatica, my mapping took long time to fetch rows but reads within a minute

+2 votes

In my mapping it tooks long time to fetch rows to the target table but it reads those rows from .CSV file within a minute.In this mapping One source is a flat file another source is a table which is having data from target.Here we create a logic to UPDATE or DELETE or INSERT in target based on comparison on the data from flat file and the source table(data from target table).while seeing sesion log it reads data from flat file within a minute but it fetches those data to the target to taget 9 rows/sec throu enter image description hereghput.Here the target table is created using unix script ie CREATE TABLE STG_LM_INSTITUTION as (SELECT * FROM LM_INSTITUTION);Target in this mapping is STG_LM_INSTITION.this is the copy of final target(LM_INSTITUTION).I think the problem is due to creation of the taget table in script but i'm not sure.Anyone please help me to solve this issue.Source flat file has 2L rows. I run this mapping with 2L rows.After 11hrs it fetches only 1L records into the target.But while running using 500rws only it fetches records in one minute to the target .

posted Dec 23, 2014 by Amit Sharma

Share this question
Facebook Share Button Twitter Share Button LinkedIn Share Button

1 Answer

0 votes

SQL_LM_INST_SEQ retrieves PK values from an Oracle sequence and that's costly, because for every inserted row a roundtrip to the database is necessary to get a new ID.

Use a Sequence Generator transformation instead - Integration Service will generate IDs on its own.

answer Dec 29, 2014 by Shweta Singh
Similar Questions
0 votes

I have some files that I would like to consolidate into a single database table. The files have similar but different formats. The files look something like this:

•ColA : string
•ColB : string
•ColC : string

•ColAA : string
•ColBB : string
•ColCC : string

•Col01 : string
•Col02 : string
•Col03 : string

The destination table looks like this:

•ColFirst : string
•ColSecond : string
•ColThird : string

I want to develop a mapping that ETLs these three files into this one database, but because the column names are different, it looks like I'll have to develop three different mappings, or three different sources, or three different somethings. The problem is that my example is contrived: I actually have many different files that all have different formats and column names, but the data is all very similar.

I would like to develop a single mapping or workflow that can handle all of this by only adding a table that holds the column mappings. Such a table would look like this based on the sample files and sample table above:

enter image description here

In this way, to edit a column mapping I only have to make an edit this this TableMappings table. I wouldn't have to make any changes at all to the mapping or workflow. Nor would I have to redeploy an application.

What would a mapping or workflow look like that could take advantage of something like this? I assume there'd be a flat file source that takes files from a folder. There would be something in the middle that uses this TableMappings table to map column names. Finally there would be a relational data object that represents my destination database table "TableDestination". I don't know how to put this together though.

0 votes

I want to copy the data into one excel file, while the mapping has been running in informatica.

+1 vote

My source having 10 records but how can i load 20 records in target, i am not bother about duplicates. How I can do it?

0 votes

Suppose if we have duplicate records in a table temp_n. Now I want to pass unique values to t1 and duplicate values to t2 in single mapping? How can we achieve?