Skip to main content

Aggregating data from a number of rows into a single rows & Converting rows into columns in Oracle

For Aggregating data from a number of rows into a single row we have various functions in oracle some of them are listagg function, xmlagg and wm_concat function

Consider the following example

Dept No
Emp Name
10
John
11
Neo
12
Anderson
12
Trinity
11
Morpheous
12
Smith

Now we want the data in the following way aggregated based on Dept No

Dept No
Emp Name
10
John
11
Morpheous,Neo
12
Anderson,Smith,Trinity

We can achieve this by using the LISTAGG function in the following way

SELECT deptno, LISTAGG(ename, ',') WITHIN GROUP (ORDER BY ename) AS employees
FROM   emp
GROUP BY deptno;

Using XMLAGG function

SELECT deptno,
 RTRIM ( xmlagg (xmlelement (c, ename || ',') order by ename).extract ('//text()') , ',' ) AS EMPNAME
FROM   emp
GROUP BY deptno;

Using WM_CONCAT

SELECT deptno, wm_concat(ename) AS employees
FROM   emp  GROUP BY deptno;
Above three functions yield the same result.

For converting rows into columns we have Pivot function in oracle

Syntax is as follows

SELECT * FROM
(
  SELECT column1, column2
  FROM tables
  WHERE conditions
)
PIVOT
(
  aggregate_function(column2)
  FOR column2
  IN ( expr1, expr2, ... expr_n) | subquery
)
ORDER BY expression [ASC/DESC];

Consider the following data for example

order_id
customer_id
product_id
50001
SMITH
10
50002
SMITH
20
50003
ANDERSON
30
50004
ANDERSON
40
50005
JONES
10
50006
JONES
20
50007
SMITH
20
50008
SMITH
10
50009
SMITH
20

The expected output is

customer_ref
10
20
30
ANDERSON
0
0
1
JONES
1
1
0
SMITH
2
3

The query will be as follows


SELECT * FROM
(
  SELECT customer_ref, product_id
  FROM orders
)
PIVOT
(
  COUNT(product_id)
  FOR product_id IN (10, 20, 30)
)
ORDER BY customer_ref:
Hope this article helps you in understanding the above functions

Comments

Popular posts from this blog

Comparing Objects in Informatica

We might face a scenario where there may be difference between PRODUCTION v/s SIT version of code or any environment or between different folders in same environment. In here we go for comparison of objects we can compare between mappings,sessions,workflows In Designer it would be present under "Mappings" tab we can find "Compare" option. In workflow manger under "Tasks & Workfows" tab we can find "Compare" option for tasks and workflows comparison respectively. However the easiest and probably the best practice would be by doing using Repository Manager.In Repository Manager under "Edit" tab we can find "Compare" option. The advantage of using Repository manager it compares all the objects at one go i.e. workflow,session and mapping. Hence reducing the effort of individually checking the mapping and session separately. Once we select the folder and corresponding workflow we Can click compare for checking out ...

Target Load Type - Normal or Bulk in Session Properties

We can see the Target load type ( Normal or Bulk) property in session under Mapping tab and we will go for Bulk to improve the performance of session to load large amount of data. SQL loader utility will be used for Bulk load and it will not create any database logs(redolog and undolog), it directly writes to data file.Transaction can not be rolled back as we don't have database logs.However,Bulk loading is very as compared to Normal loading. In target if you are using Primary Key or Primary Index or any constraints you can't use Bulk mode. We can see this property in the below snap shot.

SCD Type 2 Implementation in Informatica using dynamic lookup

We have already covered how a dynamic look up works the same can be found out from the following link http://dwbuddy.blogspot.com/2014/06/dynamic-look-up-in-informatica.html  So let's see how we can use it in a regular SCD type 2 mapping. The logic used in SCD type 2 is same as discussed in earlier blog i.e.  http://dwbuddy.blogspot.com/2015/01/scd-type-ii-implementation-in.html . The only difference being here is we have used dynamic look up instead of a static one. We go for dynamic look up when there are duplicate entries for a PK value and only the latest record present in the file should be termed as active. Let us consider the following sample data So  for EMP_ID 101 we have two records in the same day flow so the latest record i.e. with salary as 1500 should be termed as Active record. The mapping will look in the following way So in here we have a dynamic lookup couple of expression transformations,a transaction control transformation which is use...