DataStage:DataStage Parallel Job VS Server Job

本人英文水平一般,先收着以后再翻译吧。
1) The basic difference between server and parallel jobs is the degree of parallelism
Server job stages do not have in built partitoning and parallelism mechanism for extracting and loading data between different stages. 

All you can do to enhance the speed and perormance in server jobs is to enable inter process row buffering through the administrator. This helps stages to exchange data as soon as it is available in the link. 
You could use IPC stage too which helps one passive stage read data from another as soon as data is available. In other words, stages do not have to wait for the entire set of records to be read first and then transferred to the next stage. Link partitioner and link collector stages can be used to achieve a certain degree of partitioning paralellism. 
All of the above features which have to be explored in server jobs are built in datastage Px. 

2) The Px engine runs on a multiprocessor system and takes full advantage of the processing nodes defined in the configuration file. Both SMP and MMP architecture is supported by datastage Px. 

3) Px takes advantage of both pipeline parallelism and partitoning paralellism. Pipeline parallelism means that as soon as data is available between stages( in pipes or links), it can be exchanged between them without waiting for the entire record set to be read. Partitioning parallelism means that entire record set is partitioned into small sets and processed on different nodes(logical processors). For example if there are 100 records, then if there are 4 logical nodes then each node would process 25 records each. This enhances the speed at which loading takes place to an amazing degree. Imagine situations where billions of records have to be loaded daily. This is where datastage PX comes as a boon for ETL process and surpasses all other ETL tools in the market.

4) In parallel we have Dataset which acts as the intermediate data storage in the linked list, it is the best storage option it stores the data in datastage internal format.

5) In parallel we can choose to display OSH , which gives information about the how job works.

6) In Parallel Transformer there is no reference link possibility, in server stage reference could be given to transformer. Parallel stage can use both basic and parallel oriented functions.

7) Datastage server executed by datastage server environment but parallel executed under control of datastage runtime environment

8) Datastage compiled in to BASIC(interpreted pseudo code) and Parallel compiled to OSH(Orchestrate Scripting Language).

9) Debugging and Testing Stages are available only in the Parallel Extender.

10) More Processing stages are not included in Server example, Join, CDC, Lookup etc…..

11) In File stages, Hash file available only in Server and Complex falat file , dataset , lookup file set avail in parallel only.

12) Server Transformer supports basic transforms only, but in parallel both basic and parallel transforms.

13) Server transformer is basic language compatability, pararllel transformer is c++ language compatabillity

14) Look up of sequntial file is possible in parallel jobs 

15) . In parallel we can specify more file paths to fetch data from using 
file pattern similar to Folder stage in Server, while in server we can 
specify one file name in one O/P link. 

16). We can simulteneously give input as well as output link to a seq. file 
stage in Server. But an output link in parallel means a reject link, that 
is a link that collects records that fail to load into the sequential file 
for some reasons.

17). The difference is file size Restriction. 
Sequential file size in server is : 2GB 
Sequential file size in parallel is : No Limitation..

18). Parallel sequential file has filter options too. Where you can specify the file pattern.

转载自:http://www.360doc.com/content/09/0223/10/7362_2620736.shtml
代码转载自:https://pan.quark.cn/s/7f503284aed9 Hibernate的核心组件总数达到五个,具体包括:Session、SessionFactory、Transaction、Query以及Configuration。 这五个核心组件在各类开发项目中都具有普遍的应用性。 借助这些组件,不仅可以高效地进行持久化对象的读取与存储,还能够实现事务管理功能。 接下来将通过图形化的方式,逐一阐述这五个核心组件的具体细节。 依据所提供的文件内容,可以总结出以下几个关键知识点:### 1. SSH框架详细架构图尽管标题提及“SSH框架详细架构图”,但在描述部分并未直接呈现关于SSH的详细内容,而是转向介绍了Hibernate的核心接口。 然而,在此我们可以简要概述SSH框架(涵盖Spring、Struts、Hibernate)的核心理念及其在Java开发中的具体作用。 #### Spring框架- **定义**:Spring框架是一个开源架构,其设计目标在于简化企业级应用的开发流程。 - **特点**: - **分层结构**:该框架允许开发者根据实际需求选择性地采纳部分组件,而非强制使用全部功能。 - **可复用性**:Spring框架支持创建可在不同开发环境中重复利用的业务逻辑和数据访问组件。 - **核心构成**: - **核心容器**:该部分包含了Spring框架的基础功能,其核心在于`BeanFactory`,该组件通过工厂模式运作,并借助控制反转(IoC)理念,将配置和依赖管理与具体的应用代码进行有效分离。 - **Spring上下文**:提供一个配置文件,其中整合了诸如JNDI、EJB、邮件服务、国际化支持等企业级服务。 - **Spring AO...
评论
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值