IN 1053 AdministratorGuide en
IN 1053 AdministratorGuide en
IN 1053 AdministratorGuide en
10.5.3
Administrator Guide
Informatica Administrator Guide
10.5.3
November 2022
© Copyright Informatica LLC 2005, 2022
This software and documentation are provided only under a separate license agreement containing restrictions on use and disclosure. No part of this document may be
reproduced or transmitted in any form, by any means (electronic, photocopying, recording or otherwise) without prior consent of Informatica LLC.
U.S. GOVERNMENT RIGHTS Programs, software, databases, and related documentation and technical data delivered to U.S. Government customers are "commercial
computer software" or "commercial technical data" pursuant to the applicable Federal Acquisition Regulation and agency-specific supplemental regulations. As such,
the use, duplication, disclosure, modification, and adaptation is subject to the restrictions and license terms set forth in the applicable Government contract, and, to the
extent applicable by the terms of the Government contract, the additional rights set forth in FAR 52.227-19, Commercial Computer Software License.
Informatica, the Informatica logo, PowerCenter, Power Exchange, and Informatica Cloud are trademarks or registered trademarks of Informatica LLC in the United
States and many jurisdictions throughout the world. A current list of Informatica trademarks is available on the web at https://www.informatica.com/trademarks.html.
Other company and product names may be trade names or trademarks of their respective owners.
Subject to your opt-out rights, the software will automatically transmit to Informatica in the USA information about the computing and network environment in which the
Software is deployed and the data usage and system statistics of the deployment. This transmission is deemed part of the Services under the Informatica privacy policy
and Informatica will use and otherwise process this information in accordance with the Informatica privacy policy available at https://www.informatica.com/in/
privacy-policy.html. You may disable usage collection in Administrator tool.
Portions of this software and/or documentation are subject to copyright held by third parties. Required third party notices are included with the product.
The information in this documentation is subject to change without notice. If you find any problems in this documentation, report them to us at
[email protected].
Informatica products are warranted according to the terms and conditions of the agreements under which they are provided. INFORMATICA PROVIDES THE
INFORMATION IN THIS DOCUMENT "AS IS" WITHOUT WARRANTY OF ANY KIND, EXPRESS OR IMPLIED, INCLUDING WITHOUT ANY WARRANTIES OF
MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND ANY WARRANTY OR CONDITION OF NON-INFRINGEMENT.
Table of Contents 3
Informatica Network Credentials. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 29
Enter Informatica Network Credentials. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 30
Searching Informatica Knowledge Base. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 30
4 Table of Contents
Audit Reports. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 60
Service States. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 61
Process States. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 61
Job States. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 63
Informatica Administrator Accessibility Overview. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 63
Keyboard Shortcuts. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 64
Table of Contents 5
Shutting Down a Domain. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 85
Domain Properties. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 86
General Properties. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 86
Database Properties. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 87
Gateway Configuration Properties. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 88
Service Level Management. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 88
SMTP Configuration. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 89
Custom Properties for the Domain. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 90
Chapter 6: Nodes. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 91
Nodes Overview. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 91
Node Types. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 92
Gateway Nodes. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 92
Worker Nodes. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 92
Example Domain with Multiple Nodes. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 92
Node Roles. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 93
Service Role. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 93
Compute Role. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 94
Updating the Node Role. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 94
Viewing Processes on a Node with the Service Role. . . . . . . . . . . . . . . . . . . . . . . . . . . . . 95
Define and Add Nodes. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 95
Adding Nodes to the Domain. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 95
Configuring Node Properties. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 96
Shutting Down and Restarting the Node. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 98
Shutting Down a Node from the Administrator Tool. . . . . . . . . . . . . . . . . . . . . . . . . . . . . 99
Starting or Stopping a Node on Windows. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 99
Starting or Stopping a Node on UNIX. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 99
Removing the Node Association. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 100
Removing a Node. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 100
6 Table of Contents
Application Service Failover Configuration. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 109
PowerCenter Integration Service Failover and Recovery Configuration. . . . . . . . . . . . . . . . 109
Command Line Program Resilience Configuration. . . . . . . . . . . . . . . . . . . . . . . . . . . . . 110
Domain Failover Configuration. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 110
Node Restart Configuration. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 111
Oracle RAC Database Failover. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 111
Troubleshooting High Availability. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 111
Table of Contents 7
Hadoop Cluster Properties. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 143
Common Properties. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 144
Reject Directory Properties. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 146
Blaze Configuration. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 146
Spark Configuration. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 147
HBase Connection Properties. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 148
HDFS Connection Properties. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 149
HBase Connection Properties for MapR-DB. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 151
Hive Connection Properties. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 151
HTTP Connection Properties. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 155
IBM DB2 Connection Properties. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 157
IBM DB2 for i5/OS Connection Properties. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 160
IBM DB2 for z/OS Connection Properties. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 163
IMS Connection Properties. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 166
JDBC Connection Properties. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 168
JDBC V2 Connection Properties. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 171
JD Edwards EnterpriseOne Connection Properties. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 173
Kafka Connection Properties. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 174
General Properties. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 175
Kafka Broker Properties. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 175
SSL Properties. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 176
Creating a Kafka Connection Using infacmd. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 177
Kudu Connection Properties . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 177
LDAP Connection Properties. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 178
Microsoft Azure Blob Storage Connection Properties. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 179
Microsoft Azure Cosmos DB SQL API Connection Properties. . . . . . . . . . . . . . . . . . . . . . . . . 180
Microsoft Azure Data Lake Storage Gen1 Connection Properties. . . . . . . . . . . . . . . . . . . . . . . 181
Microsoft Azure Data Lake Storage Gen2 Connection Properties. . . . . . . . . . . . . . . . . . . . . . . 182
Microsoft Azure SQL Data Warehouse Connection Properties. . . . . . . . . . . . . . . . . . . . . . . . . 183
MS SQL Server Connection Properties. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 185
Netezza Connection Properties. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 189
OData Connection Properties. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 190
ODBC Connection Properties. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 191
Oracle Connection Properties. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 192
Salesforce Connection Properties. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 195
Salesforce Marketing Cloud Connection Properties. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 196
SAP Connection Properties. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 197
Sequential Connection Properties. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 200
Snowflake Connection Properties. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 202
Teradata Parallel Transporter Connection Properties. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 203
Tableau Connection Properties. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 205
Tableau V3 Connection Properties. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 206
8 Table of Contents
Twitter Streaming Connection Properties. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 207
VSAM Connection Properties. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 208
Web Services Connection Properties. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 211
Identifier Properties in Database Connections. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 212
Regular Identifiers. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 213
Delimited Identifiers. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 213
Identifier Properties. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 213
Table of Contents 9
Supported Platforms. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 232
Repositories. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 232
Service Options. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 233
Connections. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 233
Metadata Exchange Options. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 233
10 Table of Contents
Monitor Web Services. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 253
Properties View for a Web Service. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 253
Reports View for a Web Service. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 253
Operations View for a REST or SOAP Web Service. . . . . . . . . . . . . . . . . . . . . . . . . . . . . 253
Requests View for a Web Service. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 254
Monitor Workflows. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 254
Workflow Graph . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 254
View Workflow Objects. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 255
Workflow States. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 256
Workflow Object States. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 257
Mapping Task Work Item States. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 258
Canceling or Aborting a Workflow. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 259
Workflow Recovery. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 259
Workflow Logs. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 260
Job Status After Application Service Restart or Failover. . . . . . . . . . . . . . . . . . . . . . . . . . . . 262
Monitoring a Folder of Objects. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 262
Viewing the Context of an Object. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 263
Configuring the Date and Time Custom Filter. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 263
Configuring the Elapsed Time Custom Filter. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 263
Configuring the Multi-Select Custom Filter. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 263
Table of Contents 11
Analyst Service Log Events. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 277
Data Integration Service Log Events. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 277
Listener Service Log Events. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 278
Logger Service Log Events. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 278
Model Repository Service Log Events. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 278
Metadata Manager Service Log Events. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 278
PowerCenter Integration Service Log Events. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 279
PowerCenter Repository Service Log Events. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 279
Resource Manager Service Log Events. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 279
SAP BW Service Log Events. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 279
Scheduler Service Log Events. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 280
Web Services Hub Log Events. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 280
User Activity Log Events. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 280
Log Aggregator. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 281
Aggregating Application Service Logs. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 282
Processing Aggregated Application Service Logs. . . . . . . . . . . . . . . . . . . . . . . . . . . . . 282
Mapping Task Logs. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 283
12 Table of Contents
Chapter 16: Node Diagnostics. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 297
Node Diagnostics Overview. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 297
Informatica Network Login. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 298
Logging In to the Informatica Network. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 298
Generating Node Diagnostics. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 299
Downloading Node Diagnostics. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 300
Uploading Node Diagnostics. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 300
Analyzing Node Diagnostics. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 301
Identify Bug Fixes. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 301
Identify Recommendations. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 302
Table of Contents 13
PowerCenter Code Page Conversion. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 320
Choosing Characters for PowerCenter Repository Metadata. . . . . . . . . . . . . . . . . . . . . . 320
Case Study: Processing ISO 8859-1 Data. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 321
Configuring the ISO 8859-1 Environment. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 321
Case Study: Processing Unicode UTF-8 Data. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 323
Configuring the UTF-8 Environment. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 323
Index. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 361
14 Table of Contents
Preface
Use the Informatica® Administrator Guide to learn how to log into the Administrator tool and understand the
user interface. Read on how to configure, manage, and monitor the Informatica domain. Learn about domain
architecture and its components, including nodes, services, high availability, connections, and monitoring.
Informatica Resources
Informatica provides you with a range of product resources through the Informatica Network and other online
portals. Use the resources to get the most from your Informatica products and solutions and to learn from
other Informatica users and subject matter experts.
Informatica Network
The Informatica Network is the gateway to many resources, including the Informatica Knowledge Base and
Informatica Global Customer Support. To enter the Informatica Network, visit
https://network.informatica.com.
To search the Knowledge Base, visit https://search.informatica.com. If you have questions, comments, or
ideas about the Knowledge Base, contact the Informatica Knowledge Base team at
[email protected].
Informatica Documentation
Use the Informatica Documentation Portal to explore an extensive library of documentation for current and
recent product releases. To explore the Documentation Portal, visit https://docs.informatica.com.
If you have questions, comments, or ideas about the product documentation, contact the Informatica
Documentation team at [email protected].
15
Informatica Product Availability Matrices
Product Availability Matrices (PAMs) indicate the versions of the operating systems, databases, and types of
data sources and targets that a product release supports. You can browse the Informatica PAMs at
https://network.informatica.com/community/informatica-network/product-availability-matrices.
Informatica Velocity
Informatica Velocity is a collection of tips and best practices developed by Informatica Professional Services
and based on real-world experiences from hundreds of data management projects. Informatica Velocity
represents the collective knowledge of Informatica consultants who work with organizations around the
world to plan, develop, deploy, and maintain successful data management solutions.
You can find Informatica Velocity resources at http://velocity.informatica.com. If you have questions,
comments, or ideas about Informatica Velocity, contact Informatica Professional Services at
[email protected].
Informatica Marketplace
The Informatica Marketplace is a forum where you can find solutions that extend and enhance your
Informatica implementations. Leverage any of the hundreds of solutions from Informatica developers and
partners on the Marketplace to improve your productivity and speed up time to implementation on your
projects. You can find the Informatica Marketplace at https://marketplace.informatica.com.
To find your local Informatica Global Customer Support telephone number, visit the Informatica website at
the following link:
https://www.informatica.com/services-and-training/customer-success-services/contact-us.html.
To find online support resources on the Informatica Network, visit https://network.informatica.com and
select the eSupport option.
16 Preface
Chapter 1
Understanding Domains
This chapter includes the following topics:
The Informatica domain is the fundamental administrative unit in Informatica. The domain supports the
administration of the distributed services. A domain is a collection of nodes and services that you can group
in folders based on administration ownership.
A node is the logical representation of a machine in a domain. One node in the domain acts as a gateway to
receive service requests from clients and route them to the appropriate service and node. Services and
processes run on nodes in a domain. The availability of a service or process on a node depends on how you
configure the service and the node.
Services for the domain include the Service Manager and a set of application services:
• Service Manager. A service that runs on each node in the domain to manage all domain operations. The
Service Manager performs domain functions such as authentication, authorization, and logging. The
Service Manager also starts the application services configured to run on the node.
• Application Services. Services that represent server-based functionality, such as the Model Repository
Service and the Data Integration Service. The application services that run on a node depend on the way
you configure the services.
The Service Manager and application services control security. The Service Manager manages users and
groups that can log in to application clients and authenticates the users who log in to the application clients.
The Service Manager and application services authorize user requests from application clients.
Informatica Administrator (the Administrator tool), consolidates the administrative tasks for domain objects
such as services, nodes, licenses, and grids. You manage the domain and the security of the domain through
the Administrator tool.
17
If you have the high availability option, you can scale services and eliminate single points of failure for
services. Services can continue running despite temporary network or hardware failures.
Nodes
A node is a logical representation of a machine in a domain. During installation, you add the installation
machine to the domain as a node. You can add multiple nodes to a domain.
Each node in the domain runs the Service Manager that manages domain functions on that node. The Service
Manager also supports the application services that run on the node. The domain functions that the node
performs and the services that the node runs depend on the following node configurations:
Node type
The node type determines whether the node can serve as a gateway or worker node and determines the
domain functions that the node performs. One gateway node serves as the master gateway node for the
domain. The master gateway node receives service requests from clients and routes them to the
appropriate service and node. A worker node is any node not configured to serve as a gateway. The first
time that you install the Informatica services, you create a gateway node and the Informatica domain.
When you install the Informatica services on other machines, you create additional gateway nodes or
worker nodes that you join to the domain.
Node role
The node role defines the purpose of the node. A node with the service role can run application services.
A node with the compute role can perform computations requested by remote application services. A
node with both roles can run application services and locally perform computations for those services.
By default, each gateway and worker node has both the service and compute roles enabled. If a node is
assigned to a Data Integration Service grid, you might want to update the node role. Enable only the
service role to dedicate the node to running the Data Integration Service process. Enable only the
compute role to dedicate the node to running Data Integration Service mappings.
You can subscribe to alerts to receive notification about node events such as node failure or a master
gateway election. You can also generate and upload node diagnostics to the Configuration Support Manager
and review information such as available EBFs and Informatica recommendations.
Service Manager
The Service Manager is a service that manages all domain operations. It runs within Informatica services. It
runs as a service on Windows and as a daemon on UNIX. When you start Informatica services, you start the
Service Manager.
The Service Manager runs on each node in the domain. If the Service Manager is not running, the node is not
available.
• Domain support. The Service Manager performs functions on each node to support the domain. The
functions that the Service Manager performs on a node depend on the type and role of the node. For
example, the Service Manager running on the master gateway node performs all domain functions on that
node. The Service Manager running on any other type of node performs limited domain functions on that
node.
• Application service support. When a node has the service role, the Service Manager starts application
services configured to run on that node. It starts and stops services and service processes based on
requests from clients. It also directs service requests to application services. The Service Manager uses
TCP/IP to communicate with the application services.
The following table describes the domain functions that the Service Manager performs:
Function Description
Alert The Service Manager sends alerts to subscribed users. You subscribe to alerts to receive
Management notification for node failure and master gateway election on the domain, and for service process
failover for services on the domain. When you subscribe to alerts, you receive notification emails.
Alert management occurs on the master gateway node.
Authentication The Service Manager authenticates users who log in to application clients.
Authentication occurs on the master gateway node.
Authorization The Service Manager authorizes user requests for domain objects based on the privileges, roles,
and permissions assigned to the user. Requests for domain objects can come from the
Administrator tool.
Domain authorization occurs on the master gateway node.
Container When a node has the compute role, the Service Manager manages the containers on the node. A
Management container is an allocation of memory and CPU resources. An application service uses the
container to remotely perform computations on the node. For example, a Data Integration Service
that runs on a grid can remotely run a mapping within a container on a node with the compute
role.
Container management occurs on any node with the compute role.
Node The Service Manager manages node configuration metadata in the domain.
Configuration Node configuration occurs on all nodes in the domain.
Licensing The Service Manager registers license information and verifies license information when you run
application services.
Licensing occurs on the master gateway node.
Logging The Service Manager provides accumulated log events from each service in the domain. To
perform the logging function, the Service Manager runs a Log Manager and a Log Agent.
The Log Manager runs on the master gateway node. The Log Agent runs on all nodes where
PowerCenter® Integration Service sessions and workflows run and where Data Integration
Service jobs run.
Service Manager 19
Function Description
User The Service Manager manages the native and LDAP users and groups that can log in to
Management application clients. It also manages the creation of roles and the assignment of roles and
privileges to native and LDAP users and groups.
User management occurs on the master gateway node.
Monitoring The Service Manager persists, updates, retrieves, and publishes run-time statistics for integration
objects in the monitoring Model repository. The Service Manager stores the monitoring
configuration details in the domain configuration repository.
Monitoring occurs on the master gateway node.
The following image shows where the Service Manager performs domain functions:
Application Services
Application services represent server-based functionality.
Application services include services that you create and system services that are created for you when you
create the domain. A system service can have a single instance in the domain.
• Analyst Service
• Content Management Service
• Data Integration Service
• Metadata Access Service
• Metadata Manager Service
• Model Repository Service
• PowerCenter Integration Service
• PowerCenter Repository Service
When you configure an application service, you designate a node to run the service process. When a service
process runs, the Service Manager assigns a port number from the range of port numbers assigned to the
node.
The service process is the runtime representation of a service running on a node. The service type
determines how many service processes can run at a time. For example, the PowerCenter Integration Service
can run multiple service processes at a time when you run it on a grid.
If you have the high availability option, you can run a service on multiple nodes. Designate the primary node
to run the service. All other nodes are back-up nodes for the service. If the primary node is not available, the
service runs on a back-up node. You can subscribe to alerts to receive notification in the event of a service
process failover.
If you do not have the high availability option, configure a service to run on one node. If you assign the
service to multiple nodes, then the service will not start.
Analyst Service
The Analyst Service is an application service that runs the Informatica Analyst application in the Informatica
domain. The Analyst Service manages the connections between service components and the users that log in
to the Analyst tool. The Analyst Service connects to a Data Integration Service, a Model Repository Service, a
Metadata Manager Service, and a Search Service.
The Analyst Service also specifies the following directories and database connections:
Connect the Analyst Service to a Metadata Manager Service to perform data lineage operations on
scorecards in the Analyst tool. Connect the Analyst Service to a Search Service to manage search operations
in the Analyst tool.
Specify a flat file cache directory to store temporary data from flat files that you upload. Specify a business
glossary directory to stores temporary files that you export from the Business Glossary.
Specify a flat file network directory so that you can import flat files using the network path option in the
Analyst tool.
Specify a database as the exception management audit database to write an audit trail for all Human task
operations to a single database. When you specify the database, specify the schema for the audit tables also.
The database stores audit data for all work that users perform on Human task instances in the Analyst tool
Application Services 21
that the Analyst Service manages. If you do not specify a database and schema, the Analyst Service writes
audit data for each Human task instance to the database that stores the Human task data.
The Content Management Service must be available when you use the following resources:
The Content Management Service manages configuration information for address reference data. The
Data Integration Service applies the configuration information when it runs a mapping that reads the
address reference data.
The Content Management Service manages the list of the population files on the node. The Data
Integration Service applies the population configuration when it runs a mapping that reads the
population files.
The Content Management Service stores the locations of any probabilistic model file and classifier
model file on the node. The Content Management Service also manages the compilation status of each
model.
Reference tables
The Content Management Service identifies the database that stores data values for the reference table
objects in the associated Model repository.
Rule specifications
The Content Management Service generates mapplets from rule specifications. The Analyst Service
selects a Content Management Service to generate the mapplets. The Analyst tool uses the Model
Repository Service configuration to select the Content Management Service.
When you start a command from the command line or an external client to run SQL data services and
mappings in an application, the command sends the request to the Data Integration Service.
You can configure the Data Integration Service to run on the following domain objects:
On nodes
If your license includes high availability, you can configure the service to run on multiple nodes. By
default, the service runs on the primary node. If the primary node is not available, the service runs on a
back-up node. If the service process fails or the node becomes unavailable, the service fails over to
another node. If your license does not include high availability, you can configure the service to run on
one node.
If your license includes grid, you can configure the service to run on a grid. A grid is an alias assigned to
a group of nodes. The Data Integration Service dispatches jobs to available nodes assigned to the grid.
When the Data Integration Service runs on a grid, the service remains available if a service process fails
or a node becomes unavailable.
Email Service
The Email Service is a system service that manages email notifications for business glossaries, scorecards,
and workflows.
You can configure the service to run on multiple nodes. Designate the primary node to run the service. All
other nodes are backup nodes for the service. If the primary node is not available, the service runs on a
backup node.
The Metadata Access Service contains information about the Service Principal Name (SPN) and keytab
information if the Hadoop cluster uses Kerberos authentication. You can create one or more Metadata
Access Services on a node. Based on your license, the Metadata Access Service can be highly available.
HBase, HDFS, Hive, and MapR-DB connections use the Metadata Access Service when you import an object
from a Hadoop cluster. Google Cloud Storage connection uses Metadata Access Service to import metadata
from files in Google Cloud Storage. Create and configure a Metadata Access Service before you create
Google Cloud Storage, HBase, HDFS, Hive, and MapR-DB connections.
Use Metadata Manager to browse and analyze metadata from disparate source repositories. You can load,
browse, and analyze metadata from application, business intelligence, data integration, data modelling, and
relational metadata sources.
You can configure the Metadata Manager Service to run on only one node. The Metadata Manager Service is
not a highly available service. However, you can run multiple Metadata Manager Services on the same node.
Application Services 23
Developer, Informatica Analyst, Data Integration Service, and the Administrator tool store metadata in the
Model repository.
You can configure a Model repository as a monitoring Model repository. You can then set up a monitoring
Model Repository Service for the monitoring Model repository. The monitoring Model Repository Service
monitors statistics for Data Integration Service jobs. You configure the monitoring Model Repository Service
in the domain properties.
Create one Model Repository Service for each Model repository. When you create a Model Repository Service,
you can create a Model repository or use an existing Model repository. You can run multiple Model
Repository Services on the same node.
• On a grid. When you configure the service to run on a grid, it can run on multiple nodes at a time. The
PowerCenter Integration Service dispatches tasks to available nodes assigned to the grid. If you do not
have the high availability option, the task fails if any service process or node becomes unavailable. If you
have the high availability option, failover and recovery is available if a service process or node becomes
unavailable.
• On nodes. If you have the high availability option, you can configure the service to run on multiple nodes.
By default, it runs on the primary node. If the primary node is not available, it runs on a backup node. If the
service process fails or the node becomes unavailable, the service fails over to another node. If you do not
have the high availability option, you can configure the service to run on one node.
If you have the high availability option, you can configure the service to run on primary and backup nodes. By
default, the service process runs on the primary node. If the service process fails, a new process starts on the
same node. If the node becomes unavailable, a service process starts on one of the backup nodes.
If you have the PowerCenter high availability option, you can run the Listener Service on multiple nodes. If the
Listener Service process fails on the primary node, it fails over to a backup node.
The Resource Manager Service communicates with compute nodes in a Data Integration Service grid. Enable
the Resource Manager Service when you configure a Data Integration Service grid to run jobs in separate
remote processes.
You can configure the service to run on multiple nodes. Designate the primary node to run the service. All
other nodes are back-up nodes for the service. If the primary node is not available, the service runs on a back-
up node.
SAP BW Service
The SAP BW Service listens for RFC requests from SAP NetWeaver BI and initiates workflows to extract from
or load to SAP NetWeaver BI. The SAP BW Service is not highly available. You can configure it to run on one
node.
Scheduler Service
The Scheduler Service is a system service that manages schedules for profiles, scorecards, deployed
mappings, and deployed workflows.
You can configure the service to run on multiple nodes. Designate the primary node to run the service. All
other nodes are back-up nodes for the service. If the primary node is not available, the service runs on a back-
up node.
High Availability
High availability is an option that eliminates a single point of failure in a domain and provides minimal service
interruption in the event of failure. High availability consists of the following components:
• Resilience. The ability of application services to tolerate transient network failures until either the
resilience timeout expires or the external system failure is fixed.
• Failover. The migration of an application service or task to another node when the node running the
service process becomes unavailable.
High Availability 25
• Recovery. The automatic completion of tasks after a service is interrupted. Automatic recovery is
available for PowerCenter Integration Service and PowerCenter Repository Service tasks. You can also
manually recover PowerCenter Integration Service workflows and sessions. Manual recovery is not part of
high availability.
Informatica DiscoveryIQ uploads data to Informatica 15 minutes after you install and configure the
Informatica domain. Thereafter, the domain sends data to Informatica every few days. Data collection and
upload is enabled by default. You can choose to not send any usage statistics to Informatica.
If the network where you install Informatica services needs a proxy server to communicate with the external
network, configure the proxy details.
Informatica DiscoveryIQ enables Informatica to provide an environment health check after the analysis of
system statistics and domain reports. You can receive best practices and recommendations from
Informatica based on the reports to resolve issues on the domain. The usage statistics provide Informatica a
proactive insight into product implementation.
1. In the Administrator tool header area, click Manage > DiscoveryIQ Proxy Details.
2. Enter the domain, host name, and port number of the proxy server.
3. Enter the user name and password to connect to the proxy server.
4. Click OK to save the proxy server details.
If you have a native user account, you can change your password at any time with the Change Password
application. If someone else created your user account, change your password the first time you log in to the
Administrator tool.
User preferences control the options that appear in the Administrator tool when you log in. User preferences
do not affect the options that appear when another user logs in to the Administrator tool.
You can configure Informatica Network credentials for your account so that you can access the Informatica
Knowledge Base from the Administrator tool.
Password Management
You can change the password through the Change Password application.
You can open the Change Password application from the Administrator tool or with the following URL:
http://<fully qualified host name>:<port>/passwordchange/
The Service Manager uses the user password associated with a worker node to authenticate the domain
user. If you change a user password that is associated with one or more worker nodes, the Service Manager
updates the password for each worker node. The Service Manager cannot update nodes that are not running.
For nodes that are not running, the Service Manager updates the password when the nodes restart.
Note: For an LDAP user account, change the password in the LDAP directory service.
28
For a user account in the domain that uses native authentication, if you enable password complexity, use the
following guidelines when you create or change a password:
1. In the Administrator tool header area, click Manage > Change Password.
The Change Password application opens in a new browser window.
2. Enter the current password in the Password box, and the new password in the New Password and
Confirm Password boxes.
3. Click Update.
Preferences
Your preferences determine the options that appear in the Administrator tool when you log in. Your
preferences do not affect the options that appear when another user logs in to the Administrator tool.
The following table describes the options that you can configure for your preferences:
Option Description
Subscribe for Subscribes you to domain and service alerts. You must have a valid email address configured for
Alerts your user account. Default is No.
Show Custom Displays custom properties in the contents panel when you click an object in the Navigator. You
Properties use custom properties to configure Informatica behavior for special cases or to increase
performance. Hide the custom properties to avoid inadvertently changing the values. Use custom
properties only if Informatica Global Customer Support instructs you to.
To edit your preferences, click Manage > Preferences in the Administrator tool header area.
You can also view the search results for an error message in the Informatica Knowledge Base by clicking the
error message code in the Administrator tool.
Preferences 29
Enter Informatica Network Credentials
Enter your Informatica Network credentials to access the Informatica Knowledge Base from the
Administrator tool.
• Domain administrative tasks. Manage logs, domain objects, user permissions, and domain reports.
Generate and upload node diagnostics. Monitor Data Integration Service jobs and applications. Domain
objects include application services, nodes, grids, folders, database connections, operating system
profiles, and licenses.
• Security administrative tasks. Manage users, groups, roles, and privileges.
31
The Administrator tool has the following tabs:
• Manage. View and edit the properties of the domain and objects within the domain.
• Monitor. View the status of profile jobs, scorecard jobs, preview jobs, mapping jobs, SQL data services,
web services, and workflows for each Data Integration Service.
• Logs. View log events for the domain and services within the domain.
• Reports. Run a Web Services Report or License Management Report.
• Security. Manage users, groups, roles, and privileges.
• Cloud. View information about your Informatica Cloud® organization.
If the Informatica domain runs on a network with Kerberos authentication, you must configure the browser to
allow access to the Informatica web applications. In Microsoft Internet Explorer, Microsoft Edge, and Google
Chrome, add the URL of the Informatica web application to the list of trusted sites. In Safari, add the
certificate of the Informatica web application to the keychain. If you are using Chrome version 86.0.42x or
later on Windows, you must also set the AuthServerWhitelist and AuthNegotiateDelegateWhitelist
policies.
You configure the Administrator tool port when you define the domain. You can define the domain during
installation or by running the infasetup DefineDomain command line program. If you enter the domain port
instead of the Administrator tool port in the URL, the browser is directed to the Administrator tool port.
Note: If the domain fails over to a different master gateway node, the host name in the Administrator tool
URL is equal to the host name of the elected master gateway node.
I cannot log in to the Administrator tool from the same machine where I created the domain gateway node.
After installation, if you cannot log in to the Administrator tool from the same machine where you
created the domain gateway node, clear the browser cache. When you initially log in to the Administrator
tool after installation, you can only log in with the Administrator user account created during installation.
If a different user credential is stored in the browser cache, the login can fail.
If a blank page appears after you log in to the Administrator tool, verify that you enabled delegation for
all user accounts with service principals used in the Informatica domain. To enable delegation, in the
Microsoft Active Directory Service, set the Trust this user for delegation to any service (Kerberos only)
option for each user account that you set an SPN.
Manage Tab
On the Manage tab, you can view and manage the domain and the objects that it contains.
The contents that appear and the tasks you can complete on the Manage tab vary based on the view that you
select. You can select the following views:
• Domain. View and manage domain status, resource consumption, and events.
• Services and Nodes. View and manage application services and nodes.
• Connections. View and manage connections.
• Schedules. View and manage schedules for profiles, scorecards, deployed mappings, and deployed
workflows.
Manage Tab 33
Manage Tab - Domain View
The Domain view displays an overview of the domain and its contents. You can use the domain view to
monitor the domain status, resource consumption, and events. You can also perform domain actions, such
as enabling and disabling services.
The following image shows the Domain view on the Manage tab:
Use the Domain Actions menu to view more information about the domain or shut down the domain.
• View Properties. Open the Services and Nodes view and display the properties for the domain.
• View History. Open the History view and display domain events from the last day.
• View Logs. Open the Logs tab and display Service Manager log events from the last day.
• View Command History. Open the Command History panel and display the 50 most recent service
lifecycle commands that were issued from the Administrator tool.
• Shut Down Domain.
Contents panel
Displays domain objects and their states. Domain objects include services, nodes, and grids.
Method Description
Search Search for an object by name. You can use an asterisk (*) as a wildcard character in this field.
Objects in the contents panel have Actions menus. Use the Actions menus to view information about
domain objects or perform common tasks. The information you can view and the tasks that you can
perform vary depending on which object you select.
• View Properties. Open the Services and Nodes view and display the properties for the service.
• View History. Open the History view and display service events from the last day.
• View Logs. Open the Logs tab and display service log events from the last day.
• View Dependencies. Open the Dependency graph and display direct dependencies for the service.
• Recycle Service.
• Enable or Disable Service.
• View Properties. Open the Services and Nodes view and display properties for the node.
• View History. Open the History view and display node events from the last day.
• View Dependencies. Open the Dependency graph and display direct dependencies for the node.
• Shut Down Node
• View Properties. Open the Services and Nodes view and display properties for the grid.
• View Dependencies. Open the Dependency graph and display direct dependencies for the grid.
Doughnut chart that shows the number and state of the services in the domain. When you click a state in
the chart, the contents panel displays services with that state.
Bar charts that compare resource usage of a domain process to resource usage of all processes on the
machine. The Domain view contains a memory usage indicator and a CPU usage indicator.
The following table describes the details that display depending on the object that you select in the contents
panel:
Node Node name and state. Click the node name to view the node properties.
Service The Details panel displays the following content for a service:
- Service name and state. Click the service name to view the service properties.
- Node on which the service process runs. Click the node name to view the node properties.
- State of the node on which the service process runs.
- State of the service process.
Service running The Details panel displays the following content for a service that runs on a grid:
on a grid - Service name and state. Click the service name to view the service properties.
- Nodes in the grid. Click a node name to view the node properties.
- State of the nodes on which the service processes run.
- State of the service processes.
Service in high The Details panel displays the following content for a service that is highly available:
availability mode - Service name and state. Click the service name to view the service properties.
- Nodes on which the service is configured to run. Click a node name to view the node properties.
- State of the nodes on which the service processes run.
- State of the service process on the nodes.
Grid The Details panel displays the following content for a grid:
- Grid name and state. Click the grid name to view the grid properties.
- Nodes in the grid. Click a node name to view the node properties.
- State of nodes running in the grid.
You can view usage statistics for memory and CPU. Choose whether to view current statistics or to view
graphs of usage for the past 60 minutes. Click the selection arrow and choose Current or Last Hour Trend.
The information that the graphs display varies based on which domain object you select.
The following table describes the information that you can view when you select the domain or a domain
object:
Service in high availability mode The usage indicators display the following content:
- The node on which the service process is running.
- Resource usage of the service process that is running on the node.
- Resource usage of all processes running on the machine.
Service running on a grid The usage indicators display the following content:
- All nodes on which the service process runs.
- Resource usage of the service process that is running on each node.
- Resource usage of all processes running on the machine.
If a See More… link appears in the indicator, you can click it to view the complete list of nodes in the domain.
You can sort the list by node name, process usage on the nodes, or process usage on the machine. You can
also search the list for a particular node.
Navigator
Appears in the left pane of the Manage tab. The Navigator displays the following types of objects:
• Domain. You can view one domain, which is the highest object in the Navigator hierarchy.
• Folders. Use folders to organize domain objects in the Navigator. Select a folder to view information
about the folder and the objects in the folder.
• Application services. An application service represents server-based functionality. Select an
application service to view information about the service and its processes.
• System services. A system service is an application service that can have a single instance in the
domain. Select a system service to view information about the service and its processes.
• Nodes. A node represents a machine in the domain. You configure service processes to run on nodes
with the service role.
• Grids. Create a grid to run the Data Integration Service or PowerCenter Integration Service on multiple
nodes. Select a grid to view nodes assigned to the grid.
• Licenses. Create a license on the Manage tab based on a license key file provided by Informatica.
Select a license to view services assigned to the license.
You can search for nodes, services, and grids in the Navigator.
Contents panel
Appears in the right pane of the Manage tab and displays information about the domain or domain
object that you select in the Navigator.
When you select the domain in the Navigator, you can create a folder, service, node, grid, or license.
When you select the domain in the Navigator, you can shut down the domain or view logs for the domain.
When you select a node in the Navigator, you can remove a node association, recalculate the CPU profile
benchmark, or shut down the node.
When you select a service in the Navigator, you can recycle or disable the service and configure
properties for the service.
When you select a license in the Navigator, you can add an incremental key to the license.
Navigator Search
You can search for and filter nodes, application services, and grids in the Navigator.
You can perform the following tasks in the Navigator search section:
In the search text box, enter the name or partial name of the object to search for. The Navigator displays
the search results.
Click Filters, and then select the object type to filter by. If you filter by nodes, you can filter further by
node role. If you filter by services, you can filter further by service type. The Navigator displays the
filtered results.
The following image displays the Navigator search section filtered by nodes with the service and
compute roles:
Reset filters.
Domain
You can view one domain in the Services and Nodes view on the Manage tab. It is the highest object in the
Navigator hierarchy.
When you select the domain in the Navigator, the contents panel shows the following views and buttons:
In the Actions menu on the Manage tab, you can shut down the domain, view logs, or access help for the
current view.
Folders
You can use folders in the domain to organize objects and to manage security.
Folders can contain nodes, services, grids, licenses, and other folders.
When you select a folder in the Navigator, the Navigator opens to display the objects in the folder. The
contents panel displays the following information:
Note: The System_Services folder is created for you when you create the domain, and contains all of the
system services. A system service is an application service that can have a single instance in the domain.
You cannot delete, move, or edit the properties or contents of the System_Services folder.
Application Services
Application services are a group of services that represent Informatica server-based functionality.
In the Services and Nodes view on the Manage tab, you can create and manage the following application
services:
Analyst Service
Runs Informatica Analyst in the Informatica domain. The Analyst Service manages the connections
between service components and the users that log in to Informatica Analyst.
The Analyst Service connects to a Data Integration Service, a Model Repository Service, a Metadata
Manager Service, and a Search Service. The Analyst Service also specifies a flat file cache directory and
flat file network directory, a directory for business glossary export files, and an exception management
audit database.
You can create and recycle the Analyst Service in the Informatica domain to access the Analyst tool. You
can launch the Analyst tool from the Administrator tool.
When you select an Analyst Service in the Navigator, the contents panel displays the following
information:
• Properties view. View the state of the service and the URL for the Analyst Service. Manage general,
model repository, data integration, human task, metadata manager, flat file cache, flat file network,
business glossary export, logging, and custom properties.
• Processes view. View the state of the service process on each node. View and edit service process
properties on each assigned node.
When you select a Content Management Service in the Navigator, the contents panel displays the
following information:
• Properties view. View the state of the service. Manage general, master, data integration, model
repository, reference table data, temporary file, logging, and custom properties.
• Processes view. View the state of the service process on each node. View and edit service process
properties on each assigned node.
• Permissions view. View or edit the group and user permissions on the Content Management Service.
• Actions menu. Manage the service.
Completes data integration tasks for Informatica Analyst, Informatica Developer, and external clients.
When you preview or run data profiles, SQL data services, and mappings in Informatica Analyst or
Informatica Developer, the application sends requests to the Data Integration Service to perform the
data integration tasks. When you start a command from the command line or an external client to run
SQL data services and mappings in an application, the command sends the request to the Data
Integration Service.
When you select a Data Integration Service in the Navigator, the contents panel displays the following
information:
• Properties view. View the state of the service. Manage general, model repository, logging, logical
data object and virtual table cache, profiling, data object cache, and custom properties. Set the
default deployment option.
• Processes view. View the state of the service process on each node. View and edit service process
properties on each assigned node.
• Applications view. Start and stop applications and SQL data services. Back up applications. Manage
application properties.
• Actions menu. Manage the service and repository contents.
Metadata Access Service
Allows the Developer tool to access Hadoop connection information to import and preview metadata.
When importing a data object, the Developer tool sends a request to the Metadata Access Service.
When you select a Metadata Access Service in the Navigator, the contents panel displays the following
information:
• Properties view. View the state of the service. Manage general, execution, HTTP configuration and
custom properties.
• Processes view. View the state of the service process on each node. View and edit service process
properties on each assigned node.
• Permissions view. View or edit the group and user permissions on the Metadata Access Service.
• Actions menu. Manage the service.
Runs the Metadata Manager application and manages connections between the Metadata Manager
components.
• Properties view. View the state of the service and the URL of the Metadata Manager Service instance.
View or edit Metadata Manager properties.
• Associated Services view. View and configure the Integration Service associated with the Metadata
Manager Service.
• Permissions view. View or edit the group and user permissions on the Metadata Manager Service.
• Actions menu. Manage the service and repository contents.
Manages the Model repository. The Model repository stores metadata created by Informatica products,
such as Informatica Developer, Informatica Analyst, the Data Integration Service, and Informatica
Administrator. The Model repository enables collaboration among the products.
When you select a Model Repository Service in the Navigator, the contents panel displays the following
information:
• Properties view. View the state of the service. Manage general, repository database, search, and
custom properties.
• Processes view. View the state of the service process on each node. View and edit service process
properties on each assigned node.
• Actions menu. Manage the service and repository contents.
You can configure a Model repository as a monitoring Model repository. You can configure a monitoring
Model Repository Service at the domain level to monitor multiple Data Integration Service and the
objects that the Data Integration Service runs.
Runs PowerCenter sessions and workflows. Select a PowerCenter Integration Service in the Navigator to
access information about the service.
When you select a PowerCenter Integration Service in the Navigator, the contents panel displays the
following information:
• Properties view. View the state of the service. View or edit Integration Service properties.
• Associated Repository view. View or edit the repository associated with the Integration Service.
• Processes view. View the state of the service process on each node. View or edit the service process
properties on each assigned node.
• Permissions view. View or edit group and user permissions on the Integration Service.
• Actions menu. Manage the service.
Manages the PowerCenter repository. It retrieves, inserts, and updates metadata in the repository
database tables. Select a PowerCenter Repository Service in the Navigator to access information about
the service.
When you select a PowerCenter Repository Service in the Navigator, the contents panel displays the
following information:
• Properties view. View the state and operating mode of the service. Manage general and advanced
properties, node assignments, and database properties.
When you select a Listener Service in the Navigator, the contents panel displays the following
information:
• Properties view. View the state of the service and the URL of the PowerExchange Listener instance.
View or edit Listener Service properties.
• Actions menu. Contains actions that you can perform on the Listener Service, such as viewing logs or
enabling and disabling the service.
When you select a Logger Service in the Navigator, the contents panel displays the following
information:
• Properties view. View the state of the service and the URL of the PowerExchange Logger instance.
View or edit Logger Service properties.
• Actions menu. Contains actions that you can perform on the Logger Service, such as viewing logs or
enabling and disabling the service.
SAP BW Service
Listens for RFC requests from SAP BW and initiates workflows to extract from or load to SAP BW. Select
an SAP BW Service in the Navigator to access properties and other information about the service.
When you select an SAP BW Service in the Navigator, the contents panel displays the following
information:
• Properties view. View the state of the service. Manage general properties and node assignments.
• Associated Integration Service view. View or edit the Integration Service associated with the SAP BW
Service.
• Processes view. View the state of the service process on each node. View or edit the directory of the
BWParam parameter file.
• Permissions view. View or edit group and user permissions on the SAP BW Service.
• Actions menu. Manage the service.
A web service gateway for external clients. It processes SOAP requests from web service clients that
want to access PowerCenter functionality through web services. Web service clients access the
PowerCenter Integration Service and PowerCenter Repository Service through the Web Services Hub.
When you select a Web Services Hub in the Navigator, the contents panel displays the following
information:
• Properties view. View the state of the service. View or edit Web Services Hub properties.
System Services
A system service is an application service that can have a single instance in the domain. When you create the
domain, the system services are created for you. You can enable, disable, and configure system services.
You can manage system services on the Services and Nodes view on the Manage tab.
The following image shows the System Services folder in the Domain Navigator:
Email Service
The Email Service sends email notifications for business glossaries, scorecards, and workflows. Enable
the Email Service to allow users to configure email notifications.
When you select the Email Service in the System Services folder in the Navigator, the contents panel
displays the following views and buttons:
• Properties view. View the state of the service. Manage the Model repository options and email server
properties for the service.
• Processes view. View the state of the service process for each assigned node.
• Permissions view. View or edit the group and user permissions on the service.
• Actions menu. Manage the service.
The Resource Manager Service manages computing resources in the domain and dispatches jobs to
achieve optimal performance and scalability. The Resource Manager Service collects information about
nodes with the compute role. The service matches job requirements with resource availability to identify
the best compute node to run the job.
The Resource Manager Service communicates with compute nodes in a Data Integration Service grid.
Enable the Resource Manager Service when you configure a Data Integration Service grid to run jobs in
separate remote processes.
When you select the Resource Manager Service in the System Services folder in the Navigator, the
contents panel displays the following views and buttons:
• Properties view. View the state of the service. Manage the log level and the primary and backup
nodes for the service.
• Processes view. View the state of the service process for each assigned node.
• Permissions view. View or edit the group and user permissions on the service.
• Actions menu. Manage the service.
The REST Operations Hub Service is an application service in the Informatica domain that exposes
Informatica product functionality to external clients through REST APIs.
Scheduler Service
The Scheduler Service manages schedules for profiles, scorecards, deployed mappings, and deployed
workflows. Enable the Scheduler Service to create, manage, and run schedules.
When you select the Scheduler Service in the System Services folder in the Navigator, the contents panel
displays the following views and buttons:
• Properties view. View the state of the service. Manage the log level, primary and backup nodes, and
Model Repository Service options for the service.
• Processes view. View the state of the service process for each assigned node and configure process
properties.
• Permissions view. View or edit the group and user permissions on the service.
• Actions menu. Manage the service.
Nodes
A node is a logical representation of a physical machine in the domain. On the Services and Nodes view on
the Manage tab, you assign resources to nodes and configure service processes to run on nodes that have
the service role.
When you select a node in the Navigator, the contents panel displays the following information:
• Properties view. View the status of the node. View or edit node properties, such as the repository backup
directory or range of port numbers for the processes that run on the node.
• Processes view. View the status of processes configured to run on the node. Service processes run on
nodes that have the service role.
• Resources view. View or edit resources assigned to the node.
• Permissions view. View or edit group and user permissions on the node.
In the Actions menu in the Navigator, you can delete the node, move the node to a folder, refresh the contents
on the Manage tab, or access help on the current tab.
Grids
A grid is an alias assigned to a group of nodes that run PowerCenter Integration Service or Data Integration
Service jobs.
When you run a job on a grid, the Integration Service distributes the processing across multiple nodes in the
grid. For example, when you run a profile on a grid, the Data Integration Service splits the work into multiple
jobs and assigns each job to a node in the grid. You assign nodes to the grid in the Services and Nodes view
on the Manage tab.
When you select a grid in the Navigator, the contents panel displays the following information:
Licenses
You create a license object on the Manage tab based on a license key file provided by Informatica.
After you create the license, you can assign services to the license.
When you select a license in the Navigator, the contents panel displays the following information:
• Properties view. View license properties, such as supported platforms, repositories, and licensed options.
You can also edit the license description.
• Assigned Services view. View or edit the services assigned to the license.
• Options view. View the licensed PowerCenter options.
• Permissions view. View or edit user permissions on the license.
In the Actions menu in the Navigator, you can delete the license, move the license to a folder, refresh the
contents on the Manage tab, or access help on the current tab.
In the Actions menu on the Manage tab, you can add an incremental key to a license.
Navigator
When you select the domain in the Navigator, you can create a connection.
When you select a connection in the Navigator, you can delete the connection.
Displays information about the domain or the connection that you select in the Navigator.
When you select the domain in the Navigator, the contents panel shows all connections in the domain. In
the contents panel, you can filter or sort connections, or search for specific connections.
When you select a connection in the Navigator, the contents panel displays information about the
connection. The tasks that you can complete for the connection vary depending on which of the
following views you select:
When you select a connection in the Navigator, you can test the connection.
Navigator
When you select the domain in the Navigator, the All Schedules view displays the name, status, and
description of the schedules in the domain.
Properties view
When you select a schedule in the Navigator, the Properties view displays details about the schedule.
When you select the domain in the Navigator, the Scheduled Jobs view displays details about the
scheduled jobs in the domain.
When you select a schedule in the Navigator, the Scheduled Jobs view displays details about the jobs
that are running on that schedule.
The actions that you can perform in the Actions menu vary based on which view you select.
You can perform the following actions using the Actions menu:
• Create a schedule
• Edit a schedule
• Delete a schedule
• Pause a schedule
• Resume a schedule
Monitor Tab
On the Monitor tab, you can monitor Data Integration Services and objects that the Data Integration Services
run.
After you configure the monitoring Model repository on the domain, you can view monitoring statistics in the
Administrator tool.
• Ad hoc jobs. Includes profiles, enterprise data profiles, mappings, scorecards, reference tables, and
previews.
• Applications. Includes deployed mappings, logical data objects, SQL data services, web services, and
workflows.
• SQL data service connections.
• Requests. Includes SQL data service and web service requests.
• Summary Statistics view. View graphical information about object state, distribution, and resource usage
across the Data Integration Services.
• Execution Statistics view. View properties, statistics, and reports for the objects that the Data Integration
Service runs.
Use the Summary Statistics view to view graphical information about object state, distribution, and resource
usage.
Time line
Specify a source and time range for which you want to view statistics.
View doughnut or tabular charts for jobs for the selected source and time range. Select sections of the
doughnuts to filter by job type, state, and Data Integration Service.
View doughnut or tabular charts for deployed mappings and workflows for the selected source and time
range. Select sections of the doughnuts to filter by type, state, and Data Integration Service.
View doughnut or tabular charts for Data Integration Service jobs for the selected source and time range.
Select sections of the doughnuts to filter by object type, state, and Data Integration Service.
View CPU and memory usage of all of the Data Integration Service processes that are running in the
domain or on a node in the domain.
The following image shows the Execution Statistics view on the Monitor tab:
1. Navigator
2. Folders
3. Actions menu
4. Contents panel
5. Details panel
6. Views in the details panel
When you select an object in the Navigator, you can view details about the object and monitor the object.
You can select the following types of objects in the Navigator in the Execution Statistics view:
Domain
View the states and properties of Data Integration Services in the domain.
View general properties about the Data Integration Service, and view statistics about objects that the
Data Integration Service runs.
View a list of objects in the folder. The folder is a logical grouping of objects. When you select a folder, a
list of objects appears in the contents panel. The contents panel shows multiple columns that show
properties about each object. You can configure the columns that appear in the contents panel.
The following table shows the folders that appear in the Navigator:
Folder Location
Depending on the type of object you select in the Navigator, the contents panel might display the following
views:
Properties view
Shows general properties and run-time statistics about the selected object. General properties might
include the name and description of the object. Statistics vary based on the selected object type.
Statistics view
Shows historical statistics about jobs in an application, or jobs that the Data Integration Service ran. For
example, when you select an application, you can view the number of deployed mapping jobs that failed
in the last four hours.
Reports view
Shows reports for the selected object. The reports contain key metrics for the object. For example, you
can view reports to determine the longest running jobs on a Data Integration Service during a particular
time period.
Summary Statistics view
Shows throughput and resource usage statistics for ad hoc mapping jobs, deployed mapping jobs, or
mappings in a workflow.
Shows graphs of the throughput and resource usage for ad hoc mapping jobs, deployed mapping jobs, or
mappings in a workflow.
Shows averaged data from multiple runs for a specific job. For example, you can view the minimum,
maximum, and average duration of the mapping job. You can view the average amount of CPU that the
job consumes when it runs.
Connections view
Shows connections defined for the selected object. You can view statistics about each connection, such
as the number of closed, aborted, and total connections.
Requests view
Shows details about requests. There are two types of requests: SQL queries and web service requests.
Users can use a third-party client tool to run SQL queries against the virtual tables in an SQL data
service. Users can use a web service client to run web service requests against a web service. Each web
service request runs a web service operation.
A request is a web services request or an SQL query that a user runs against a virtual table in an SQL
data service.
Shows virtual tables defined in an SQL data service. You can also view properties and cache refresh
details for each virtual table.
Operations view
The following table describes the types of jobs and statistics that you can view:
Connections Displays the following statistics about SQL data service connections to virtual databases:
- Total. Total number of connections.
- Closed. Number of database connections on which SQL data service requests have previously
run.
- Aborted. Number of connections that were manually aborted, or that were aborted when the Data
Integration Service was recycled or disabled in abort mode.
Requests Displays the following statistics about SQL data service and web service requests:
- Total. Total number of requests.
- Completed. Number of completed requests.
- Aborted. Requests that were aborted when the Data Integration Service was recycled, or
disabled in abort mode.
- Failed. Number of failed requests.
The reports that appear in the Reports view vary depending on which object you select. To view reports in the
Reports view, you must configure them in Monitor tab Actions > Report and Statistic Settings. By default, no
reports appear in the Reports view.
Shows ad hoc jobs that ran the longest during the specified time period. The report shows the job name,
ID, type, state, and duration. You can view this report in the Reports view when you monitor a Data
Integration Service in the Monitor tab.
Shows mapping jobs that ran the longest during the specified time period. The report shows the job
name, state, ID, and duration. You can view this report in the Reports view when you monitor a Data
Integration Service in the Monitor tab.
Shows profile jobs that ran the longest during the specified time period. The report shows the job name,
state, ID, and duration. You can view this report in the Reports view when you monitor a Data Integration
Service in the Monitor tab.
Shows reference table process jobs that ran the longest during the specified time period. Reference
table jobs are jobs where you export or import reference table data. The report shows the job name,
state, ID, and duration. You can view this report in the Reports view when you monitor a Data Integration
Service in the Monitor tab.
Shows scorecard jobs that ran the longest during the specified time period. The report shows the job
name, state, ID, and duration. You can view this report in the Reports view when you monitor a Data
Integration Service in the Monitor tab.
Shows SQL data service connections that were open the longest during the specified time period. The
report shows the connection ID, SQL data service, connection state, and duration. You can view this
report in the Reports view when you monitor a Data Integration Service, an SQL data service, or an
application in the Monitor tab.
Shows SQL data service requests that ran the longest during the specified time period. The report shows
the request ID, SQL data service, request state, and duration. You can view this report in the Reports view
when you monitor a Data Integration Service, an SQL data service, or an application in the Monitor tab.
Shows web service requests that were open the longest during the specified time period. The report
shows the request ID, web service operation, request state, and duration. You can view this report in the
Reports view when you monitor a Data Integration Service, a web service, or an application in the
Monitor tab.
Shows all workflows that were running the longest during the specified time period. The report shows
the workflow name, state, instance ID, and duration. You can view this report in the Reports view when
you monitor a Data Integration Service or an application in the Monitor tab.
Shows workflows that do not include a Human task that were running the longest during the specified
time period. The report shows the workflow name, state, instance ID, and duration. You can view this
report in the Reports view when you monitor a Data Integration Service or an application in the Monitor
tab.
Shows the minimum, maximum, and average duration for SQL data service and web service requests
during a specified time period. You can view this report in the Reports view when you monitor a Data
Integration Service, an SQL data service, a web service, or an application in the Monitor tab.
Shows the total number of SQL data service requests from each IP address during the specified time
period. You can view this report in the Reports view when you monitor a Data Integration Service, an SQL
data service, or an application in the Monitor tab.
Shows SQL data service connections that received the most connection requests during the specified
time period. The report shows the connection ID, SQL data service, and the total number of connection
requests. You can view this report in the Reports view when you monitor a Data Integration Service, an
application, or an SQL data service in the Monitor tab.
Shows users that ran the most number of ad hoc jobs during the specified time period. The report shows
the user name and the total number of jobs that the user ran. You can view this report in the Reports
view when you monitor a Data Integration Service in the Monitor tab.
Shows IP addresses that received the most number of web service requests during the specified time
period. You can view this report in the Reports view when you monitor a Data Integration Service, an
application, a web service, or web service operation in the Monitor tab.
Shows the most frequent errors for ad hoc jobs, regardless of job type, during the specified time period.
The report shows the job type, error ID, and error count. You can view this report in the Reports view
when you monitor a Data Integration Service in the Monitor tab.
Shows the most frequent errors for SQL data service requests during the specified time period. The
report shows the error ID and error count. You can view this report in the Reports view when you monitor
a Data Integration Service, an SQL data service, or an application in the Monitor tab.
Shows the most frequent faults for web service requests during the specified time period. The report
shows the fault ID and fault count. You can view this report in the Reports view when you monitor a Data
Integration Service, a web service, or an application in the Monitor tab.
When you select an ad hoc mapping job, deployed mapping job, or mapping in a workflow in the contents
panel, the details panel displays the Summary Statistics view. The Summary Statistics view displays run-time
statistics about the throughput and resource usage for the job. The run-time period begins when the Data
Integration Service starts reading from the first row.
You can view the following throughput statistics for the job:
• Executing node. Node where the Data Integration Service that ran the job is running.
• Average CPU usage. The average amount of CPU that the Data Integration Service used to run the job.
• Average Memory usage. The average amount of memory that the Data Integration Service used to run the
job.
When you select an ad hoc mapping job, deployed mapping job, or mapping in a workflow in the contents
panel of the Execution Statistics view, the details panel displays the Detailed Statistics view. The Detailed
Statistics view displays run-time statistics about the throughput and resources usage for the job. The run-
time period begins when the Data Integration Service starts reading from the first row.
The Detailed Statistics view displays the following graphs:
Throughput graph
Plots the number of rows read and written against the run time of the job.
Plots the percentage of the Data Integration Service CPU that was allocated to run the job against the
run time of the job.
Plots the amount of memory, in megabytes, that the Data Integration Service allocated to run the job
against the run time of the job.
Logs Tab
The Logs tab shows logs.
On the Logs tab, you can view the following types of logs:
• Domain log. Domain log events are log events generated from the domain functions that the Service
Manager performs.
• Service log. Service log events are log events generated by each application service.
• User Activity log. User Activity log events monitor user activity in the domain.
The Logs tab displays the following components for each type of log:
On the Reports tab, you can run the following domain reports:
• License Management Report. Run a report to monitor the number of software options purchased for a
license and the number of times a license exceeds usage limits. Run a report to monitor the usage of
logical CPUs and PowerCenter Repository Services. You run the report for a license.
• Web Services Report. Run a report to analyze the performance of web services running on a Web Services
Hub. You run the report for a time interval.
Security Tab
You administer Informatica security on the Security tab of the Administrator tool.
1. In the Search section, select whether you want to search for users, groups, or roles.
2. Enter the name or partial name to search for.
You can include an asterisk (*) in a name to use a wildcard character in the search. For example, enter
“ad*” to search for all objects starting with “ad”. Enter “*ad” to search for all objects ending with “ad”.
3. Click Go.
The Search Results section appears and displays a maximum of 100 objects. If your search returns more
than 100 objects, narrow your search criteria to refine the search results.
4. Select an object in the Search Results section to display information about the object in the contents
panel.
The Navigator on the Security tab displays one of the following sections based on what you are viewing:
• Groups section. Select a group to view the properties of the group, the users assigned to the group, and
the roles and privileges assigned to the group.
Reports Tab 57
• Users section. Select a user to view the properties of the user, the groups the user belongs to, and the
roles and privileges assigned to the user.
• Roles section. Select a role to view the properties of the role, the users and groups that have the role
assigned to them, and the privileges assigned to the role.
• Operating Profiles section. Select an operating profile to view the properties of the operating system
profile, and the permissions assigned to users and groups that use the operating system profile.
• LDAP Configuration section. Select a configuration to view the LDAP server connection details, the LDAP
security domain that contains users and groups imported from the LDAP directory service, and the LDAP
synchronization schedule.
The Navigator provides different ways to complete a task. You can use any of the following methods to
manage groups, users, and roles:
• Click the Actions menu. Each section of the Navigator includes an Actions menu to manage groups, users,
roles, operating system profiles, or LDAP configurations.
• Right-click an object. Right-click an object in the Navigator to display the options available in the Actions
menu.
• Use keyboard shortcuts. Use keyboard shortcuts to move to different sections of the Navigator.
Groups
A group is a collection of users and groups that can have the same privileges, roles, and permissions.
The Groups section of the Navigator organizes groups into security domain folders. A security domain is a
collection of user accounts and groups in an Informatica domain. Native authentication uses the Native
security domain which contains the users and groups created and managed in the Administrator tool. LDAP
authentication uses LDAP security domains which contain users and groups imported from the LDAP
directory service.
When you select a security domain folder in the Groups section of the Navigator, the contents panel displays
all groups belonging to the security domain.
When you select a group in the Navigator, the contents panel displays the following tabs:
• Overview. Displays general properties of the group and users assigned to the group.
• Privileges. Displays the privileges and roles assigned to the group for the domain and for application
services in the domain.
• Permissions. Displays the level of access that users within the group have perform tasks on domain
objects, including nodes, grids and application services . Also displays the level of access that users
within the group have to perform tasks on connection objects and operating system profiles.
Users
A user with an account in the Informatica domain can log in to the following application clients:
• Informatica Administrator
• PowerCenter Client
• Informatica Developer
• Informatica Analyst
• Metadata Manager
The Users section of the Navigator organizes users into security domain folders. A security domain is a
collection of user accounts and groups in an Informatica domain. Native authentication uses the Native
When you select a security domain folder in the Users section of the Navigator, the contents panel displays
all users belonging to the security domain.
When you select a user in the Navigator, the contents panel displays the following tabs:
• Overview. Displays general properties of the user and all groups to which the user belongs.
• Privileges. Displays the privileges and roles assigned to the user for the domain and for application
services in the domain.
• Permissions. Displays the level of access that the user has to perform tasks on domain objects, including
nodes, grids and application services . Also displays the level of access that the user has to perform tasks
on connection objects and operating system profiles.
Roles
A role is a collection of privileges that you assign to a user or group. Privileges determine the actions that
users can perform. You assign a role to users and groups for the domain and for application services in the
domain.
The Roles section of the Navigator organizes roles into the following folders:
• System-defined Roles. Contains roles that you cannot edit or delete. The Administrator role is a system-
defined role.
• Custom Roles. Contains roles that you can create, edit, and delete. The Administrator tool includes some
custom roles that you can edit and assign to users and groups.
When you select a folder in the Roles section of the Navigator, the contents panel displays all roles belonging
to the folder.
When you select a role in the Navigator, the contents panel displays the following tabs:
• Overview. Displays general properties of the role and the users and groups that have the role assigned for
the domain and application services.
• Privileges. Displays the privileges assigned to the role for the domain and application services.
The Operating System Profiles section of the Navigator lists the operating system profiles configured in the
domain.
When you select an operating system profile in the Navigator, the contents panel displays the following tabs:
• Properties. Displays general properties of the operating system profile configured for the Data Integration
Service, for the PowerCenter Integration Service, or for both application services.
• Permissions. Displays the permissions assigned to users and groups that use the operating system
profile. Also indicates whether the operating system profile is the default profile assigned to a user or
group.
Security Tab 59
LDAP Configuration
You can configure an Informatica domain to enable users and groups imported from one or more LDAP
directory services to log in to Informatica nodes, services, and application clients.
The LDAP Configuration section of the Navigator lists the LDAP configurations the domain uses.
When you select an LDAP configuration, the following tabs appear under the LDAP Configuration tab:
• Overview. Lists the connection details for the LDAP server that contains the directory service from which
you want to import users and groups.
• Security Domains. Lists the details for the LDAP security domain that contains users and groups imported
from the LDAP directory service.
• Schedule. Lists the details for the synchronization schedule specifying when the Service Manager updates
the security domain with the users and groups in the LDAP directory service.
Account Management
To improve security in the Informatica domain, you can enforce lockout of user and administrator accounts
after a specified number of failed login attempts.
The Account Lockout Configuration section of the Account Management page displays whether account
lockout is enabled for user accounts and administrator accounts. The section also indicates the maximum
number of failed login attempts allowed.
The Locked Out Native Users section of the page lists locked out user accounts in the native security domain.
You can unlock a user account in the native security domain.
The Locked Out LDAP Users section of the page lists locked out user accounts in an LDAP security domain.
You can unlock a user account in the Informatica domain. However, the LDAP administrator must unlock the
user account in the LDAP server. The user cannot log in to the Informatica domain until the LDAP
administrator unlocks the user account.
Audit Reports
Audit reports provide information about users and groups in the Informatica domain, and about the privileges,
roles, and permissions assigned to each user or group.
You select the audit report to generate from the Select Report Type menu. You can generate the following
audit reports:
Displays contact information and status details of user accounts in the domain. You can select the users
or groups for which you want to generate the report.
Displays information about users and the groups to which they belong. You can select the users or
groups for which you want to generate the report.
Privileges
Displays information about privileges assigned to the users and groups in the domain. You can select
the users or groups for which you want to generate the report.
Roles
Displays information about the roles assigned to the users and groups in the domain. You can select the
roles for which you want to generate the report.
Displays information about the domain objects for which users and groups have permission. You can
select the users or groups for which you want to generate the report.
Service States
You can identify the state of an application service by the icon that displays in the Administrator tool.
The following table describes the icons that indicate service states:
State Icon
Available
Disabled
Unavailable
Process States
You can identify the state of a Data Integration Service process or a PowerCenter Integration Service process
by the icon that displays in the Administrator tool.
The state icons also indicate the type of node on which the process runs. If the primary node has high
availability, a yellow diamond is superimposed on the process state icon. If the process runs on a grid, a grid
icon is superimposed on the process state icon.
The following table describes the icons that indicate process states:
State Icon
Aborted
Aborted (Grid)
Disabled
Service States 61
State Icon
Disabled (Grid)
Failed
Failed (Grid)
Running
Running (Grid)
Standing by or Delayed
Starting
Starting (Grid)
Stopped
Stopped (Grid)
Stopping
Stopping (Grid)
The following table describes the icons associated with each job state:
State Icon
Aborted
Canceled
Completed
Failed
Queued or Pending
Running
Starting
Stopped
Stopping
Terminated
Unknown
To turn the JAWS Virtual PC cursor on and off, use the keyboard shortcut Insert+Z.
Note: To use the JAWS screen reader with the Administrator tool, you must use Internet Explorer 11.
Job States 63
Keyboard Shortcuts
You can use keyboard shortcuts to navigate and work with the Administrator tool interface.
You can add, edit, and change values in the Administrator tool. Keyboard focus in the Administrator tool is
indicated by a blue border around the interface label. A dotted line appears around a selected object
indicating that the object is in focus. Tooltips appear when the label item receives keyboard focus or on
mouse-over. The navigation order of objects in the editor is from top to bottom and left to right.
Press Tab.
Press Shift+Tab.
Press the Space arrow key. To go back to the main menu, press Esc.
Edit the value of grid content such as the Access and Revoke fields in the Assign Permission and Edit Direct
Permissions dialog box
Note: You must enter appropriate values for all the form elements marked with an asterisk (*).
Move focus from Update Frequency drop-down menu to Time Range check box in the Statistics and Reports list grid in
the Report and Statistic Settings dialog box of the Monitor tab or Monitoring tool
Press Esc.
You cannot access the split bars in the Administrator tool and increase or decrease the size of the panels
using the keyboard. You cannot select multiple items with the Ctrl key in the Audit Reports tab under
Security.
• View current status, resource usage, and details for the domain and objects in the domain.
• View dependencies among objects in the domain.
• Perform domain actions such as shutting down the domain, enabling and disabling services, and shutting
down nodes.
• View recent service commands that users issued from the Administrator tool.
• View historical information about status, resource usage, and events in the domain.
65
Dependency Graph
The Dependency graph displays dependencies among services, nodes, and grids in the Informatica domain.
The following image shows the Dependency graph for a Model Repository Service:
You can use the Dependency graph to perform the following tasks:
When you view dependencies for an object, the Dependency graph displays the upstream and downstream
dependencies. Upstream dependencies are objects on which the selected object depends. Downstream
dependencies are objects that depend on the selected object.
When you enable, disable, or recycle services from the Dependency graph, the actions appear in the
Command History panel.
Service Node on which the service process runs. Services that depend on the service.
Service running The service has the following upstream Services that depend on the service.
on a grid dependencies:
- Nodes on which the service process
runs.
- The grid on which the service
processes run.
Service running in Primary and backup nodes on which the Services that depend on the service.
HA mode service processes can run.
Grid Nodes assigned to the grid. Services that are running on the grid.
3. In the Dependency graph, you can optionally complete the following tasks:
• Select Show Legend to view information about the icons and lines used in the graph.
• Click and drag to view different parts of the graph.
• Zoom in or zoom out of the graph.
• To exit the Dependency graph, click Close.
Downstream services are services that depend on other services. For example, a Data Integration Service
depends on a Model Repository Service. You recycle or disable downstream services using the Actions menu
for the service on which they depend. When you disable downstream services, the service processes abort.
Dependency Graph 67
Command History
The Command History panel on the Domain view displays recent service lifecycle commands that users
issued from the Administrator tool. Service lifecycle commands include enable, disable, and recycle.
To view the command history, click Domain Actions > View Command History.
You can view the following information about the commands in the Command History panel:
• Service Name. Name of the service for which the command was issued.
• Service Type.
• Command.
• Status. Can be Failed, Success, or Queued.
• Status Updated
• Comments. Comments that users entered when they recycled or disabled the service.
• Message. Log messages associated with the command.
Optionally, you can show or hide columns in the command history. To change the columns, right-click the
column header, and then select or clear columns.
Note: The command history is erased when you shut down or restart the master gateway node.
History View
The History view displays historical data for the domain, a service, or a node. You can view historical data for
the contents panel, Service State Summary, resource usage indicators, and Details panel. You can also view
information about events in the domain.
1. Time line
2. Event circle
3. Last Refreshed
4. Time line slider
5. Events panel
Use the time line to choose which time range and events you would like to view. When events occur
around a point in time, the point is marked with a red event circle. When you select a circle, the Events
panel displays the events that occurred around that point in time. By default, the time line displays
crashes and unplanned shutdowns that occurred in the last one day.
Events panel
Displays events that occurred during a specified time range. When you open the History view, the panel
lists crashes and unplanned shutdowns that occurred in the last one day.
Contents panel
Displays current and historical domain contents and states. When you open the History view, the
contents panel shows domain objects and states at the last refresh. When you drag the time line slider
to a point in time, the contents panel shows domain objects and states at that point in time.
Displays current and historical service state summaries. When you open the History view from the
domain Actions menu, the Service State Summary appears and displays the number and states of
services at the last refresh. When you drag the time line slider to a point in time, the Service State
Summary displays the number of services that were available or unavailable at that time.
Details panel
Displays current and historical state for a service or node. When you open the History view, the Details
panel displays the state of the service or node at the last refresh. When you drag the time line slider to a
point in time, the Details panel displays the object state at that point in time.
Display current and historical resource usage information. When you open the History view, the usage
indicators display the usage statistics as they appeared at the last refresh. When you drag the time line
slider to a point in time, the indicators display statistics for that point in time.
The data that you view in the History view is stored in the monitoring Model repository. Before you can view
historical data, you must configure the monitoring Model repository in the Manage > Services and Nodes >
Monitoring Configuration tab. You can view the per-minute data for up to two weeks in the past. After two
weeks, the statistics that you view are hourly averages. You can view hourly averaged data for up to one year
in the past.
Note: The monitoring Model Repository Service that you configure to store the historical data cannot capture
data about events that occur while it is unavailable or disabled. For example, Start events for the monitoring
Model Repository Service and Crash events for the node that the monitoring Model Repository Service
process runs on do not appear in the History view.
Viewing History
You can view historical statistics for the domain, a service, or a node.
The amount of history that you can view depends on the monitoring Model repository options that you
configure in the Monitoring Configuration tab. You must configure a monitoring Model repository in the
Monitoring Configuration tab before you can view historical statistics.
1. To access the History view, click the Actions menu for the domain, a service, or a node, and then select
View History.
The History view opens. The auto-refresh pauses at the Last Refreshed time stamp. The time line and
Events panel display and display crashes and unplanned shutdowns that occurred in the last one day.
2. To change the time range, select a time range in the Time Range list.
History View 69
When you change the time range, the time line refreshes and displays the selected time range.
3. To choose a custom time range, select Custom in the Time Range list.
You can select a custom time range from one hour to one month.
4. To change the events that display, select events in the Events list.
You can view information about the following events:
• Crash
• Failed to Start
• Unplanned Shutdown
• Planned Shutdown
• Start
When you select different events, the time line refreshes and reflects the change.
5. To view historical statistics about the domain, drag the time line slider to a point in time.
The History view refreshes and displays the domain status at that point in time.
6. To exit the History view, click Close.
Note: Current services display as Unavailable when the time range begins before they were created.
Viewing Events
You can view events for the domain, a node, or a service.
1. To view events, click the Actions menu for the domain, a node, or a service, and then choose View
History.
The History view opens and displays the Events panel. By default, the Events panel lists crashes and
unplanned shutdowns that have occurred in the last one day.
1. You can view the following information about events:
• Object name
• Object type
• Event type
• Time the event occurred
• Name of the associated node
• Comments about recycled or disabled services
• Comments about shut down nodes
2. Optionally, you can complete the following actions in the Events panel:
Option Description
Search events Type the search criteria in the search field, and then press enter. To clear the search, click
the Reset Filters icon.
Sort a column To sort a column in ascending order, click the column header. To sort the column in
descending order, click the column header again.
Add or remove To change the columns that appear in the Events panel, right-click a column header, and
columns then select or deselect a column.
Reset filter When you click a circle on the time line, you can clear the selection by clicking Reset filters.
History View 71
Chapter 5
Domain Management
This chapter includes the following topics:
• Manage alerts. Configure, enable, and disable domain and service alerts for users.
• Create folders. Create folders to organize domain objects and manage security by setting permission on
folders.
• Manage domain security. Configure secure communication between domain components.
• Manage user security. Assign privileges and permissions to users and groups.
• Manage application services. Enable, disable, recycle, and remove application services. Enable and
disable service processes.
• Manage nodes. Configure node properties, such as the backup directory and resources, and shut down
nodes.
• Configure gateway nodes. Configure nodes to serve as a gateway.
• Shut down the domain. Shut down the domain to complete administrative tasks on the domain.
72
• Manage domain configuration. Back up the domain configuration on a regular basis. You might need to
restore the domain configuration from a backup to migrate the configuration to another database user
account. You might also need to reset the database information for the domain configuration if it
changes.
• Complete domain tasks. You can monitor the statuses of all application services and nodes, view
dependencies among application services and nodes, and shut down the domain.
• Configure domain properties. For example, you can change the database properties, SMTP properties for
alerts, and domain resiliency properties.
To manage nodes and services through a single interface, all nodes and services must be in the same
domain. You cannot access multiple Informatica domains in the same Administrator tool window. You can
share metadata between domains when you register or unregister a local repository in the local Informatica
domain with a global repository in another Informatica domain.
Alert Management
Alerts provide users with domain and service alerts. Domain alerts provide notification about node failure and
master gateway election. Service alerts provide notification about service process failover.
Property Description
Host Name The SMTP outbound mail server host name. For example, enter the Microsoft Exchange
Server for Microsoft Outlook.
Port Port used by the outgoing mail server. Valid values are from 1 to 65535. Default is 25.
User name The user name for authentication upon sending if required by the outbound mail server.
Alert Management 73
Property Description
Password The user password for authentication upon sending if required by the outbound mail server.
Sender Email The email address that the Service Manager uses in the From field when sending notification
Address emails. If you leave this field blank, the Service Manager uses Administrator@<host
name> as the sender.
7. Click OK.
Subscribing to Alerts
After you complete the SMTP configuration, you can subscribe to alerts.
1. Verify that the domain administrator has entered a valid email address for your user account on the
Security page.
If the email address or the SMTP configuration is not valid, the Service Manager cannot deliver the alert
notification.
2. In the Administrator tool header area, click Manage > Preferences.
The Preferences page appears.
3. In the User Preferences section, click Edit.
The Edit Preferences dialog box appears.
4. Select Subscribe for Alerts.
5. Click OK.
6. Click OK.
The Service Manager sends alert notification emails based on your domain privileges and permissions.
The following table lists the alert types and events for notification emails:
Viewing Alerts
When you subscribe to alerts, you can receive domain and service notification emails for certain events.
When a domain or service event occurs that triggers a notification, you can track the alert status in the
following ways:
• The Service Manager sends an alert notification email to all subscribers with the appropriate privilege and
permission on the domain or service.
• The Log Manager logs alert notification delivery success or failure in the domain or service log.
For example, the Service Manager sends the following notification email to all alert subscribers with the
appropriate privilege and permission on the service that failed:
From: Administrator@<database host>
To: Jon Smith
Folder Management
Use folders in the domain to organize objects and to manage security.
Folders can contain nodes, services, grids, licenses, and other folders. You might want to use folders to
group services by type. For example, you can create a folder called IntegrationServices and move all
Integration Services to the folder. Or, you might want to create folders to group all services for a functional
area, such as Sales or Finance.
When you assign a user permission on the folder, the user inherits permission on all objects in the folder.
• View services and nodes. View all services in the folder and the nodes where they run. Click a node or
service name to access the properties for that node or service.
• Create folders. Create folders to group objects in the domain.
• Move objects to folders. When you move an object to a folder, folder users inherit permission on the
object in the folder. When you move a folder to another folder, the other folder becomes a parent of the
moved folder.
• Remove folders. When you remove a folder, you can delete the objects in the folder or move them to the
parent folder.
Note: The System_Services folder is created for you when you create the domain, and contains all of the
system services. A system service is an application service that can have a single instance in the domain.
You cannot delete, move, or edit the properties or contents of the System_Services folder.
Creating a Folder
You can create a folder in the domain or in another folder.
1. In the Administrator tool, click the Manage tab > Services and Nodes view.
2. In the Domain Navigator, select the domain or folder in which you want to create a folder.
3. On the Navigator Actions menu, click New > Folder.
Folder Management 75
4. Edit the following properties:
Name Name of the folder. The name is not case sensitive and must be unique within the
domain. It cannot exceed 80 characters or begin with @. It also cannot contain
spaces or the following special characters:
`~%^*+={}\;:'"/?.,<>|!()][
Description Description of the folder. The description cannot exceed 765 characters.
5. Click OK.
Note: The domain serves as a folder when you move objects in and out of folders.
1. In the Administrator tool, click the Manage tab > Services and Nodes view.
2. In the Domain Navigator, select an object.
3. On the Navigator Actions menu, select Move to Folder.
4. In the Select Folder dialog box, select a folder, and click OK.
Removing a Folder
When you remove a folder, you can delete the objects in the folder or move them to the parent folder.
1. In the Administrator tool, click the Manage tab > Services and Nodes view.
2. In the Domain Navigator, select a folder.
3. On the Navigator Actions menu, select Delete.
4. Confirm that you want to delete the folder.
You can delete the contents only if you have the appropriate privileges and permissions on all objects in
the folder.
5. Choose to wait until all processes complete or to abort all processes.
6. Click OK.
You can configure secure communication between services within the domain.
You can configure secure communication between Informatica domain components and web browsers
or web service clients.
Each method of configuring secure communication is independent of the other methods. When you configure
secure communication for one set of components, you do not need to configure secure communication for
any other set.
Note: If you change a secure domain to a non-secure domain or from a non-secure domain to a secure
domain, you must delete the domain configuration in the Developer tool and PowerCenter client tools and
configure the domain again in the client.
Privileges determine the actions that users can complete on domain objects. Permissions define the level of
access a user has to a domain object. Domain objects include the domain, folders, nodes, grids, licenses,
database connections, operating system profiles, and application services.
Even if a user has the domain privilege to complete certain actions, the user might also require permission to
complete the action on a particular object. For example, a user has the Manage Services domain privilege
which grants the user the ability to edit application services. However, the user also must have permission on
the application service. A user with the Manage Services domain privilege and permission on the
Development Repository Service but not on the Production Repository Service can edit the Development
Repository Service but not the Production Repository Service.
To log in to the Administrator tool, a user must have the Access Informatica Administrator domain privilege.
If a user has the Access Informatica Administrator privilege and permission on an object, but does not have
the domain privilege that grants the ability to modify the object type, then the user can view the object. For
example, if a user has permission on a node, but does not have the Manage Nodes and Grids privilege, the
user can view the node properties but cannot configure, shut down, or remove the node.
If a user does not have permission on a selected object in the Navigator, the contents panel displays a
message indicating that permission on the object is denied.
Note: You can perform all of the common administration tasks for system services, except for removing the
system service.
The behavior of a service when it starts service processes depends on how it is configured:
• If the service is configured for high availability, then the service starts the service process on the primary
node. The service processes on the backup nodes are in Standing By state.
• If the service is configured to run on a grid, then the service starts service processes on all nodes that
have the service role.
The state of a service depends on the state of its processes. A service can have the following states:
• Available. You have enabled the service and at least one service process is running. The service is
available to process requests.
• Unavailable. You have enabled the service and none of its processes are running. This can be because the
service processes are disabled or failed to start. The service is not available to process requests.
• Disabled. You have disabled the service.
You can disable a service to perform a management task, such as changing the data movement mode for a
PowerCenter Integration Service. You might want to disable the service process on a node if you need to shut
down the node for maintenance. When you disable a service, all associated service processes stop, but they
remain enabled.
The following table describes the different states of a service and its processes:
Enabled Standing By The service process is enabled but is not running because another service
process is running as the primary service process. It is on standby to run
in case of service failover.
Disabled Disabled The service is enabled but the service process is not running on the node.
Enabled Failed The service and service process are enabled, but the service process could
not start.
1. In the Administrator tool, click the Manage tab > Services and Nodes view.
2. In the Domain Navigator, select a service.
3. In the contents panel, select the Processes view.
The Processes view displays the state of the processes.
1. In the Administrator tool, click the Manage tab > Services and Nodes view.
2. In the Domain Navigator, select the domain.
3. In the Properties view, configure the following restart properties:
Maximum Restart Number of times within a specified period that the domain attempts to restart an
Attempts application service process when it fails. The value must be greater than or equal to 1.
Default is 3.
Within Restart Maximum period of time that the domain spends attempting to restart an application
Period (sec) service process when it fails. If a service fails to start after the specified number of
attempts within this period of time, the service does not restart. Default is 900.
Disable the service before you delete the service to ensure that the service is not running any processes. If
you do not disable the service, you may have to choose to wait until all processes complete or abort all
processes when you delete the service.
1. In the Administrator tool, click the Manage tab > Services and Nodes view.
2. In the Domain Navigator, select the application service.
3. In the Manage tab Actions menu, select Delete.
4. In the warning message that appears, click Yes to stop other services that depend on the application
service.
5. If the Disable Service dialog box appears, choose to wait until all processes complete or abort all
processes, and then click OK.
I think that a service is using incorrect environment variable values. How can I find out which
environment variable values are used by a service.
Set the error severity level for the node to debug. When the service starts on the node, the Domain log will
display the environment variables that the service is using.
One gateway node in the domain serves as the master gateway node for the domain. The Service Manager on
the master gateway node accepts service requests and manages the domain and services in the domain. If
the domain has one gateway node and it becomes unavailable, the domain cannot accept service requests. If
the domain has multiple gateway nodes and the master gateway node becomes unavailable, the Service
Managers on the other gateway nodes elect a new master gateway node. The new master gateway node
accepts service requests. Only one gateway node can be the master gateway node at any particular time.
You can make the following changes to the gateway configuration for the domain:
You can convert a worker node to serve as a gateway node if the worker node is running and has the
service role enabled. When you convert a worker node to a gateway node, you must specify the log
directory for the node. If you have multiple gateway nodes, configure all gateway nodes to write log files
to the same directory on a shared disk.
After you convert a worker node to a gateway node, the Service Manager on the master gateway node
writes the domain configuration database connection to the nodemeta.xml file of the new gateway node.
You can convert a gateway node to serve as a worker node if another node in the domain is configured
as a gateway node.
If you convert a master gateway node to serve as a worker node, you must restart the node to make the
Service Managers elect a new master gateway node. If you do not restart the node, the node continues
as the master gateway node until you restart the node or the node becomes unavailable.
1. In the Administrator tool, click the Manage tab > Services and Nodes view.
2. In the Domain Navigator, select the domain.
3. In the contents panel, select the Properties view.
4. In the Properties view, click Edit in the Gateway Configuration Properties section.
5. To convert a worker node to a gateway node, complete the following steps:
a. Select the check box next to the node.
b. If the domain uses a secure domain configuration database, specify the truststore file and
password for the database.
c. Configure the directory path for the log files for each node that you convert to a gateway node.
If you have multiple gateway nodes, configure all gateway nodes to write log files to the same
directory on a shared disk.
Note: You must use the infacmd isp SwitchToWorkerNode command to convert a a worker node to
a gateway node in a domain configured to use SAML authentication. See the Informatica Command
Reference for instructions on using the infacmd isp SwitchToWorkerNode command.
6. To convert a gateway node to a worker node, clear the check box next to the node.
• Back up the domain configuration. Back up the domain configuration on a regular basis. You may need to
restore the domain configuration from a backup if the domain configuration in the database becomes
corrupt.
• Restore the domain configuration. You may need to restore the domain configuration if you migrate the
domain configuration to another database user account. Or, you may need to restore the backup domain
configuration to a database user account.
• Migrate the domain configuration. You may need to migrate the domain configuration to another database
user account.
• Configure the connection to the domain configuration database. Each gateway node must have access to
the domain configuration database. You configure the database connection when you create a domain. If
you change the database connection information or migrate the domain configuration to a new database,
you must update the database connection information for each gateway node.
• Configure custom properties. Configure domain properties that are unique to your environment or that
apply in special cases. Use custom properties only if Informatica Global Customer Support instructs you
to do so.
Note: The domain configuration database and the Model repository cannot use the same database user
schema.
Run the infasetup BackupDomain command to back up the domain configuration to a binary file.
Note: If the infasetup BackupDomain command fails with a Java memory error, increase the system memory
available for infasetup. To increase system memory, set the -Xmx value in the INFA_JAVA_COMD_OPTS
environment variable.
When you run this command, infasetup backs up the domain configuration database tables. To restore the
domain to another database, you must back up the ISP_RUN_LOG table contents manually to get the previous
workflow and session logs.
Additionally, use the database backup utility to manually back up additional repository tables that the
infasetup command does not back up.
Informatica restores the domain configuration from the current version. If you have a backup file from an
earlier product version, you must use the earlier version to restore the domain configuration.
You can restore the domain configuration to the same or a different database user account. If you restore the
domain configuration to a database user account with existing domain configuration, you must configure the
command to overwrite the existing domain configuration. If you do not configure the command to overwrite
the existing domain configuration, the command fails.
Each node in a domain has a host name and port number. When you restore the domain configuration, you
can disassociate the host names and port numbers for all nodes in the domain. You might do this if you want
to run the nodes on different machines. After you restore the domain configuration, you can assign new host
names and port numbers to the nodes. Run the infasetup DefineGatewayNode or DefineWorkerNode
command to assign a new host name and port number to a node.
If you restore the domain configuration to another database, you must reset the database connections for all
gateway nodes.
Important: You lose all data in the summary tables when you restore the domain configuration.
1. Disable the application services. Disable the application services in complete mode to ensure that you
do not abort any running service process. You must disable the application services to ensure that no
service process is running when you shut down the domain.
2. Shut down the domain. You must shut down the domain to ensure that no change to the domain occurs
while you are restoring the domain.
3. Run the infasetup RestoreDomain command to restore the domain configuration to a database. The
RestoreDomain command restores the domain configuration in the backup file to the specified database
user account.
4. Assign new host names and port numbers to the nodes in the domain if you disassociated the previous
host names and port numbers when you restored the domain configuration. Run the infasetup
DefineGatewayNode or DefineWorkerNode command to assign a new host name and port number to a
node.
5. Reset the database connections for all gateway nodes if you restored the domain configuration to
another database. All gateway nodes must have a valid connection to the domain configuration
database.
Important: Summary tables are lost when you restore the domain configuration.
Disable application services in complete mode to ensure that you do not abort running service processes.
To update the node with the new database connection information, complete the following steps:
To update the node after you change the user or password, complete the following steps:
To redefine the node after you change the host name or port number, complete the following steps:
If the Informatica domain uses Kerberos authentication, all service and node SPNs have the same Kerberos
realm name. After you change the Informatica domain name, you must generate SPNs and keytab files with
the new Informatica domain name.
To rename the domain, complete the following tasks:
1. If the domain contains a PowerCenter global repository, you must unregister all local repositories from
the global repository.
2. Shut down the domain. Shut down the domain through the Administrator tool, ensuring that all nodes are
shut down.
3. Back up the domain with the infasetup BackupDomain command.
4. Back up the sitekey and keytab files. By default, the files are in the following location:
<Informatica installation directory>/isp/config/keys
5. Update the domain and nodes.
To update the domain name, run the infasetup updateDomainName command from any gateway node.
Run the updateGatewayNode and updateWorkerNode commands with the updated domain name for all
the gateway and worker nodes.
6. On PowerCenter, register the local repository with a connected global repository with the updated
domain name with the pmrep Register command.
7. You can create SPN and keytab files with the updated domain name for Kerberos authentication. Copy
the keytab files in the keys directory. You can continue to use the older site key file. If you need to
regenerate the site key when it is missing or corrupted, you must provide the older domain name.
8. Optionally, you can run the License Management Report in the Administrator tool to review the updated
domain name.
9. You must configure the Informatica clients to use the updated domain name.
When you shut down a domain, the Service Manager on the master gateway node stops all application
services and Informatica services in the domain. Any service processes running on nodes in the domain are
aborted. To avoid possible loss of data or metadata and allow running processes to complete, you can shut
down each node from the Administrator tool or from the operating system.
Before you shut down a domain, verify that all processes, including workflows, have completed and no users
are logged in to repositories in the domain.
Domain Properties
On the Manage tab, you can configure domain properties including database properties, gateway
configuration, and service levels.
To view and edit properties, click the Manage tab. In the Navigator, select a domain. Then click the
Properties view in the contents panel. The contents panel shows the properties for the domain.
You can configure the properties to change the domain. You cannot change the database properties in the
Admin Console. You need to change these properties using the Command UpdateGatewayNode. You can
change SMTP properties for alerts, and the domain resiliency properties.
• General properties. Edit general properties, such as service resilience and dispatch mode.
• Database properties. View the database properties, such as database name and database host.
• Gateway configuration properties. Configure a node to serve as gateway and specify the location to write
log events.
• Service level management. Create and configure service levels.
• SMTP configuration. Edit the SMTP settings for the outgoing mail server to enable alerts.
• Custom properties. Edit custom properties that are unique to the Informatica environment or that apply in
special cases. When you create a domain, it has no custom properties. Use custom properties only at the
request of Informatica Global Customer Support.
General Properties
In the General Properties area, you can configure general properties for the domain.
The following table describes the properties that you can edit in the General Properties area:
Property Description
Resilience Timeout The number of seconds that an application service tries to connect or reconnect to the
PowerCenter Repository Service or the PowerCenter Integration Service. Valid values are
from 0 to 1000000. Default is 30 seconds.
Limit on Resilience The maximum number of seconds that application clients or application services can try to
Timeout connect or reconnect to the PowerCenter Repository Service or the PowerCenter Integration
Service. Default is 180 seconds.
Restart Period The maximum amount of time in seconds that the domain spends trying to restart an
application service process. Valid values are from 0 to 1000000.
Maximum Restart The number of times that the domain tries to restart an application service process. Valid
Attempts within values are from 0 to 1000. If you set the value as 0, the domain does not try to restart the
Restart Period service process.
Dispatch Mode The mode that the Load Balancer uses to dispatch PowerCenter Integration Service tasks to
nodes in a grid. Select one of the following dispatch modes:
- MetricBased
- RoundRobin
- Adaptive
Enable Secure Configures services to use the TLS protocol to transfer data securely within the domain.
Communication When you enable secure communication for the domain, services use secure connections to
communicate with other Informatica application services and clients.
Verify that all domain nodes are available before you enable secure communication for the
domain. If a node is not available, the secure communication changes cannot be applied to
the Service Manager of the node. To apply changes, restart the domain. Set this property to
True or False.
Service Resilience The maximum number of seconds that application clients and application services can try to
Timeout connect to the Data Integration Service or to the Model Repository Service. The default is 180
seconds.
Database Properties
In the Database Properties area, you can view the database properties for the domain, such as database
name and database host. You cannot edit these properties in the Admin Console. You need to update these
properties using the command UpdateGatewayNode.
Property Description
Database Type The type of database that stores the domain configuration metadata.
Database User The user account for the database containing the domain configuration information.
Database TLS enabled Indicates whether the database for the domain configuration repository is a secure database.
True if the domain configuration repository database is secure. You can use a secure domain
configuration repository if secure communication is enabled for the Informatica domain.
Domain Properties 87
Note: The service manager uses the DataDirect drivers included with the Informatica installation. Informatica
does not support the use of any other database driver.
If you edit gateway configuration properties, previous logs do not appear. Also, the changed properties apply
to restart and failover scenarios only.
To sort gateway configuration properties, click the header of the column by which you want to sort.
The following table describes the properties that you can edit in the Gateway Configuration Properties area:
Property Description
Gateway To configure the node as a gateway node, select this option. If the domain uses a
secure domain configuration database, you must specify the truststore file and
password for the database.
To configure the node as a worker node, clear this option.
Log Directory Path The directory path for the log event files. If the Log Manager cannot write to the
directory path, it writes log events to the node.log file on the master gateway node.
If you configure multiple gateway nodes for the domain, set the database truststore file and password for all
gateway nodes.
Property Description
Database Truststore Location Path and file name of the truststore file for the secure database.
Note: To use a secure domain configuration repository database, the secure communication option must be
enabled for the domain.
Service levels set priorities among tasks that are waiting to be dispatched. When the Load Balancer has more
tasks to dispatch than the PowerCenter Integration Service can run at the time, the Load Balancer places
Because service levels are domain properties, you can use the same service levels for all repositories in a
domain. You create and edit service levels in the domain properties or by using infacmd.
You can edit but you cannot delete the Default service level, which has a dispatch priority of 5 and a
maximum dispatch wait time of 1800 seconds.
To edit a service level, click the link for the service level.
To delete a service level, select the service level and click the Delete button.
The following table describes the properties that you can edit in the Service Level Management area:
Property Description
Name The name of the service level. The name is not case sensitive and must be unique
within the domain. It cannot exceed 128 characters or begin with the @ character. It
also cannot contain spaces or the following special characters:
` ~ % ^ * + = { } \ ; : / ? . < > | ! ( ) ] [
After you add a service level, you cannot change its name.
Dispatch Priority A number that sets the dispatch priority for the service level. The Load Balancer
dispatches high priority tasks before low priority tasks. Dispatch priority 1 is the
highest priority. Valid values are from 1 to 10. Default is 5.
Maximum Dispatch Wait The amount of time in seconds that the Load Balancer waits before it changes the
Time (seconds) dispatch priority for a task to the highest priority. Setting this property ensures that
no task waits forever in the dispatch queue. Valid values are from 1 to 86400. Default
is 1800.
SMTP Configuration
Use the SMTP Configuration properties to configure SMTP settings for the domain. The outgoing mail server
uses the SMTP settings to send alerts and scorecard notifications.
The following table describes the properties that you can edit in the SMTP Configuration area:
Property Description
Host Name The SMTP outbound mail server host name. For example, enter the Microsoft Exchange Server for
Microsoft Outlook.
Port Port used by the outgoing mail server. Valid values are from 1 to 65535. Default is 25.
User name The user name for authentication upon sending if required by the outbound mail server.
Password The user password for authentication upon sending if required by the outbound mail server.
Sender Email The email address that the Service Manager uses in the From field when sending notification
Address emails. If you leave this field blank, the Service Manager uses Administrator@<host name> as
the sender.
Domain Properties 89
Custom Properties for the Domain
Configure custom properties that are unique to specific environments.
You might need to apply custom properties in special cases. When you define a custom property, enter the
property name and an initial value. Define custom properties only at the request of Informatica Global
Customer Support.
Nodes
This chapter includes the following topics:
• Nodes Overview, 91
• Node Types, 92
• Node Roles, 93
• Define and Add Nodes, 95
• Configuring Node Properties, 96
• Shutting Down and Restarting the Node, 98
• Removing the Node Association, 100
• Removing a Node, 100
Nodes Overview
A node is the logical representation of a machine in the domain. When you configure a domain with multiple
nodes, you can scale service processing across multiple machines. The Service Manager runs on all nodes in
the domain to support the domain and application services. If the Service Manager is not running, the node is
not available.
An installation on multiple machines consists of a master gateway node, which hosts the domain, and
additional gateway nodes and worker nodes that run Informatica application services. The node type
determines whether the node can serve as a gateway or a worker node and determines the domain functions
that the node performs. You define the node type when you install Informatica services and join the node to
the domain. You can use the Administrator tool to change the node type after installation.
By default, each node in the domain can run application services and computational processes. The node
role determines whether a node can run application services, computational processes, or both. If the node
has the service role, you can view the application service processes running on the node. Before you remove
or shut down a node, verify that all running processes are stopped. You might need to shut down the node if
you need to perform maintenance on the machine or to ensure that domain configuration changes take
effect.
Use the Manage tab of the Administrator tool to manage nodes, including configuring node properties,
updating a node role, and removing nodes from a domain. The properties that you can configure depend on
the node role.
If your license includes grid, you can configure the Data Integration Service or the PowerCenter Integration
Service to run on a grid. A grid is an alias assigned to a group of nodes. When you run jobs on a grid of nodes,
you improve scalability and performance by distributing jobs to processes running on multiple nodes in the
grid. When the PowerCenter Integration Service runs on a grid, you can configure it to check the resources
91
available on each node. Assign connection resources and define custom and file/directory resources on a
node that is assigned to a PowerCenter Integration Service grid.
Node Types
The node type determines whether the node can serve as a gateway or worker node and determines the
domain functions that the node performs.
You define the node type when you install Informatica services and join the node to the domain. You can use
the Administrator tool to change the node type after installation. You change the node type in the gateway
configuration properties for the domain.
Related Topics:
• “Gateway Configuration” on page 80
Gateway Nodes
A gateway node is any node that you configure to serve as a gateway for the domain. A gateway node can run
application services and perform computations, and it can serve as a master gateway node. One gateway
node acts as the master gateway at any given time. The master gateway node is the entry point to the
domain.
The Service Manager on the master gateway node performs all domain functions on the master gateway
node. The Service Managers running on other gateway nodes perform limited domain functions on those
nodes.
You can configure more than one node to serve as a gateway. If the master gateway node becomes
unavailable, the Service Managers on other gateway nodes elect another master gateway node. If you
configure only one node to serve as the gateway and the node becomes unavailable, the domain cannot
accept service requests.
Worker Nodes
A worker node is any node that you do not configure to serve as a gateway for the domain. A worker node
can run application services and perform computations, but it cannot serve as a gateway. The Service
Manager performs limited domain functions on a worker node.
All nodes run the Service Manager. Node 1 is the master gateway node and runs two application services.
Node 2 is a back-up gateway node and runs one application service. Node 3 is a worker node and runs one
application service. If Node 1 becomes unavailable, Node 2 is elected as the new master gateway node. The
Service Manager on Node 2 then performs all domain functions. When Node 1 restarts, it becomes a back-up
gateway node and the Service Manager performs limited domain functions.
92 Chapter 6: Nodes
The following image shows a domain with two gateway nodes and one worker node:
Node Roles
The node role defines the purpose of the node. A node with the service role can run application services. A
node with the compute role can perform computations requested by remote application services. A node with
both roles can run application services and locally perform computations for those services.
By default, each gateway and worker node has both the service and compute roles enabled. Each node must
have at least one role enabled.
You can configure a Data Integration Service grid so that some nodes are dedicated to running application
service processes while other nodes are dedicated to performing computations. When you enable only the
compute role on a node in a Data Integration Service grid, the node does not have to run the service process.
The machine uses all available processing power to run mappings. You can add additional nodes with only
the compute role to the grid to increase scalability of Data Integration Service mappings.
For more information about setting up a Data Integration Service grid, see the Informatica Application Service
Guide.
Service Role
A node with the service role can run application services.
When you enable the service role on a node, the Service Manager supports application services configured to
run on that node.
Node Roles 93
Compute Role
A node with the compute role can perform computations requested by remote application services.
When a node has the compute role, the Service Manager manages the containers on the node. A container is
an allocation of memory and CPU resources. An application service uses the container to remotely perform
computations on the node. For example, a Data Integration Service grid includes Node 1 with the service role
and Node 2 with the compute role. The Data Integration Service process that runs on Node 1 runs a mapping
within a container on Node 2.
A node requires the compute role when the Data Integration Service runs jobs on the node. When the Data
Integration Service runs on a single node, the node must have both the service and compute roles. When the
Data Integration Service runs on a grid, at least one of the nodes in the grid must have the compute role.
A node does not require the compute role when the Data Integration Service does not run jobs on the node. In
this case, you can disable the compute role on the node. However, because the container management
function of the Service Manager is a lightweight process, enabling or disabling the compute role does not
have a performance impact.
When you disable the compute role on a node, you must specify whether to stop, complete, or abort
computations that might be running on the node.
Enable only the service role to dedicate the node to running the Data Integration Service process. Enable only
the compute role to dedicate the node to running Data Integration Service mappings.
Note: Before you can disable the service role on a node, you must shut down all application service
processes running on the node and remove the node as a primary or back-up node for any application
service. You cannot disable the service role on a gateway node.
1. In the Administrator tool, click the Manage tab > Services and Nodes view.
2. In the Domain Navigator, select a node assigned to a Data Integration Service grid.
3. In the Properties view, click Edit for the general properties.
The Edit General Properties dialog box appears.
4. Select or clear the service and compute roles to update the node role.
5. Click OK.
6. If you disabled the compute role, the Disable Compute Role dialog box appears. Perform the following
steps:
a. Select one of the following modes to disable the compute role:
• Complete. Allows jobs to run to completion before disabling the role.
• Stop. Stops all jobs and then disables the role.
• Abort. Tries to stop all jobs before aborting them and disabling the role.
b. Click OK.
7. If you updated the role on a node assigned to a Data Integration Service or a Data Integration Service
grid, recycle the Data Integration Service.
94 Chapter 6: Nodes
Viewing Processes on a Node with the Service Role
You can view the status of all application service processes configured to run on a node with the service role.
Before you shut down or remove a node, you can view the status of each application service process to
determine which service processes you need to disable.
When a node does not have the service role, no application service processes run on the node.
1. In the Administrator tool, click the Manage tab > Services and Nodes view.
2. In the Domain Navigator, select a node with the service role.
3. In the contents panel, select the Processes view.
The view displays the status of each application service process configured to run on the node.
Informatica installer
Run the infasetup DefineGatewayNode or infasetup DefineWorkerNode command on each machine that
you want to define as a node. You might use infasetup to define a node if you decide to move a node
from one domain to another domain.
When the Informatica installer or infasetup defines a node, the program creates nodemeta.xml. This file is the
node configuration file for the node. A gateway node uses information in nodemeta.xml to connect to the
domain configuration database. A worker node uses the information in nodemeta.xml to connect to the
domain. The file is stored in the following directory on each node:
When you define a node using the Informatica installer, the installer adds the node to the domain with both
the service and compute roles enabled. When you log in to the Administrator tool, the node appears in the
Navigator.
When you define a node with infasetup, you must manually add the node to the domain. You can add a node
to the domain in the Administrator tool or with the infacmd isp AddDomainNode command. When you add
the node, you specify the roles to enable on the node.
You can use the Administrator tool to add a node to the domain before you define the node. In this case, the
Administrator tool displays a message saying that you need to run the Informatica installer to associate the
node with a physical host name and port number. The name that you enter for the node must be the same
name that you use when you define the node.
Use the Administrator tool to add a node to the domain in the following situations:
1. In the Administrator tool, click the Manage tab > Services and Nodes view.
2. In the Domain Navigator, select the folder where you want to add the node. If you do not want the node
to appear in a folder, select the domain.
3. On the Navigator Actions menu, click New > Node.
The Create Node dialog box appears.
4. Enter the node name.
The name must be the same node name that you use when you define the node.
5. If you want to change the folder for the node, click Browse and choose a new folder or the domain.
6. Optionally update the node role.
By default, each node has both the service and compute roles. If a node is assigned to a Data Integration
Service grid, you might want to update the node role to dedicate the node to running the Data Integration
Service process or to running mappings.
7. Click OK.
If you add a node to the domain before you define the node using the Informatica installer or infasetup,
the Administrator tool displays a message saying that you need to run the installer to associate the node
with a physical host name and port number.
Related Topics:
• “Node Roles” on page 93
1. In the Administrator tool, click the Manage tab > Services and Nodes view.
2. In the Domain Navigator, select a node.
3. Click the Properties view.
The Properties view displays the node properties in separate sections.
4. In the Properties view, click Edit for the section that contains the property you want to set.
5. Edit the following properties:
Name Name of the node. The name is not case sensitive and must be unique within the domain. It
cannot exceed 128 characters or begin with @. It also cannot contain spaces or the following
special characters:
`~%^*+={}\;:'"/?.,<>|!()][
Description Description of the node. The description cannot exceed 765 characters.
96 Chapter 6: Nodes
Node Property Description
Gateway Node Indicates whether the node can serve as a gateway. If this property is disabled, then the node
is a worker node.
Service Role Indicates whether the node has the service role. If enabled, application services can run on the
node. If disabled, application services cannot run on the node. Disable the property only if the
node is assigned to a Data Integration Service grid and you want to dedicate the node to
running mappings.
Default is enabled.
Compute Role Indicates whether the node has the compute role. If enabled, the node can perform
computations. If disabled, the node cannot perform computations.
A node requires the compute role when the Data Integration Service runs jobs on the node. If
the Data Integration Service does not run jobs on the node, you can disable the compute role.
However, enabling or disabling the compute role does not have a performance impact.
Default is enabled.
Backup Directory to store repository backup files. The directory must be accessible by the node.
Directory
Error Severity Level of error logging for the node. These messages are written to the Log Manager
Level application service and Service Manager log files. Set one of the following message levels:
- ERROR. Writes ERROR code messages to the log.
- WARNING. Writes WARNING and ERROR code messages to the log.
- INFO. Writes INFO, WARNING, and ERROR code messages to the log.
- TRACING. Writes TRACE, INFO, WARNING, and ERROR code messages to the log.
- DEBUG. Writes DEBUG, TRACE, INFO, WARNING, and ERROR code messages to the log.
Default is WARNING.
Minimum Port Minimum port number used by service processes on the node. To apply changes, restart
Number Informatica services. The default value is the value entered when the node was defined.
Maximum Port Maximum port number used by service processes on the node. To apply changes, restart
Number Informatica services. The default value is the value entered when the node was defined.
CPU Profile Ranks the node's CPU performance against a baseline system. Used by the Load Balancer
Benchmark component of the PowerCenter Integration Service.
For example, if the CPU is running 1.5 times as fast as the baseline machine, the value of this
property is 1.5. You can calculate the benchmark by clicking Actions > Recalculate CPU
Profile Benchmark. The calculation takes approximately five minutes and uses 100% of one
CPU on the machine. Or, you can update the value manually.
Default is 1.0. Minimum is 0.001. Maximum is 1,000,000.
Used in adaptive dispatch mode. Ignored in round-robin and metric-based dispatch modes.
Maximum Maximum number of running session tasks or command tasks allowed for each PowerCenter
Processes Integration Service process running on the node. Used by the Load Balancer component of the
PowerCenter Integration Service.
For example, if you set the value to 5, up to 5 command tasks and 5 session tasks can run at
the same time.
Set this threshold to a high number, such as 200, to cause the Load Balancer to ignore it. To
prevent the Load Balancer from dispatching tasks to this node, set this threshold to 0.
Default is 10. Minimum is 0. Maximum is 1,000,000,000.
Used in all dispatch modes.
Maximum CPU Maximum number of runnable threads waiting for CPU resources on the node. Used by the
Run Queue Load Balancer component of the PowerCenter Integration Service.
Length Set this threshold to a low number to preserve computing resources for other applications. Set
this threshold to a high value, such as 200, to cause the Load Balancer to ignore it.
Default is 10. Minimum is 0. Maximum is 1,000,000,000.
Used in metric-based and adaptive dispatch modes. Ignored in round-robin dispatch mode.
Maximum Maximum percentage of virtual memory allocated on the node relative to the total physical
Memory % memory size. Used by the Load Balancer component of the PowerCenter Integration Service.
Set this threshold to a value greater than 100% to allow the allocation of virtual memory to
exceed the physical memory size when dispatching tasks. Set this threshold to a high value,
such as 1,000, if you want the Load Balancer to ignore it.
Default is 150. Minimum is 0. Maximum is 1,000,000,000.
Used in metric-based and adaptive dispatch modes. Ignored in round-robin dispatch mode.
Log Collection The directory that stores the logs for the application service when you run the log aggregator.
Directory The directory must be accessible from all the nodes in the domain. If the log collection
directory is not accessible by other nodes, the aggregated logs do not appear in the
aggregated logs listgrid. The users who run node processes must have read-write permissions
on the directory.
Configure the log collection directory for the master gateway node in the domain.
Core Dump The directory that stores the core dump files for the domain processes when you run the log
Directory aggregator.
Configure the core dump directory for all the nodes in the domain.
6. Click OK.
You can shut down a node from the Administrator tool or from the operating system. When you shut down a
node, you stop Informatica services and abort all application service processes and computations running on
the node.
98 Chapter 6: Nodes
Warning: To avoid loss of data or metadata when you shut down a node, disable all running application
service processes in complete mode.
1. In the Administrator tool, click the Manage tab > Services and Nodes view.
2. In the Domain Navigator, select a node.
3. In the Navigator Actions menu, select Shutdown Node.
If the node has the service role, the Administrator tool displays the list of application service processes
running on that node.
4. Optionally, choose whether the shutdown is planned or unplanned.
5. Optionally, enter comments about the shutdown.
6. Click OK to stop all service processes and shut down the node, or click Cancel to cancel the operation.
To associate a different host machine with the node, you must run the installation program or infasetup
DefineGatewayNode or infasetup DefineWorkerNode command on the new host machine, and then restart
the node on the new host machine.
1. In the Administrator tool, click the Manage tab > Services and Nodes view.
2. In the Navigator, select a node.
3. In the Services and Nodes view Actions menu, select Remove Node Association.
Removing a Node
When you remove a node from a domain, it is no longer visible in the Navigator. If the node is running when
you remove it, the node shuts down and aborts all application service processes.
Note: To avoid loss of data or metadata when you remove a node, disable all running application service
processes in complete mode.
1. In the Administrator tool, click the Manage tab > Services and Nodes view.
2. In the Domain Navigator, select a node.
3. In the Navigator Actions menu, select Delete.
4. In the warning message that appears, click OK.
High Availability
This chapter includes the following topics:
The following high availability components make services highly available in an Informatica domain:
• Resilience. An Informatica domain can tolerate temporary connection failures until either the resilience
timeout expires or the failure is fixed.
• Restart and failover. A process can restart on the same node or on a backup node after the process
becomes unavailable.
• Recovery. Operations can complete after a service is interrupted. After a service process restarts or fails
over, it restores the service state and recovers operations.
When you plan a highly available Informatica environment, configure high availability for both the internal
Informatica components and systems that are external to Informatica. Internal components include the
domain, application services, application clients, and command line programs. External systems include the
network, hardware, database management systems, FTP servers, message queues, and shared storage.
High availability features for the Informatica environment are available based on your license.
Example
As you open a mapping in the PowerCenter Designer workspace, the PowerCenter Repository Service
becomes unavailable and the request fails. The domain contains multiple nodes for failover and the
PowerCenter Designer is resilient to temporary failures.
101
The PowerCenter Designer tries to establish a connection to the PowerCenter Repository Service within the
resilience timeout period. The PowerCenter Repository Service fails over to another node because it cannot
restart on the same node.
The PowerCenter Repository Service restarts within the resilience timeout period, and the PowerCenter
Designer reestablishes the connection.
After the PowerCenter Designer reestablishes the connection, the PowerCenter Repository Service recovers
from the failed operation and fetches the mapping into the PowerCenter Designer workspace.
Resilience
The domain tolerates temporary connection failures between application clients, application services, and
nodes.
A temporary connection failure might occur because an application service process fails or because of a
network failure. When a temporary connection failure occurs, the Service Manager tries to reestablish
connections between the application clients, application services, and nodes.
Based on your license, the following application clients are resilient to the services that they connect to:
Developer Tool Client
The Developer tool client tries to reconnect to the Data Integration Service or the Data Integration
Service grid when a temporary network failure occurs.
If a job is running and the Developer tool cannot reconnect to the Data Integration Service or the Data
Integration Service grid within the re-connection timeout period, the Developer tool does not resubmit the
job to a Data Integration Service or a Data Integration Service grid on a different node. The Developer
tool client fails the job.
PowerCenter Client
The PowerCenter Client tries to reconnect to the PowerCenter Repository Service and the PowerCenter
Integration Service when a temporary network failure occurs.
If you perform a PowerCenter Client action that requires connection to the repository while the
PowerCenter Client is trying to reestablish the connection, the PowerCenter Client prompts you to try the
operation again after the PowerCenter Client reestablishes the connection. If the PowerCenter Client is
unable to reestablish the connection during the resilience timeout period, the PowerCenter Client
prompts you to reconnect to the repository manually.
Command line programs try to reconnect to the domain or an application service when a temporary
network failure occurs while a command line program is running.
If the command line program is running on a Data Integration Service or a Data Integration Service grid
and the command line program cannot reconnect to the Data Integration Service or the Data Integration
Service grid within the re-connection timeout period, the command line program does not resubmit the
job to a Data Integration Service or a Data Integration Service grid on a different node. The command
line program fails the command.
The Developer does not notice the loss of connection and he is unaffected by the 120 seconds connection
loss. However, the following messages appear in the Notifications tab on the PowerCenter Workflow
Monitor:
Repository Service notifications are enabled.
DATE TIME-[REP_55101] Connection to the Repository Service [Repository_Service_Name] is
broken.
DATE TIME-[REP_55114] Reconnecting to the Repository Service [Repository_Service_Name].
The resilience time is 180 seconds.
DATE TIME-Reconnected to Repository Service [Repository_Service_Name] successfully.
Based on your license, the following application services are resilient to the temporary connection failure of
their clients:
Data Integration Service
The Data Integration Service is resilient to temporary connection failures to other services and the
Developer tool client.
The PowerCenter Integration Service is resilient to temporary connection failures to other services, the
PowerCenter client, and external components such databases and FTP servers.
The PowerCenter Repository Service is resilient to temporary connection failures to other services, such
as the PowerCenter Integration Service. It is also resilient to the temporary connection failures to the
repository database.
Node Resilience
When a domain contains multiple nodes, the nodes are resilient to temporary failures in communication from
other nodes in the domain.
Every node in the domain sends a communication signal to the master gateway node at periodic
intervals of 15 seconds. For nodes with the service role, the communication includes a list of application
services running on the node.
All nodes have a resilience timeout of 90 seconds. If a node fails to connect to the master gateway node
within the resilience timeout period, the master gateway node marks the node unavailable. If the node
that fails to connect has the service role, the master gateway node also reassigns its application
services to a back-up node. This ensures that services on a node continue to run despite node failures.
Resilience 103
The master gateway node becomes unavailable.
You can configure more than one node to serve as a gateway. If the master gateway node becomes
unavailable, the Service Managers on the other gateway nodes elect another master gateway node.
If you configure one node to serve as the gateway and the node becomes unavailable, all other nodes
shut down.
You can use the resilience timeout and limit on resilience timeout configured for the domain for PowerCenter
application services if you do not set it for the application service. Command line programs use the service
resilience timeout. If the service limit on resilience timeout is smaller than the resilience timeout for the
connecting client, the client uses the services limit as the resilience timeout.
The following figure shows some sample connections and resilience configurations in a domain with
PowerCenter application services:
The following table describes the resilience timeout and the limits shown in the figure above:
The Service Manager on the master gateway node accepts application service request and manages the
domain. If a master gateway node is not available, the domain shuts down. Configure the domain to failover
to another node by configuring multiple gateway nodes.
Based on your license, you can also configure backup nodes for application services. The Service Manager
can restart or failover the following application services if a failure occurs:
Domain Failover
The Service Manager on the master gateway node accepts service requests and manages the domain and
services in the domain. The domain can failover to another node when the domain has multiple gateway
nodes. Configure multiple gateway nodes to prevent domain shutdown when the master gateway node is
unavailable.
The master gateway node maintains a connection to the domain configuration repository. If the domain
configuration repository becomes unavailable, the master gateway node tries to reconnect when a user
performs an operation. If the master gateway node cannot connect to the domain configuration repository,
the master gateway node may shut down.
If the domain has multiple gateway nodes and the master gateway node becomes unavailable, the Service
Managers on the other gateway nodes elect another master gateway node. The domain tries to connect to
the domain configuration repository with each gateway node. If none of the gateway nodes can connect, the
domain shuts down and all domain operations fail. When a master gateway fails over, the client tools retrieve
information about the alternate domain gateways from the domains.infa file.
Note: Application services running on the master gateway node will not fail over when another master
gateway node is elected unless the application service has a backup node configured.
The following situations describe how the Service Manager restarts or fails over an application service:
• If the primary node running the service process becomes unavailable, the service fails over to a back-up
node. The primary node might be unavailable if it shuts down or if the connection to the node becomes
unavailable.
• If the primary node running the service process is available, the domain tries to restart the process based
on the restart options configured in the domain properties. If the process does not restart, the Service
Manager may mark the process as failed. The service then fails over to a back-up node and starts another
process. If the Service Manager marks the process as failed, the administrator must enable the process
after addressing any configuration problem.
If a service process fails over to a back-up node, it does not fail back to the primary node when the node
becomes available. You can disable the service process on the back-up node to cause it to fail back to the
primary node.
The following image shows how you can configure primary and back-up nodes for an application service:
Recovery
Recovery is the completion of operations after an interrupted service is restored. The state of operation for a
service contains information about the service process.
Based on your license, the following components can recover after an interrupted service is restored:
The Service Manager for each node in the domain maintains the state of service processes running on
that node. If the master gateway shuts down, the newly elected master gateway collects the state
information from each node to restore the state of the domain.
The PowerCenter Repository Service maintains the state of operation in the PowerCenter repository. The
state of operation includes information about repository locks, requests in progress, and connected
clients. After restart or failover, the PowerCenter Repository Service can recover operations from the
point of interruption.
The PowerCenter Integration Service maintains the state of operation in the shared storage configured
for the service. The state of operation includes information about scheduled, running, and completed
tasks for the service.
The PowerCenter Integration Service maintains PowerCenter session and workflow state of operation
based on the recovery strategy you configure for the session and workflow. When the PowerCenter
Integration Service restarts or fails over a service process, it can automatically recover interrupted
workflows that are configured for recovery.
The Data Integration Service maintains the state of operation in the Model repository. The state of
operation includes the state of the workflow and workflow tasks and the values of the workflow
variables and parameters during the interrupted workflow instance.
When a Data Integration Service process restarts or fails over a service process, you can manually
restart interrupted workflows that are enabled for workflow recovery.
You can configure the following Informatica domain components to be highly available:
Domain
One node in the domain acts as a gateway to receive service requests from clients and routes them to
the appropriate service and node. To prevent domain shutdown when the master gateway node is
unavailable, configure more than one gateway node.
Nodes
Informatica services are processes that run on each node. You can configure Informatica services to
restart automatically if it terminates unexpectedly.
Application Services
Based on your license, you can configure the following high availability features for application services:
• To minimize the application service downtime, configure backup nodes for application services.
• To specify the resilience period for application services, review default settings and configure
resilience timeout periods for application services.
Application Clients
Application clients provide access to Informatica functionality, and they run on user machines.
Application clients send requests to the Service Manager or application services.
You can configure resilience timeout periods for command line programs. You cannot configure a
PowerCenter client resilience timeout.
External Systems
Use highly available versions of external systems such as source and target databases, message
queues, and FTP servers.
Network
Make the network highly available by configuring redundant components such as routers, cables, and
network adapters.
When an application service connects to another application service in the domain, the service that initiates
the connection is a client of the other service.
You can configure application service resilience timeouts for the following application services:
PowerCenter Application Services
You can configure the resilience timeout and resilience timeout limits in the advanced properties of the
PowerCenter Integration Service and PowerCenter Repository Service. The resilience timeout for
application services that connects to a PowerCenter Integration Service or PowerCenter Repository
Service is determined by one of the following values:
• The service Resilience Timeout property. You can configure the resilience timeout for the service in
the service properties. To disable resilience for a service, set the resilience timeout to 0.
• The domain Resilience Timeout property. To use the resilience timeout configured for the domain,
set the resilience timeout for the service to blank.
• The service Limit on Resilience Timeout property. If the service limit on resilience timeout is smaller
than the resilience timeout for the connecting client, the client uses the limit as the resilience
timeout. To use the limit on resilience timeout configured for the domain, set the service resilience
limit to blank.
• The domain Limit on Resilience Timeout property. To use the resilience timeout configured for the
domain, set the limit on resilience timeout for the service to blank.
You can configure the resilience timeout for the SAP BW Service in the general properties for the service.
The SAP BW Service resilience timeout property is called the Retry Period.
Note: A client cannot be resilient to service interruptions if you disable the service in the Administrator tool. If
you disable the service process, the client is resilient to the interruption in service.
When you configure a backup node, verify that the node has access to run-time files that each application
service requires to process data integration tasks such as workflows and mappings. For example, a workflow
might require parameter files, input files, or output files.
The state of operation files store the state of each workflow and session operation. The PowerCenter
Integration Service always stores the state of each workflow and session operation in files in the
$PMStorageDir directory of the PowerCenter Integration Service process.
Process state information includes information about which node was running the master PowerCenter
Integration Service process and which node was running each session. You can configure the PowerCenter
Integration Service to store process state information on a cluster file system or in the PowerCenter
repository database.
Nodes that run the PowerCenter Integration Service must be on the same cluster file system so that they can
share resources. Also, nodes within a cluster must be on the cluster file system’s heartbeat network. Use a
highly available cluster file system that is configured for I/O fencing. The hardware requirements and
configuration of an I/O fencing solution are different for each file system.
The following cluster file systems are certified by Informatica for use for PowerCenter Integration Service
failover and session recovery:
Storage Array Network
Contact the file system vendors directly to evaluate which file system matches your requirements.
During failover, automatic recovery of workflows resume when the service process can access the database
tables.
When you use the infacmd , pmcmd, or pmrep command line programs to connect to the domain or an
application service the resilience timeout is determined by the command line option, an environment variable,
or the default resilience timeout.
Use the following guidelines when you configure command line program resilience:
You can set the resilience timeout for infacmd by using the -ResilienceTimeout command line option
each time you run a command. You can set the resilience timeout for pmcmd by using the -timeout
command line option each time you run a command. When you use pmrep connect to connect to a
repository, you can use the -t command line option to set the resilience timeout for pmrep commands
that use the connection.
Environment variable.
If you do not set the timeout option in the infacmd and pmcmd command line syntax, the infacmd and
pmcmd command line programs use the value of the environment variable
INFA_CLIENT_RESILIENCE_TIMEOUT that is configured on the client machine. If you do not set the
timeout option when you use pmrep connect to connect to the repository, pmrep commands use the
value of the environment variable INFA_CLIENT_RESILIENCE_TIMEOUT that is configured on the client
machine.
Default value
If you do not use the command line option or the environment variable, the pmcmd and pmrep command
line program uses the default resilience timeout of 180 seconds. If you do not use the command line
option or the environment variable, the infacmd command line program uses the value of the domain
Service Level Timeout property as the default resilience timeout.
Limit on timeout
If the limit on resilience timeout for the PowerCenter Integration Service or the PowerCenter Repository
Service is smaller than the command line resilience timeout, the command line program uses the limit as
the resilience timeout.
Note: PowerCenter does not provide resilience for a repository client when the PowerCenter Repository
Service is running in exclusive mode.
The first time that you install Informatica services, you create one gateway node. After you install
Informatica, you can define additional gateway nodes. To define a gateway node, add a gateway node to the
domain or configure a worker node to serve as a gateway node.
To restart the Informatica services when a node restarts, complete the following steps:
• In a UNIX environment, you can create a script to automatically start the Informatica services when the
node starts.
• In a Windows environment, go to the Control Panel and configure the Informatica services to start
automatically.
You can configure restart for all nodes, regardless of node type or node role.
The following operations in the PowerCenter Repository Service are resilient to the database failover in
Oracle RAC setup:
• ExecuteQuery
• ObjectExport
• ObjectImport
• PurgeVersion
• RollbackDeployment
You cannot perform any administrator operations with Oracle RAC database failover for Informatica domain
and Model Repository Service.
I am not sure where to look for status information regarding client connections to the
PowerCenter repository.
In PowerCenter Client applications such as the PowerCenter Designer and the PowerCenter Workflow
Manager, an error message appears if the connection cannot be established during the timeout period.
Detailed information about the connection failure appears in the Output window. If you are using pmrep, the
connection error information appears at the command line. If the PowerCenter Integration Service cannot
establish a connection to the repository, the error appears in the PowerCenter Integration Service log, the
workflow log, and the session log.
I have the high availability option, but my FTP server is not resilient when the network connection
fails.
The FTP server is an external system. To achieve high availability for FTP transmissions, you must use a
highly available FTP server. For example, Microsoft IIS 6.0 does not natively support the restart of file
uploads or file downloads. File restarts must be managed by the client connecting to the IIS server. If the
transfer of a file to or from the IIS 6.0 server is interrupted and then reestablished within the client resilience
timeout period, the transfer does not necessarily continue as expected. If the write process is more than half
complete, the target file may be rejected.
I have the high availability option, but the Informatica domain is not resilient when machines are
connected through a network switch.
If you are using a network switch to connect machines in the domain, use the auto-select option for the
switch.
Connections
This chapter includes the following topics:
Connections Overview
A connection is a repository object that defines a connection in the domain configuration repository.
The Data Integration Service uses database connections to process jobs for the Developer tool and the
Analyst tool. Jobs include mappings, data profiles, scorecards, and SQL data services.
You can create and manage connections in the Administrator tool, the Developer tool, and the Analyst tool.
The tasks that you can perform in each tool depend on the tool that you use. For example, you can create an
SAP NetWeaver connection in the Developer tool and manage it in the Administrator tool, but you cannot
create or manage it in the Analyst tool.
Note: These connections are independent of the connections that you create in the PowerCenter Workflow
Manager.
Connection Management
After you create a connection, you can view the connection, configure connection properties, and delete the
connection.
After you create a connection, you can perform the following actions on the connection:
Configure connection pooling to optimize processing for the Data Integration Service. Connection
pooling is a framework to cache connections.
View the connection properties through the Connections view on the Manage tab.
113
Edit the connection.
You can change the connection name and the description. You can also edit connection details such as
the user name, password, and connection strings. When you update a database connection that has
connection pooling disabled, all updates take effect immediately.
The Data Integration Service identifies connections by the connection ID instead of the connection
name. When you rename a connection, the Developer tool and the Analyst tool update the jobs that use
the connection.
Deployed applications and parameter files identify a connection by name, not by connection ID.
Therefore, when you rename a connection, you must redeploy all applications that use the connection.
You must also update all parameter files that use the connection parameter.
When you delete a connection, objects that use the connection are no longer valid. If you accidentally
delete a connection, you can re-create it by creating another connection with the same connection ID as
the deleted connection.
You can refresh the connections list to see the latest list of connections for the domain. Refresh the
connections list after a user adds, deletes, or renames a connection in the Developer tool or the Analyst
tool.
Creating a Connection
In the Administrator tool, you can create relational database, social media, and file systems connections.
8. Click Finish.
The Administrator tool displays the latest list of connections when you start the Administrator tool. You
might want to refresh the connections list when a user adds, deletes, or renames a connection in the
Developer tool or the Analyst tool.
Viewing a Connection
View connections in the Administrator tool.
Deleting a Connection
You can delete a database connection in the Administrator tool.
When you delete a connection in the Administrator tool, you also delete it from the Developer tool and the
Analyst tool.
Pass-through Security
Pass-through security is the capability to connect to an SQL data service or an external source with the client
user credentials instead of the credentials from a connection object.
Users might have access to different sets of data based on the job in the organization. Client systems restrict
access to databases by the user name and the password. When you create an SQL data service, you might
combine data from different systems to create one view of the data. However, when you define the
connection to the SQL data service, the connection has one user name and password.
If you configure pass-through security, you can restrict users from some of the data in an SQL data service
based on their user name. When a user connects to the SQL data service, the Data Integration Service ignores
the user name and the password in the connection object. The user connects with the client user name or the
LDAP user name.
A web service operation mapping might need to use a connection object to access data. If you configure
pass-through security and the web service uses WS-Security, the web service operation mapping connects to
a source using the user name and password provided in the web service SOAP request.
Example
An organization combines employee data from multiple databases to present a single view of employee data
in an SQL data service. The SQL data service contains data from the Employee and Compensation databases.
The Employee database contains name, address, and department information. The Compensation database
contains salary and stock option information.
A user might have access to the Employee database but not the Compensation database. When the user runs
a query against the SQL data service, the Data Integration Service replaces the credentials in each database
connection with the user name and the user password. The query fails if the user includes salary information
from the Compensation database.
When you deploy an SQL data service or a web service, you can choose to cache the logical data objects in a
database. You must specify the database in which to store the data object cache. The Data Integration
Service validates the user credentials for access to the cache database. If a user can connect to the cache
database, the user has access to all tables in the cache. The Data Integration Service does not validate user
credentials against the source databases when caching is enabled.
For example, you configure caching for the EmployeeSQLDS SQL data service and enable pass-through
security for connections. The Data Integration Service caches tables from the Compensation and the
Employee databases. A user might not have access to the Compensation database. However, if the user has
access to the cache database, the user can select compensation data in an SQL query.
When you configure pass-through security, the default is to disallow data object caching for data objects that
depend on pass-through connections. When you enable data object caching with pass-through security, verify
that you do not allow unauthorized users access to some of the data in the cache. When you enable caching
for pass-through security connections, you enable data object caching for all pass-through security
connections.
1. Select a connection.
2. Click the Properties view.
3. Edit the connection properties.
The Edit Connection Properties dialog box appears.
4. To choose pass-through security for the connection, select the Pass-through Security Enabled option.
5. Optionally, select the Data Integration Service for which you want to enable object caching for pass-
through security.
6. Click the Properties view.
7. Edit the pass-through security options.
The number of connection pool libraries depends on the number of running Data Integration Service
processes or DTM processes. Each Data Integration Service process or DTM process maintains its own
connection pool library. The values of the pooling properties are for each connection pool library.
For example, if you set maximum connections to 15, then each connection pool library can have a maximum
of 15 idle connections in the pool. If the Data Integration Service runs jobs in separate local processes and
three DTM processes are running, then you can have a maximum of 45 idle connection instances.
To decrease the total number of idle connection instances, set the minimum number of connections to 0 and
decrease the maximum idle time for each database connection.
The following list describes database connection pooling properties that you can edit in the Pooling view for
a database connection:
Enables connection pooling. When you enable connection pooling, each connection pool retains idle
connection instances in memory. To delete the pools of idle connections, you must restart the Data
Integration Service.
If connection pooling is disabled, the DTM process or the Data Integration Service process stops all
pooling activity. The DTM process or the Data Integration Service process creates a connection instance
each time it processes a job. It drops the instance when it finishes processing the job.
Default is enabled for DB2 for i5/OS, DB2 for z/OS, IBM DB2, Microsoft SQL Server, Oracle, and ODBC
connections. Default is disabled for Adabas, IMS, Sequential, and VSAM connections.
Minimum # of Connections
The minimum number of idle connection instances that a pool maintains for a database connection after
the maximum idle time is met. Set this value to be equal to or less than the maximum number of idle
connection instances. Default is 0.
Maximum # of Connections
The maximum number of idle connection instances that a pool maintains for a database connection
before the maximum idle time is met. Set this value to be more than the minimum number of idle
connection instances. Default is 15.
The number of seconds that a connection instance that exceeds the minimum number of connection
instances can remain idle before the connection pool drops it. The connection pool ignores the idle time
when the connection instance does not exceed the minimum number of idle connection instances.
Default is 120.
Connection Properties
This chapter includes the following topics:
119
• Microsoft Azure Blob Storage Connection Properties, 179
• Microsoft Azure Cosmos DB SQL API Connection Properties, 180
• Microsoft Azure Data Lake Storage Gen1 Connection Properties, 181
• Microsoft Azure Data Lake Storage Gen2 Connection Properties, 182
• Microsoft Azure SQL Data Warehouse Connection Properties, 183
• MS SQL Server Connection Properties, 185
• Netezza Connection Properties, 189
• OData Connection Properties, 190
• ODBC Connection Properties, 191
• Oracle Connection Properties, 192
• Salesforce Connection Properties, 195
• Salesforce Marketing Cloud Connection Properties, 196
• SAP Connection Properties, 197
• Sequential Connection Properties, 200
• Snowflake Connection Properties, 202
• Teradata Parallel Transporter Connection Properties, 203
• Tableau Connection Properties, 205
• Tableau V3 Connection Properties, 206
• Twitter Streaming Connection Properties, 207
• VSAM Connection Properties, 208
• Web Services Connection Properties, 211
• Identifier Properties in Database Connections, 212
This chapter contains connection properties for each of the connections you can create and manage through
Informatica clients.
Option Description
Location Node name for the location of the PowerExchange Listener that connects to Adabas. The
node name is defined in the first parameter of the NODE statement in the PowerExchange
dbmover.cfg configuration file.
Password Password for the database user name or a valid PowerExchange passphrase.
A PowerExchange passphrase can be from 9 to 128 characters in length and can contain the
following characters:
- Uppercase and lowercase letters
- The numbers 0 to 9
- Spaces
- The following special characters:
’ - ; # \ , . / ! % & * ( ) _ + { } : @ | < > ?
Note: The first character is an apostrophe.
Passphrases cannot include single quotation marks (‘), double quotation marks (“), or
currency symbols.
To use passphrases, ensure that the PowerExchange Listener runs with a security setting of
SECURITY=(1,N) or higher in the DBMOVER member. For more information, see "SECURITY
Statement" in the PowerExchange Reference Manual.
The allowable characters in the IBM IRRPHREX exit do not affect the allowable characters in
PowerExchange passphrases.
Note: A valid RACF passphrase can be up to 100 characters in length. PowerExchange
truncates passphrases longer than 100 characters when passing them to RACF for validation.
Code Page Required. Name of the code page to use for reading from or writing to the data source.
Usually, this value is an ISO code page name, such as ISO-8859-6.
Pass-through Enables pass-through security for the connection. When you enable pass-through security for
security enabled a connection, the domain uses the client user name and password to log into the
corresponding database, instead of the credentials defined in the connection object.
Encryption Type The type of encryption that the Data Integration Service uses. Select one of the following
options:
- None
- AES
Default is None.
Note: Informatica recommends that you use Secure Sockets Layer (SSL) authentication
instead of configuring the Encryption Type and Level connection properties. SSL
authentication provides stricter security and is used by several Informatica products.
For more information about implementing SSL authentication in a PowerExchange network,
see the PowerExchange Reference Manual.
[Encryption] Level If you select AES for Encryption Type, select one of the following options to indicate the
encryption level that the Data Integration Service uses:
- 1. Use a 128-bit encryption key.
- 2. Use a 192-bit encryption key.
- 3. Use a 256-bit encryption key.
If you do not select AES for Encryption Type, this option is ignored.
Default is 1.
Pacing size Optional. Amount of data that the source system can pass to the PowerExchange Listener.
Set the pacing size if an external application, database, or the Data Integration Service node
is a bottleneck. User lower values for faster performance.
The minimum value and default value is 0. A value of 0 provides the best performance.
Interpret as rows Optional. Select this option to express the pacing size as a number of rows. Clear this option
to express the pacing size in kilobytes. By default, this option is not selected and the pacing
size is in kilobytes.
Compression Optional. Select this option to enable source data compression. By compressing data, you
can decrease the amount of data that Informatica applications send over the network. By
default, this option is not selected and compression is disabled.
Offload processing Optional. Controls whether to offload some bulk data processing from the source machine to
the Data Integration Service machine. Select one of the following options:
- AUTO. The Data Integration Service determines whether to use offload processing.
- Yes. Use offload processing.
- No. Do not use offload processing.
Default is AUTO.
Worker threads Optional. Number of threads that the Data Integration Service uses to process bulk data when
offload processing is enabled. For optimal performance, this value should not exceed the
number of available processors on the Data Integration Service machine. Valid values are 1
through 64. Default is 0, which disables multithreading.
Array size Optional. The number of records in the storage array for the worker threads. This option is
applicable when you set the Worker Threads option to a value greater than 0. Valid values are
1 to 5000. Default is 25.
Write mode Optional. Mode in which Data Integration Service sends data to the PowerExchange Listener.
Select one of the following write modes:
- CONFIRMWRITEON. Sends data to the PowerExchange Listener and waits for a response
before sending more data. Select this option when error recovery is a priority. However, this
option might degrade performance.
- CONFIRMWRITEOFF. Sends data to the PowerExchange Listener without waiting for a
response. Use this option if you can reload the target table when an error occurs.
- ASYNCHRONOUSWITHFAULTTOLERANCE. Sends data to the PowerExchange Listener
without waiting for a response. This option also enables error detection. This option
combines the speed of CONFIRMWRITEOFF and the data integrity of CONFIRMWRITEON.
Default is CONFIRMWRITEON.
Property Description
Name The name of the connection. The name is not case sensitive and must be unique within the domain. You
can change this property after you create the connection. The name cannot exceed 128 characters,
contain spaces, or contain the following special characters:~ ` ! $ % ^ & * ( ) - + = { [ } ] | \ : ; " ' < , > . ? /
ID String that the Data Integration Service uses to identify the connection. The ID is not case sensitive. It
must be 255 characters or less and must be unique in the domain. You cannot change this property
after you create the connection. Default value is the connection name.
Description The description of the connection. The description cannot exceed 4,000 characters.
The Details tab contains the connection attributes of the Amazon Redshift connection. The following table
describes the connection attributes:
Property Description
Authentication Type The authentication method to log in to Amazon Redshift. You can select one of the following
authentication methods:
- Default. Uses the username and password to connect to Amazon Redshift.
- Redshift IAM Authentication via AssumeRole. Uses AssumeRole for IAM authentication to
connect to Amazon Redshift.
Access Key ID Amazon S3 bucket access key ID to stage data in the S3 bucket using unload and copy
commands.
Secret Access Key Amazon S3 bucket secret access key ID to stage the data in the S3 bucket.
Master Symmetric Optional. Provide a 256-bit AES encryption key in the Base64 format when you enable client-
Key side encryption. You can generate a key using a third-party tool.
If you specify a value, ensure that you specify the encryption type as client-side encryption in
the advanced target properties.
Cluster Region Optional. The AWS cluster region in which the bucket you want to access resides.
Select a cluster region if you choose to provide a custom JDBC URL that does not contain a
cluster region name in the JDBC URL connection property.
If you specify a cluster region in both Cluster Region and JDBC URL connection properties,
the Data Integration Service ignores the cluster region that you specify in the JDBC URL
connection property.
To use the cluster region name that you specify in the JDBC URL connection property, select
None as the cluster region in this property.
Select one of the following cluster regions:
- Asia Pacific (Mumbai)
- Asia Pacific (Seoul)
- Asia Pacific (Singapore)
- Asia Pacific (Sydney)
- Asia Pacific (Tokyo)
- AWS GovCloud (US)
- Canada (Central)
- China (Beijing)
- China (Ningxia)
- EU (Ireland)
- EU (Frankfurt)
- EU (London)
- EU (Paris)
- South America (Sao Paulo)
- US East (Ohio)
- US East (N. Virginia)
- US West (N. California)
- US West (Oregon)
Default is None.
You can only read data from or write data to the cluster regions supported by AWS SDK used
by PowerExchange for Amazon Redshift.
Customer Master Key Optional. Specify the customer master key ID generated by AWS Key Management Service
ID (AWS KMS) or the Amazon Resource Name (ARN) of your custom key for cross-account
access. You must generate the customer master key corresponding to the region where
Amazon S3 bucket resides. You can specify any of the following values:
Customer generated customer master key
Enables client-side or server-side encryption. Only the administrator user of the account
can use the default customer master key ID to enable client-side encryption.
IAM Assume Role The Amazon Resource Number (ARN) of the IAM role assumed by the user to use the
ARN dynamically generated temporary security credentials.
Set the value of this property if you want to use the temporary security credentials to access
Amazon Redshift.
For more information about how to get the ARN of the IAM role, see the AWS documentation.
Cluster Identifier The unique identifier of the cluster that hosts Amazon Redshift for which you are requesting
the security credentials.
Specify the cluster name.
Auto Create DBUser Select to create a new Amazon Redshift database user at run time.
Default is disabled.
Note: Use this option to create a new Amazon Redshift database user. If you create a new
user directly using the AWS command line, the user is not created in Amazon Redshift.
Expiration Time The time duration that the password for the Amazon Redshift database user expires.
Specify a value between 900 seconds and 3600 seconds.
Default is 900.
Use EC2 Role to Select to enable the EC2 role to assume another IAM role specified in the IAM AssumeRole
Assume Role ARN option.
IAM User Access Key The access key of the IAM user that has permissions to assume the IAM AssumeRole ARN.
IAM User Secret The secret access key of the IAM user that has permissions to assume the IAM Assume Role
Access Key ARN.
Property Description
Name The name of the connection. The name is not case sensitive and must be unique within the domain.
You can change this property after you create the connection. The name cannot exceed 128
characters, contain spaces, or contain the following special characters:~ ` ! $ % ^ & * ( ) - + = { [ } ] |
\:;"'<,>.?/
ID String that the Data Integration Service uses to identify the connection. The ID is not case sensitive.
It must be 255 characters or less and must be unique in the domain. You cannot change this
property after you create the connection. Default value is the connection name.
Description Optional. The description of the connection. The description cannot exceed 4,000 characters.
Access Key Access key to access the Amazon S3 bucket. Provide the access key value based on the following
authentication methods:
- Basic authentication: provide the actual access key value.
- IAM authentication: do not provide the access key value.
- Temporary security credentials via assume role: provide access key of an IAM user with no
permissions to access Amazon S3 bucket.
IAM Role ARN The ARN of the IAM role assumed by the user to use the dynamically generated temporary security
credentials.
Enter the value of this property if you want to use the temporary security credentials to access the
AWS resources.
If you want to use the temporary security credentials with IAM authentication, do not provide the
Access Key and Secret Key connection properties. If you want to use the temporary security
credentials without IAM authentication, you must enter the value of the Access Key and Secret Key
connection properties.
For more information about how to obtain the ARN of the IAM role, see the AWS documentation.
Folder Path The complete path to Amazon S3 objects. The path must include the bucket name and any folder
name.
Do not use a slash at the end of the folder path. For example, <bucket name>/<my folder
name>.
Master Optional. Provide a 256-bit AES encryption key in the Base64 format when you enable client-side
Symmetric Key encryption. You can generate a master symmetric key using a third-party tool.
Region Name Select the AWS region in which the bucket you want to access resides.
Select one of the following regions:
- Asia Pacific (Mumbai)
- Asia Pacific (Seoul)
- Asia Pacific (Singapore)
- Asia Pacific (Sydney)
- Asia Pacific (Tokyo)
- AWS GovCloud (US)
- AWS GovCloud (US-East)
- Canada (Central)
- China (Beijing)
- China (Hong Kong)
- China (Ningxia)
- EU (Ireland)
- EU (Frankfurt)
- EU (London)
- EU (Paris)
- EU (Stockholm)
- Middle East (Bahrain)
- South America (Sao Paulo)
- US East (Ohio)
- US East (N. Virginia)
- US West (N. California)
- US West (Oregon)
Default is US East (N. Virginia).
Not applicable for S3 compatible storage.
Customer Optional. Specify the customer master key ID or alias name generated by AWS Key Management
Master Key ID Service (AWS KMS) or the Amazon Resource Name (ARN) of your custom key for cross-account
access. You must generate the customer master key for the same region where Amazon S3 bucket
reside.
You can specify any of the following values:
Customer generated customer master key
Enables client-side or server-side encryption. Only the administrator user of the account can use
the default customer master key ID to enable client-side encryption.
Federated SSO SAML 2.0-enabled identity provider for the federated user single sign-on to use with the AWS
IdP account.
PowerExchange for Amazon S3 supports only the ADFS 3.0 identity provider.
Select None if you do not want to use federated user single sign-on.
Property Description
Federated User Name User name of the federated user to access the AWS account through the identity
provider.
Federated User Password Password for the federated user to access the AWS account through the identity
provider.
IdP SSO URL Single sign-on URL of the identity provider for AWS.
SAML Identity Provider ARN ARN of the SAML identity provider that the AWS administrator created to register the
identity provider as a trusted provider.
Role ARN ARN of the IAM role assumed by the federated user.
The following table describes the general connection properties for a blockchain connection:
Property Description
Name The name of the connection. The name is not case sensitive and must be unique within the
domain. You can change this property after you create the connection. The name cannot
exceed 128 characters, contain spaces, or contain the following special characters:
~ ` ! $ % ^ & * ( ) - + = { [ } ] | \ : ; " ' < , > . ? /
ID The string that the Data Integration Service uses to identify the connection. The ID is not
case sensitive. It must be 255 characters or less and must be unique in the domain. You
cannot change this property after you create the connection. Default value is the
connection name.
Description The description of the connection. Enter a string that you can use to identify the
connection. The description cannot exceed 4,000 characters.
Swagger File Path The absolute path of the swagger file path that contains the REST API to communicate with
the blockchain. The swagger file must be a JSON file that is stored on the Data Integration
Service machine. If the swagger file is in a different file format, such as YAML, convert the
file to JSON format.
Base URL Required. The base URL that is used to access assets on the blockchain.
Auth Type* Authentication method that the run-time engine uses to connect to the REST server. You
can use none, basic, digest, or OAuth.
Auth Password* Password for the user name to authenticate to the REST server.
OAuth Consumer Key* Required for the OAuth authentication type. Client key that is associated with the REST
server.
OAuth Consumer Required for the OAuth authentication type. Client password to connect to the REST server.
Secret*
OAuth Token* Required for the OAuth authentication type. Access token to connect to the REST server.
OAuth Token Secret* Required for the OAuth authentication type. Password associated with the OAuth token.
Proxy Type* Type of proxy. You can use no proxy, platform proxy, or custom.
TrustStore File Path* The absolute path of the truststore file that contains the SSL certificate.
KeyStore File Path* The absolute path of the keystore file that contains the keys and certificates required to
establish a two-way secure connection with the REST server.
Advanced Properties List of advanced properties to access an asset on the blockchain. Specify the advanced
properties using name-value pairs that are separated by a semicolon.
You can use the following advanced properties:
- X-API-KEY. Required if you authenticate to the REST server using an API key.
The advanced properties that you configure in the connection override the values for the
corresponding advanced properties in the blockchain data object. For example, if the
connection and the data object both specify a base URL, the value in the connection
overrides the value in the data object.
Note: The advanced properties have the precedence level,Operation level > Object level >
Connection level. The properties configured at the operation level will override the
properties configured at the object or connection level.
Cookies Required based on how the REST API is implemented. List of cookie properties to specify
the cookie information that is passed to the REST server. Specify the properties using
name-value pairs that are separated by a semicolon.
The cookie properties that you configure in the connection override the values for the
corresponding cookie properties in the blockchain data object.
* The property is ignored. To use the functionality, configure the property as an advanced property and provide a name-
value pair based on the property name in the swagger file.
For example, configure the following name-value pair to use basic authorization:
Authorization=Basic <credentials>
Note: You cannot use Test Connection to validate a Blockchain connection.
Note: The order of the connection properties might vary depending on the tool where you view them.
Property Description
Name The name of the connection. The name is not case sensitive and must be unique within the domain.
You can change this property after you create the connection. The name cannot exceed 128
characters, contain spaces, or contain the following special characters:
~ ` ! $ % ^ & * ( ) - + = { [ } ] | \ : ; " ' < , > . ? /
ID String that the Data Integration Service uses to identify the connection.
The ID is not case sensitive. The ID must be 255 characters or less and must be unique in the
domain. You cannot change this property after you create the connection.
Default value is the connection name.
Description Optional. The description of the connection. The description cannot exceed 4,000 characters.
Password Password corresponding to the user name to access the Cassandra server.
SQL Identifier Type of character that the database uses to enclose delimited identifiers in SQL or CQL queries. The
Character available characters depend on the database type.
Select None if the database uses regular identifiers. When the Data Integration Service generates
SQL or CQL queries, the service does not place delimited characters around any identifiers.
Select a character if the database uses delimited identifiers. When the Data Integration Service
generates SQL or CQL queries, the service encloses delimited identifiers within this character.
SSL Not applicable for PowerExchange for Cassandra JDBC or when disabled SSL mode is selected.
Truststore Absolute path and file name of the SSL truststore file that contains certificates of the trusted SSL
Path server.
SSL Not applicable for PowerExchange for Cassandra JDBC or when disabled SSL mode is selected.
Truststore Password for the SSL truststore.
Password
SSL Keystore Not applicable for PowerExchange for Cassandra JDBC or when disabled SSL mode is selected.
Path Absolute path and file name of the SSL keystore file that contains private keys and certificates for
the SSL server.
SSL Keystore Not applicable for PowerExchange for Cassandra JDBC or when disabled SSL mode is selected.
Password Password for the SSL keystore.
Additional Enter one or more JDBC connection parameters in the following format:
Connection <param1>=<value>;<param2>=<value>;<param3>=<value>
Properties
PowerExchange for Cassandra JDBC supports the following JDBC connection parameters:
- BinaryColumnLength
- DecimalColumnScale
- EnableCaseSensitive
- EnableNullInsert
- EnablePaging
- RowsPerPage
- StringColumnLength
- VTTableNameSeparator
When you configure a Confluent Kafka connection, you configure the following properties:
• List of Kafka brokers or Confluent Kafka brokers that the connection reads from or writes to.
• Number of seconds the Integration Service attempts to reconnect to the database if the connection fails.
• Version of the Confluent Kafka messaging broker.
Property Description
Name Name of the connection. The name is not case sensitive and must be unique within the domain. You can
change this property after you create the connection. The name cannot exceed 128 characters, contain
spaces, or contain the following special characters:
~ ` ! $ % ^ & * ( ) - + = { [ } ] | \ : ; " ' < , > . ? /
ID The string that the Data Integration Service uses to identify the connection. The ID is not case sensitive.
It must be 255 characters or less and must be unique in the domain. You cannot change this property
after you create the connection. Default value is the connection name.
Description Description of the connection. Enter a string that you can use to identify the connection. The
description cannot exceed 4,000 characters.
Location Domain where you want to create the connection. Select the domain name.
Property Description
Kafka Broker List Comma-separated list of Confluent Kafka brokers that maintain the
configuration of the Confluent Kafka messaging broker.
To specify a Confluent Kafka broker, use the following format: <IP
address>:<port>
Retry Timeout Number of seconds after which the Data Integration Service attempts to
reconnect to the Confluent Kafka broker to read or write data. If the
source or target is not available for the time you specify, the mapping
execution stops to avoid any data loss.
Schema Registry URL Location and port of the schema registry provider on which to connect.
Property Description
SSL Mode Optional. SSL mode indicates the encryption type to use for the connection.
You can choose one of the following SSL modes:
- Disabled
- One way
- Two way
The default value is Disabled.
SSL TrustStore File Required when One way SSL mode is selected.
Path Absolute path and file name of the SSL truststore file that contains certificates of the
trusted SSL server.
SSL KeyStore File Path Required when Two way SSL mode is selected.
Absolute path and file name of the SSL keystore file that contains private keys and
certificates for the SSL server.
Additional Security Optional. Comma-separated list of connection properties to connect to the Confluent Kafka
Properties broker in a secured way.
A Databricks connection is a cluster type connection. You can create and manage a Databricks connection in
the Administrator tool or the Developer tool. You can use infacmd to create a Databricks connection.
Configure properties in the Databricks connection to enable communication between the Data Integration
Service and the Databricks cluster.
Property Description
Name The name of the connection. The name is not case sensitive and must be unique within the
domain. You can change this property after you create the connection. The name cannot exceed
128 characters, contain spaces, or contain the following special characters:~ ` ! $ % ^ & * ( ) - + =
{[}]|\:;"'<,>.?/
ID String that the Data Integration Service uses to identify the connection. The ID is not case
sensitive. It must be 255 characters or less and must be unique in the domain. You cannot change
this property after you create the connection. Default value is the connection name.
Description Optional. The description of the connection. The description cannot exceed 4,000 characters.
Cluster Name of the cluster configuration associated with the Databricks environment.
Configuration Required if you do not configure the cloud provisioning configuration.
Cloud Name of the cloud provisioning configuration associated with a Databricks cloud platform.
Provisioning Required if you do not configure the cluster configuration.
Configuration
Staging Directory The directory where the Databricks Spark engine stages run-time files.
If you specify a directory that does not exist, the Data Integration Service creates it at run time.
If you do not provide a directory path, the run-time staging files are written to /<cluster staging
directory>/DATABRICKS.
Advanced List of advanced properties that are unique to the Databricks environment.
Properties You can configure run-time properties for the Databricks environment in the Data Integration
Service and in the Databricks connection. You can override a property configured at a high level
by setting the value at a lower level. For example, if you configure a property in the Data
Integration Service custom properties, you can override it in the Databricks connection. The Data
Integration Service processes property overrides based on the following priorities:
1. Databricks connection advanced properties
2. Data Integration Service custom properties
Note: Informatica does not recommend changing these property values before you consult with
third-party documentation, Informatica documentation, or Informatica Global Customer Support. If
you change a value without knowledge of the property, you might experience performance
degradation or other unexpected results.
Advanced Properties
Configure the following properties in the Advanced Properties of the Databricks configuration section:
infaspark.json.parser.mode
Specifies the parser how to handle corrupt JSON records. You can set the value to one of the following
modes:
Specifies whether the parser can read a multiline record in a JSON file. You can set the value to true or
false. Default is false. Applies only to non-native distributions that use Spark version 2.2.x and above.
infaspark.flatfile.writer.nullValue
When the Databricks Spark engine writes to a target, it converts null values to empty strings (" "). For
example, 12, AB,"",23p09udj.
The Databricks Spark engine can write the empty strings to string columns, but when it tries to write an
empty string to a non-string column, the mapping fails with a type mismatch.
To allow the Databricks Spark engine to convert the empty strings back to null values and write to the
target, configure the property in the Databricks Spark connection.
infaspark.pythontx.exec
Required to run a Python transformation on the Databricks Spark engine. Set to the location of the
Python executable binary on the worker nodes in the Databricks cluster.
When you provision the cluster at run time, set this property in the Databricks cloud provisioning
configuration. Otherwise, set on the Databricks connection.
Required to run a Python transformation on the Databricks Spark engine. Set to the location of the
Python installation directory on the worker nodes in the Databricks cluster.
When you provision the cluster at run time, set this property in the Databricks cloud provisioning
configuration. Otherwise, set on the Databricks connection.
Note: The order of the connection properties might vary depending on the tool where you view them.
When you create a Greenplum connection, you enter information for metadata and data access.
Property Description
ID String that the Data Integration Service uses to identify the connection. The ID is not case sensitive. It
must be 255 characters or fewer and must be unique in the domain. You cannot change this property
after you create the connection. Default value is the connection name.
Description Description of the connection. The description cannot exceed 765 characters.
The user name, password, driver name, and connection string are required to import the metadata. The
following table describes the properties for metadata access:
Property Description
User Name User name with permissions to access the Greenplum database.
PowerExchange for Greenplum uses the host name, port number, and database name to create a control file
to provide load specifications to the Greenplum gpload bulk loading utility. It uses the Enable SSL option and
the certificate path to establish secure communication to the Greenplum server over SSL.
The following table describes the connection properties for data access:
Property Description
Port Number Greenplum server port number. If you enter 0, the gpload utility reads from the environment variable
$PGPORT. Default is 5432.
Enable SSL Select this option to establish secure communication between the gpload utility and the Greenplum
server over SSL.
Certificate Path Path where the SSL certificates for the Greenplum server are stored.
For information about the files that need to be present in the certificates path, see the gpload
documentation.
Note: The order of the connection properties might vary depending on the tool where you view them.
Property Description
Name The name of the connection. The name is not case sensitive and must be unique within the
domain. You can change this property after you create the connection. The name cannot exceed
128 characters, contain spaces, or contain the following special characters:~ ` ! $ % ^ & * ( ) - + =
{[}]|\:;"'<,>.?/
ID String that the Data Integration Service uses to identify the connection.
The ID is not case sensitive. The ID must be 255 characters or less and must be unique in the
domain. You cannot change this property after you create the connection.
Default value is the connection name.
Description Optional. The description of the connection. The description cannot exceed 4,000 characters.
Service Account Specifies the client_email value present in the JSON file that you download after you create a
ID service account.
Service Account Specifies the private_key value present in the JSON file that you download after you create a
Key service account.
APIVersion API that PowerExchange for Google Analytics uses to read from Google Analytics reports.
Select Core Reporting API v3.
Note: PowerExchange for Google Analytics does not support Analytics Reporting API v4.
Note: The order of the connection properties might vary depending on the tool where you view them.
Property Description
Service Account Specifies the client_email value present in the JSON file that you download after you create a
ID service account in Google BigQuery.
Service Account Specifies the private_key value present in the JSON file that you download after you create a
Key service account in Google BigQuery.
Connection The mode that you want to use to read data from or write data to Google BigQuery.
mode Select one of the following connection modes:
- Simple. Flattens each field within the Record data type field as a separate field in the mapping.
- Hybrid. Displays all the top-level fields in the Google BigQuery table including Record data type
fields. PowerExchange for Google BigQuery displays the top-level Record data type field as a
single field of the String data type in the mapping.
- Complex. Displays all the columns in the Google BigQuery table as a single field of the String
data type in the mapping.
Default is Simple.
Schema Specifies a directory on the client machine where the must create a JSON file with the sample
Definition File schema of the Google BigQuery table. The JSON file name is the same as the Google BigQuery
Path table name.
Alternatively, you can specify a storage path in Google Cloud Storage where the must create a
JSON file with the sample schema of the Google BigQuery table. You can download the JSON file
from the specified storage path in Google Cloud Storage to a local machine.
Project ID Specifies the project_id value present in the JSON file that you download after you create a
service account in Google BigQuery.
If you have created multiple projects with the same service account, enter the ID of the project
that contains the dataset that you want to connect to.
Storage Path This property applies when you read or write large volumes of data.
Path in Google Cloud Storage where the creates a local stage file to store the data temporarily.
You can either enter the bucket name or the bucket name and folder name.
For example, enter gs://<bucket_name> or gs://<bucket_name>/<folder_name>
Use Legacy SQL Not applicable for PowerExchange for Google BigQuery.
For Custom
Query
Region ID The region name where the Google BigQuery dataset resides.
For example, if you want to connect to a Google BigQuery dataset that resides in Las Vegas
region, specify us-west4 as the Region ID.
Note: In the Storage Path connection property, ensure that you specify a bucket name or the
bucket name and folder name that resides in the same region as the dataset in Google BigQuery.
For more information about the regions supported by Google BigQuery, see the following Google
BigQuery documentation:https://cloud.google.com/bigquery/docs/locations
Optional Specifies whether you can configure certain source and target functionalities through custom
Properties properties.
You can select one of the following options:
- None. Select if you do not want to configure any custom properties.
- Required. If you want to specify custom properties to configure the source and target
functionalities.
Default is None.
Provide Optional Comma-separated key-value pairs of custom properties to enable certain source and target
Properties functionalities.
Appears only when you select Required in the Optional Properties.
For more information about the list of custom properties that you can specify, see the Informatica
Knowledge Base article: https://kb.informatica.com/faq/7/Pages/26/632722.aspx
Connection Modes
You can configure a Google BigQuery connection to use one of the following connection modes:
Simple mode
If you use simple mode, PowerExchange for Google BigQuery flattens each field within the Record data
type field as a separate field in the Google BigQuery data object.
Hybrid mode
If you use hybrid mode, PowerExchange for Google BigQuery displays all the top-level fields in the
Google BigQuery table including Record data type fields. PowerExchange for Google BigQuery displays
the top-level Record data type field as a single field of the String data type in the Google BigQuery data
object.
Complex mode
If you use complex mode, PowerExchange for Google BigQuery displays all the columns in the Google
BigQuery table as a single field of the String data type in the Google BigQuery data object.
Note: The order of the connection properties might vary depending on the tool where you view them.
Property Description
Name The name of the connection. The name is not case sensitive and must be unique within the
domain. You can change this property after you create the connection. The name cannot exceed
128 characters, contain spaces, or contain the following special characters:~ ` ! $ % ^ & * ( ) - + =
{[}]|\:;"'<,>.?/
ID String that the Data Integration Service uses to identify the connection.
The ID is not case sensitive. The ID must be 255 characters or less and must be unique in the
domain. You cannot change this property after you create the connection.
Default value is the connection name.
Description Optional. The description of the connection. The description cannot exceed 4,000 characters.
Project ID Specifies the project_id value present in the JSON file that you download after you create a service
account.
If you have created multiple projects with the same service account, enter the ID of the project
that contains the bucket that you want to connect to.
Service Account Specifies the client_email value present in the JSON file that you download after you create a
ID service account.
Service Account Specifies the private_key value present in the JSON file that you download after you create a
Key service account.
Instance ID Name of the instance that you created in Google Cloud Spanner.
Note: The order of the connection properties might vary depending on the tool where you view them.
The following table describes the Google Cloud Storage connection properties:
Property Description
Name The name of the connection. The name is not case sensitive and must be unique within the
domain. You can change this property after you create the connection. The name cannot exceed
128 characters, contain spaces, or contain the following special characters:~ ` ! $ % ^ & * ( ) - + =
{[}]|\:;"'<,>.?/
ID String that the Data Integration Service uses to identify the connection.
The ID is not case sensitive. The ID must be 255 characters or less and must be unique in the
domain. You cannot change this property after you create the connection.
Default value is the connection name.
Description Optional. The description of the connection. The description cannot exceed 4,000 characters.
Project ID Specifies the project_id value present in the JSON file that you download after you create a service
account.
If you have created multiple projects with the same service account, enter the ID of the project that
contains the bucket that you want to connect to.
Service Account Specifies the client_email value present in the JSON file that you download after you create a
ID service account.
Service Account Specifies the private_key value present in the JSON file that you download after you create a
Key service account.
Property Description
Name The name of the connection. The name is not case sensitive and must be unique within the
domain. You can change this property after you create the connection. The name cannot exceed
128 characters, contain spaces, or contain the following special characters: ~ ` ! $ % ^ & *
( ) - + = { [ } ] | \ : ; " ' < , > . ? /
ID The string that the Data Integration Service uses to identify the connection. The ID is not case
sensitive. It must be 255 characters or less and must be unique in the domain. You cannot change
this property after you create the connection. Default value is the connection name.
Description The description of the connection. Enter a string that you can use to identify the connection. The
description cannot exceed 4,000 characters.
Connection Type The connection type. Select Pub Sub connection type.
Client Email The client_email value available in the JSON file that you download after you create a service
account.
Client Id The client_id value available in the JSON file that you download after you create a service
account.
Private Key Id The private_key_id value available in the JSON file that you download after you create a
service account.
Private Key The private_key value available in the JSON file that you download after you create a service
account.
Project Id The project_id value available in the JSON file that you download after you create a service
account.
You can configure run-time properties for the Hadoop environment in the Data Integration Service, the
Hadoop connection, and in the mapping. You can override a property configured at a high level by setting the
value at a lower level. For example, if you configure a property in the Data Integration Service custom
properties, you can override it in the Hadoop connection or in the mapping. The Data Integration Service
processes property overrides based on the following priorities:
1. Mapping custom properties set using infacmd ms runMapping with the -cp option
2. Mapping run-time properties for the Hadoop environment
3. Hadoop connection advanced properties for run-time engines
4. Hadoop connection advanced general properties, environment variables, and classpaths
5. Data Integration Service custom properties
Note: When a mapping uses Hive Server 2 to run a job or parts of a job, you cannot override properties that
are configured on the cluster level in preSQL or post-SQL queries or SQL override statements. Workaround:
Instead of attempting to use the cluster configuration on the domain to override cluster properties, pass the
override settings to the JDBC URL. For example: beeline -u "jdbc:hive2://<domain
host>:<port_number>/tpch_text_100" --hiveconf hive.execution.engine=tez
The following table describes the general connection properties for the Hadoop connection:
Property Description
Name The name of the connection. The name is not case sensitive and must be unique within the
domain. You can change this property after you create the connection. The name cannot
exceed 128 characters, contain spaces, or contain the following special characters:
~ ` ! $ % ^ & * ( ) - + = { [ } ] | \ : ; " ' < , > . ? /
ID String that the Data Integration Service uses to identify the connection. The ID is not case
sensitive. It must be 255 characters or less and must be unique in the domain. You cannot
change this property after you create the connection. Default value is the connection name.
Description The description of the connection. Enter a string that you can use to identify the connection.
The description cannot exceed 4,000 characters.
Cluster The name of the cluster configuration associated with the Hadoop environment.
Configuration Required if you do not configure the Cloud Provisioning Configuration.
Cloud Provisioning Name of the cloud provisioning configuration associated with a cloud platform such as
Configuration Amazon AWS or Microsoft Azure.
Required if you do not configure the Cluster Configuration.
Cluster Classpath* The classpath to access the Hadoop jar files and the required libraries.
The $DEFAULT_CLUSTER_CLASSPATH variable contains a list of paths to the default jar files
and libraries.
You can configure run-time properties for the Hadoop environment in the Data Integration
Service, the Hadoop connection, and in the mapping. You can override a property configured at
a high level by setting the value at a lower level. For example, if you configure a property in the
Data Integration Service custom properties, you can override it in the Hadoop connection or in
the mapping. The Data Integration Service processes property overrides based on the following
priorities:
1. Mapping custom properties set using infacmd ms runMapping with the -cp option
2. Mapping run-time properties for the Hadoop environment
3. Hadoop connection advanced properties for run-time engines
4. Hadoop connection advanced general properties, environment variables, and classpaths
5. Data Integration Service custom properties
Note: When a mapping uses Hive Server 2 to run a job or parts of a job, you cannot override
properties that are configured on the cluster level in preSQL or post-SQL queries or SQL
override statements. Workaround: Instead of attempting to use the cluster configuration on the
domain to override cluster properties, pass the override settings to the JDBC URL. For
example: beeline -u "jdbc:hive2://<domain host>:<port_number>/
tpch_text_100" --hiveconf hive.execution.engine=tez
* Informatica does not recommend changing these property values before you consult with third-party documentation,
Informatica documentation, or Informatica Global Customer Support. If you change a value without knowledge of the
property, you might experience performance degradation or other unexpected results.
Common Properties
The following table describes the common connection properties that you configure for the Hadoop
connection:
Property Description
Impersonation Required if the Hadoop cluster uses Kerberos authentication. Hadoop impersonation user. The
User Name user name that the Data Integration Service impersonates to run mappings in the Hadoop
environment.
Data Engineering Integration supports operating system profiles on all Hadoop distributions. In
the Hadoop run-time environment, the Data Integration Service pushes the processing to the
Hadoop cluster and the run-time engines run mappings with the operating system profile
specified Hadoop impersonation properties.
Temporary Table Hadoop compression library for a compression codec class name.
Compression Note: The Spark engine does not support compression settings for temporary tables. When you
Codec run mappings on the Spark engine, the Spark engine stores temporary tables in an uncompressed
file format.
Codec Class Codec class name that enables data compression and improves performance on temporary
Name staging tables.
Hive Staging Namespace for Hive staging tables. Use the name default for tables that do not have a
Database Name specified database name.
If you do not configure a namespace, the Data Integration Service uses the Hive database name
in the Hive target connection to create staging tables.
When you run a mapping in the native environment to write data to Hive, you must configure the
Hive staging database name in the Hive connection. The Data Integration Service ignores the
value you configure in the Hadoop connection.
Environment SQL SQL commands to set the Hadoop environment. The Data Integration Service executes the
environment SQL at the beginning of each Hive script generated by a HiveServer2 job.
The following rules and guidelines apply to the usage of environment SQL:
- You can use environment SQL to define Hadoop or Hive parameters that you want to use in the
PreSQL commands or in custom queries.
- If you use multiple values for the Environment SQL property, ensure that there is no space
between the values.
Engine Type The Data Integration Service uses HiveServer2 to process portions of some jobs by running
HiveServer2 tasks on the Spark engine. When you import the cluster configuration through the
Administrator tool, you can choose to create connections. The engine type property is populated
by default based on the distribution.
When you manually create a connection, you must configure the engine type.
You can specify the engine type based on the following Hadoop distributions:
- Amazon EMR. Tez
- Azure HDI. Tez
- Cloudera CDH. MRv2
- Cloudera CDP. Tez
- Dataproc. MRv2
- Hortonworks HDP. Tez
- MapR. MRv2
Advanced List of advanced properties that are unique to the Hadoop environment. The properties are
Properties common to the Blaze and Spark engines. The advanced properties include a list of default
properties.
You can configure run-time properties for the Hadoop environment in the Data Integration
Service, the Hadoop connection, and in the mapping. You can override a property configured at a
high level by setting the value at a lower level. For example, if you configure a property in the
Data Integration Service custom properties, you can override it in the Hadoop connection or in the
mapping. The Data Integration Service processes property overrides based on the following
priorities:
1. Mapping custom properties set using infacmd ms runMapping with the -cp option
2. Mapping run-time properties for the Hadoop environment
3. Hadoop connection advanced properties for run-time engines
4. Hadoop connection advanced general properties, environment variables, and classpaths
5. Data Integration Service custom properties
Note: When a mapping uses Hive Server 2 to run a job or parts of a job, you cannot override
properties that are configured on the cluster level in preSQL or post-SQL queries or SQL override
statements. Workaround: Instead of attempting to use the cluster configuration on the domain to
override cluster properties, pass the override settings to the JDBC URL. For example: beeline -
u "jdbc:hive2://<domain host>:<port_number>/tpch_text_100" --hiveconf
hive.execution.engine=tez
Note: Informatica does not recommend changing these property values before you consult with
third-party documentation, Informatica documentation, or Informatica Global Customer Support.
If you change a value without knowledge of the property, you might experience performance
degradation or other unexpected results.
Property Description
Write Reject Files If you use the Blaze engine to run mappings, select the check box to specify a location to move
to Hadoop reject files. If checked, the Data Integration Service moves the reject files to the HDFS location
listed in the property, Reject File Directory.
By default, the Data Integration Service stores the reject files based on the RejectDir system
parameter.
Reject File The directory for Hadoop mapping files on HDFS when you run mappings.
Directory
Blaze Configuration
The following table describes the connection properties that you configure for the Blaze engine:
Property Description
Blaze Staging The HDFS file path of the directory that the Blaze engine uses to store temporary files. Verify that
Directory the directory exists. The YARN user, Blaze engine user, and mapping impersonation user must have
write permission on this directory.
Default is /blaze/workdir. If you clear this property, the staging files are written to the Hadoop
staging directory /tmp/blaze_<user name>.
Blaze User The owner of the Blaze service and Blaze service logs.
Name When the Hadoop cluster uses Kerberos authentication, the default user is the Data Integration
Service SPN user. When the Hadoop cluster does not use Kerberos authentication and the Blaze user
is not configured, the default user is the Data Integration Service user.
Minimum Port The minimum value for the port number range for the Blaze engine. Default is 12300.
Maximum Port The maximum value for the port number range for the Blaze engine. Default is 12600.
YARN Queue The YARN scheduler queue name used by the Blaze engine that specifies available resources on a
Name cluster.
Note: If YARN preemption is enabled on the cluster, verify with the Hadoop administrator that
preemption is disabled on the queue associated with the Blaze engine.
Blaze Job The host name and port number for the Blaze Job Monitor.
Monitor Use the following format:
Address <hostname>:<port>
Where
- <hostname> is the host name or IP address of the Blaze Job Monitor server.
- <port> is the port on which the Blaze Job Monitor listens for remote procedure calls (RPC).
For example, enter: myhostname:9080
Blaze YARN Node label that determines the node on the Hadoop cluster where the Blaze engine runs. If you do
Node Label not specify a node label, the Blaze engine runs on the nodes in the default partition.
If the Hadoop cluster supports logical operators for node labels, you can specify a list of node
labels. To list the node labels, use the operators && (AND), || (OR), and ! (NOT).
Note: You cannot use node labels on a Cloudera CDH cluster.
Advanced List of advanced properties that are unique to the Blaze engine. The advanced properties include a
Properties list of default properties.
You can configure run-time properties for the Hadoop environment in the Data Integration Service,
the Hadoop connection, and in the mapping. You can override a property configured at a high level
by setting the value at a lower level. For example, if you configure a property in the Data Integration
Service custom properties, you can override it in the Hadoop connection or in the mapping. The Data
Integration Service processes property overrides based on the following priorities:
1. Mapping custom properties set using infacmd ms runMapping with the -cp option
2. Mapping run-time properties for the Hadoop environment
3. Hadoop connection advanced properties for run-time engines
4. Hadoop connection advanced general properties, environment variables, and classpaths
5. Data Integration Service custom properties
Note: When a mapping uses Hive Server 2 to run a job or parts of a job, you cannot override
properties that are configured on the cluster level in preSQL or post-SQL queries or SQL override
statements. Workaround: Instead of attempting to use the cluster configuration on the domain to
override cluster properties, pass the override settings to the JDBC URL. For example: beeline -u
"jdbc:hive2://<domain host>:<port_number>/tpch_text_100" --hiveconf
hive.execution.engine=tez
Note: Informatica does not recommend changing these property values before you consult with
third-party documentation, Informatica documentation, or Informatica Global Customer Support. If
you change a value without knowledge of the property, you might experience performance
degradation or other unexpected results.
Spark Configuration
The following table describes the connection properties that you configure for the Spark engine:
Property Description
Spark Staging The HDFS file path of the directory that the Spark engine uses to store temporary files for running
Directory jobs. The YARN user, Data Integration Service user, and mapping impersonation user must have
write permission on this directory.
If you do not specify a file path, by default, the temporary files are written to the Hadoop staging
directory /tmp/SPARK_<user name>.
When you run Sqoop jobs on the Spark engine, the Data Integration Service creates a Sqoop staging
directory within the Spark staging directory to store temporary files: <Spark staging
directory>/sqoop_staging
Spark Event Optional. The HDFS file path of the directory that the Spark engine uses to log events.
Log Directory
YARN Queue The YARN scheduler queue name used by the Spark engine that specifies available resources on a
Name cluster. The name is case sensitive.
Advanced List of advanced properties that are unique to the Spark engine. The advanced properties include a
Properties list of default properties.
You can configure run-time properties for the Hadoop environment in the Data Integration Service,
the Hadoop connection, and in the mapping. You can override a property configured at a high level
by setting the value at a lower level. For example, if you configure a property in the Data Integration
Service custom properties, you can override it in the Hadoop connection or in the mapping. The Data
Integration Service processes property overrides based on the following priorities:
1. Mapping custom properties set using infacmd ms runMapping with the -cp option
2. Mapping run-time properties for the Hadoop environment
3. Hadoop connection advanced properties for run-time engines
4. Hadoop connection advanced general properties, environment variables, and classpaths
5. Data Integration Service custom properties
Note: When a mapping uses Hive Server 2 to run a job or parts of a job, you cannot override
properties that are configured on the cluster level in preSQL or post-SQL queries or SQL override
statements. Workaround: Instead of attempting to use the cluster configuration on the domain to
override cluster properties, pass the override settings to the JDBC URL. For example: beeline -u
"jdbc:hive2://<domain host>:<port_number>/tpch_text_100" --hiveconf
hive.execution.engine=tez
Note: Informatica does not recommend changing these property values before you consult with
third-party documentation, Informatica documentation, or Informatica Global Customer Support. If
you change a value without knowledge of the property, you might experience performance
degradation or other unexpected results.
Property Description
Name The name of the connection. The name is not case sensitive and must be unique
within the domain. You can change this property after you create the connection.
The name cannot exceed 128 characters, contain spaces, or contain the following
special characters:
~ ` ! $ % ^ & * ( ) - + = { [ } ] | \ : ; " ' < , > . ? /
ID String that the Data Integration Service uses to identify the connection. The ID is
not case sensitive. It must be 255 characters or less and must be unique in the
domain. You cannot change this property after you create the connection. Default
value is the connection name.
Description The description of the connection. The description cannot exceed 4,000
characters.
Note: The order of the connection properties might vary depending on the tool where you view them.
Property Description
Name Name of the connection. The name is not case sensitive and must be unique within the domain. The
name cannot exceed 128 characters, contain spaces, or contain the following special characters:
~ ` ! $ % ^ & * ( ) - + = { [ } ] | \ : ; " ' < , > . ? /
ID String that the Data Integration Service uses to identify the connection. The ID is not case sensitive.
It must be 255 characters or less and must be unique in the domain. You cannot change this
property after you create the connection. Default value is the connection name.
Description The description of the connection. The description cannot exceed 765 characters.
Location The domain where you want to create the connection. Not valid for the Analyst tool.
NameNode The URI to access the storage system. You can find the value for fs.defaultFS in the core-
URI site.xml configuration set of the cluster configuration.
If you create connections when you import the cluster configuration, the NameNode URI property is
populated by default, and it is updated each time you refresh the cluster configuration.
If you use a Cloudera CDP Public Cloud compute cluster and the HDFS is on a Cloudera Data Lake
cluster, set the property spark.yarn.access.hadoopFileSystems in the Spark properties of the Hadoop
Connection to the same value as set here.
HDFS hdfs://<namenode>:<port>
where:
- <namenode> is the host name or IP address of the NameNode.
- <port> is the port that the NameNode listens for remote procedure calls (RPC).
hdfs://<nameservice> in case of NameNode high availability.
MapR-FS maprfs:///
WASB in wasb://<container_name>@<account_name>.blob.core.windows.net/<path>
HDInsight
where:
- <container_name> identifies a specific Azure Storage Blob container.
Note: <container_name> is optional.
- <account_name> identifies the Azure Storage Blob object.
Example:
wasb://infabdmoffering1storage.blob.core.windows.net/
infabdmoffering1cluster/mr-history
ADLS in adl://home
HDInsight
When you create a cluster configuration from an Azure HDInsight cluster, the cluster configuration uses
either ADLS or WASB as the primary storage. You cannot create a cluster configuration with ADLS or WASB
as the secondary storage. You can edit the NameNode URI property in the HDFS connection to connect to a
local HDFS location.
The following table describes the HBase connection properties for MapR-DB:
Property Description
Name Name of the connection. The name is not case sensitive and must be unique
within the domain. You can change this property after you create the connection.
The name cannot exceed 128 characters, contain spaces, or contain the following
special characters:
~ ` ! $ % ^ & * ( ) - + = { [ } ] | \ : ; " ' < , > . ? /
ID String that the Data Integration Service uses to identify the connection. The ID is
not case sensitive. It must be 255 characters or less and must be unique in the
domain. You cannot change this property after you create the connection. Default
value is the connection name.
Description Description of the connection. The description cannot exceed 4,000 characters.
Cluster Configuration The name of the cluster configuration associated with the Hadoop environment.
MapR-DB Database Path Database path that contains the MapR-DB table that you want to connect to. Enter
a valid MapR cluster path.
When you create an HBase data object for MapR-DB, you can browse only tables
that exist in the MapR-DB path that you specify in the Database Path field. You
cannot access tables that are available in sub-directories in the specified path.
For example, if you specify the path as /user/customers/, you can access the
tables in the customers directory. However, if the customers directory contains
a sub-directory named regions, you cannot access the tables in the following
directory:
/user/customers/regions
Note: The order of the connection properties might vary depending on the tool where you view them.
Property Description
Name The name of the connection. The name is not case sensitive and must be unique
within the domain. You can change this property after you create the connection.
The name cannot exceed 128 characters, contain spaces, or contain the following
special characters:
~ ` ! $ % ^ & * ( ) - + = { [ } ] | \ : ; " ' < , > . ? /
ID String that the Data Integration Service uses to identify the connection. The ID is
not case sensitive. It must be 255 characters or less and must be unique in the
domain. You cannot change this property after you create the connection. Default
value is the connection name.
Description The description of the connection. The description cannot exceed 4000
characters.
Location The domain where you want to create the connection. Not valid for the Analyst
tool.
LDAP username LDAP user name of the user that the Data Integration Service impersonates to run
mappings on a Hadoop cluster. The user name depends on the JDBC connection
string that you specify in the Metadata Connection String or Data Access
Connection String for the native environment.
If the Hadoop cluster uses Kerberos authentication, the principal name for the
JDBC connection string and the user name must be the same. Otherwise, the user
name depends on the behavior of the JDBC driver. With Hive JDBC driver, you can
specify a user name in many ways and the user name can become a part of the
JDBC URL.
If the Hadoop cluster does not use Kerberos authentication, the user name
depends on the behavior of the JDBC driver.
If you do not specify a user name, the Hadoop cluster authenticates jobs based on
the following criteria:
- The Hadoop cluster does not use Kerberos authentication. It authenticates jobs
based on the operating system profile user name of the machine that runs the
Data Integration Service.
- The Hadoop cluster uses Kerberos authentication. It authenticates jobs based
on the SPN of the Data Integration Service. LDAP username will be ignored.
Environment SQL SQL commands to set the Hadoop environment. In native environment type, the
Data Integration Service executes the environment SQL each time it creates a
connection to a Hive metastore. If you use the Hive connection to run profiles on
a Hadoop cluster, the Data Integration Service executes the environment SQL at
the beginning of each Hive session.
The following rules and guidelines apply to the usage of environment SQL in both
connection modes:
- Use the environment SQL to specify Hive queries.
- Use the environment SQL to set the classpath for Hive user-defined functions
and then use environment SQL or PreSQL to specify the Hive user-defined
functions. You cannot use PreSQL in the data object properties to specify the
classpath. If you use Hive user-defined functions, you must copy the .jar files to
the following directory:
<Informatica installation directory>/services/shared/hadoop/
<Hadoop distribution name>/extras/hive-auxjars
- You can use environment SQL to define Hadoop or Hive parameters that you
want to use in the PreSQL commands or in custom queries.
- If you use multiple values for the Environment SQL property, ensure that there is
no space between the values.
SQL Identifier Character The type of character used to identify special characters and reserved SQL
keywords, such as WHERE. The Data Integration Service places the selected
character around special characters and reserved SQL keywords. The Data
Integration Service also uses this character for the Support mixed-case
identifiers property.
Property Description
JDBC Driver Name of the Hive JDBC driver class. If you leave this option blank, the Developer tool uses the
Class Name default Apache Hive JDBC driver shipped with the distribution. If the default Apache Hive JDBC
driver does not fit your requirements, you can override the Apache Hive JDBC driver with a third-
party Hive JDBC driver by specifying the driver class name.
Metadata The JDBC connection URI used to access the metadata from the Hadoop server.
Connection You can use PowerExchange for Hive to communicate with a HiveServer service or HiveServer2
String service. To connect to HiveServer, specify the connection string in the following format:
jdbc:hive2://<hostname>:<port>/<db>
Where
- <hostname> is name or IP address of the machine on which HiveServer2 runs.
- <port> is the port number on which HiveServer2 listens.
- <db> is the database name to which you want to connect. If you do not provide the database
name, the Data Integration Service uses the default database details.
To connect to HiveServer2, use the connection string format that Apache Hive implements for
that specific Hadoop Distribution. For more information about Apache Hive connection string
formats, see the Apache Hive documentation.
For user impersonation, you must add hive.server2.proxy.user=<xyz> to the JDBC
connection URI. If you do not configure user impersonation, the current user's credentials are
used connect to the HiveServer2.
If the Hadoop cluster uses SSL or TLS authentication, you must add ssl=true to the JDBC
connection URI. For example: jdbc:hive2://<hostname>:<port>/<db>;ssl=true
If you use self-signed certificate for SSL or TLS authentication, ensure that the certificate file is
available on the client machine and the Data Integration Service machine. For more information,
see theData Engineering Integration Guide.
Bypass Hive JDBC driver mode. Select the check box to use the embedded JDBC driver mode.
JDBC Server To use the JDBC embedded mode, perform the following tasks:
- Verify that Hive client and Informatica services are installed on the same machine.
- Configure the Hive connection properties to run mappings on a Hadoop cluster.
If you choose the non-embedded mode, you must configure the Data Access Connection String.
Informatica recommends that you use the JDBC embedded mode.
Fine Grained When you select the option to observe fine grained authorization in a Hive source, the mapping
Authorization observes the following:
- Row and column level restrictions. Applies to Hadoop clusters where Sentry or Ranger security
modes are enabled.
- Data masking rules. Applies to masking rules set on columns containing sensitive data by
Dynamic Data Masking.
If you do not select the option, the Blaze and Spark engines ignore the restrictions and masking
rules, and results include restricted or sensitive data.
Data Access The connection string to access data from the Hadoop data store. To connect to HiveServer,
Connection specify the non-embedded JDBC mode connection string in the following format:
String
jdbc:hive2://<hostname>:<port>/<db>
Where
- <hostname> is name or IP address of the machine on which HiveServer2 runs.
- <port> is the port number on which HiveServer2 listens.
- <db> is the database to which you want to connect. If you do not provide the database name,
the Data Integration Service uses the default database details.
To connect to HiveServer2, use the connection string format that Apache Hive implements for the
specific Hadoop Distribution. For more information about Apache Hive connection string formats,
see the Apache Hive documentation.
For user impersonation, you must add hive.server2.proxy.user=<xyz> to the JDBC
connection URI. If you do not configure user impersonation, the current user's credentials are
used connect to the HiveServer2.
If the Hadoop cluster uses SSL or TLS authentication, you must add ssl=true to the JDBC
connection URI. For example: jdbc:hive2://<hostname>:<port>/<db>;ssl=true
If you use self-signed certificate for SSL or TLS authentication, ensure that the certificate file is
available on the client machine and the Data Integration Service machine. For more information,
see the Data Engineering Integration Guide.
Hive Staging HDFS directory for Hive staging tables. You must grant execute permission to the Hadoop
Directory on impersonation user and the mapping impersonation users.
HDFS This option is applicable and required when you write data to a Hive target in the native
environment.
Note: The order of the connection properties might vary depending on the tool where you view them.
Property Description
Name Name of the connection. The name is not case sensitive and must be unique within the domain.
The name cannot exceed 128 characters, contain spaces, or contain the following special
characters:
~ ` ! $ % ^ & * ( ) - + = { [ } ] | \ : ; " ' < , > . ? /
ID String that the Data Integration Service uses to identify the connection. The ID is not case
sensitive. It must be 255 characters or less and must be unique in the domain. You cannot
change this property after you create the connection. Default value is the connection name.
Username User name to connect to the web service. Enter a user name if you enable HTTP authentication
or WS-Security.
If the Web Service Consumer transformation includes WS-Security ports, the transformation
receives a dynamic user name through an input port. The Data Integration Service overrides the
user name defined in the connection.
Password Password for the user name. Enter a password if you enable HTTP authentication or WS-
Security.
If the Web Service Consumer transformation includes WS-Security ports, the transformation
receives a dynamic password through an input port. The Data Integration Service overrides the
password defined in the connection.
End Point URL URL for the web service that you want to access. The Data Integration Service overrides the URL
defined in the WSDL file.
If the Web Service Consumer transformation includes an endpoint URL port, the transformation
dynamically receives the URL through an input port. The Data Integration Service overrides the
URL defined in the connection.
Timeout Number of seconds that the Data Integration Service waits for a response from the web service
provider before it closes the connection. Specify a timeout value between 1 and 10,000 seconds.
HTTP Type of user authentication over HTTP. Select one of the following values:
Authentication - None. No authentication.
Type - Automatic. The Data Integration Service chooses the authentication type of the web service
provider.
- Basic. Requires you to provide a user name and password for the domain of the web service
provider. The Data Integration Service sends the user name and the password to the web
service provider for authentication.
- Digest. Requires you to provide a user name and password for the domain of the web service
provider. The Data Integration Service generates an encrypted message digest from the user
name and password and sends it to the web service provider. The provider generates a
temporary value for the user name and password and stores it in the Active Directory on the
Domain Controller. It compares the value with the message digest. If they match, the web
service provider authenticates you.
- NTLM. Requires you to provide a domain name, server name, or default user name and
password. The web service provider authenticates you based on the domain you are
connected to. It gets the user name and password from the Windows Domain Controller and
compares it with the user name and password that you provide. If they match, the web service
provider authenticates you. NTLM authentication does not store encrypted passwords in the
Active Directory on the Domain Controller.
Trust Certificates File containing the bundle of trusted certificates that the Data Integration Service uses when
File authenticating the SSL certificate of the web service. Enter the file name and full directory path.
Default is <Informatica installation directory>/services/shared/bin/ca-
bundle.crt.
Client Certificate Client certificate that a web service uses when authenticating a client. Specify the client
File Name certificate file if the web service needs to authenticate the Data Integration Service.
Client Certificate Password for the client certificate. Specify the client certificate password if the web service
Password needs to authenticate the Data Integration Service.
Client Certificate Format of the client certificate file. Select one of the following values:
Type - PEM. Files with the .pem extension.
- DER. Files with the .cer or .der extension.
Specify the client certificate type if the web service needs to authenticate the Data Integration
Service.
Private Key File Private key file for the client certificate. Specify the private key file if the web service needs to
Name authenticate the Data Integration Service.
Private Key Password for the private key of the client certificate. Specify the private key password if the web
Password service needs to authenticate the Data Integration Service.
Private Key Type Type of the private key. PEM is the supported type.
Note: The order of the connection properties might vary depending on the tool where you view them.
Property Description
Name Name of the connection. The name is not case sensitive and must be unique within the
domain. The name cannot exceed 128 characters, contain spaces, or contain the
following special characters:
~ ` ! $ % ^ & * ( ) - + = { [ } ] | \ : ; " ' < , > . ? /
ID String that the Data Integration Service uses to identify the connection. The ID is not
case sensitive. It must be 255 characters or less and must be unique in the domain.
You cannot change this property after you create the connection. Default value is the
connection name.
Description The description of the connection. The description cannot exceed 765 characters.
Pass-through security Enables pass-through security for the connection. When you enable pass-through
enabled security for a connection, the domain uses the client user name and password to log
into the corresponding database, instead of the credentials defined in the connection
object.
Connection String for data The DB2 connection URL used to access metadata from the database.
access dbname
Where dbname is the alias configured in the DB2 client.
AdvancedJDBCSecurityOpti Database parameters for metadata access to a secure database. Informatica treats the
ons value of the AdvancedJDBCSecurityOptions field as sensitive data and stores the
parameter string encrypted.
To connect to a secure database, include the following parameters:
- EncryptionMethod. Required. Indicates whether data is encrypted when transmitted
over the network. This parameter must be set to SSL.
- ValidateServerCertificate. Optional. Indicates whether Informatica validates the
certificate that is sent by the database server.
If this parameter is set to True, Informatica validates the certificate that is sent by
the database server. If you specify the HostNameInCertificate parameter,
Informatica also validates the host name in the certificate.
If this parameter is set to false, Informatica does not validate the certificate that is
sent by the database server. Informatica ignores any truststore information that you
specify.
- HostNameInCertificate. Optional. Host name of the machine that hosts the secure
database. If you specify a host name, Informatica validates the host name included
in the connection string against the host name in the SSL certificate.
- cryptoProtocolVersion. Optional. If you enable TLS for the IBM DB2 instance, set the
cryptoProtocolVersion parameter as follows:
cryptoProtocolVersion=TLSv<version number>.
For example, cryptoProtocolVersion=TLSv1.2
Note: The version number must be the same as the TLS version you configured for
the server.
- TrustStore. Required. Path and file name of the truststore file.
Note: If you configure SSL or TLS and specify only the file name, you must copy the
truststore file to the following directory to test the connection: <Informatica
server installation directory>/tomcat/bin
- TrustStorePassword. Required. Password for the truststore file for the secure
database.
Note: Informatica appends the secure JDBC parameters to the connection string. If you
include the secure JDBC parameters directly to the connection string, do not enter any
parameters in the AdvancedJDBCSecurityOptions field.
Data Access Properties: The connection string used to access data from the database.
Connection String For IBM DB2 this is <database name>
Code Page The code page used to read from a source database or to write to a target database or
file.
Environment SQL SQL commands to set the database environment when you connect to the database.
The Data Integration Service runs the connection environment SQL each time it
connects to the database.
Transaction SQL SQL commands to set the database environment when you connect to the database.
The Data Integration Service runs the transaction environment SQL at the beginning of
each transaction.
SQL Identifier Character Type of character that the database uses to enclose delimited identifiers in SQL
queries. The available characters depend on the database type.
Select (None) if the database uses regular identifiers. When the Data Integration
Service generates SQL queries, the service does not place delimited characters around
any identifiers.
Select a character if the database uses delimited identifiers. When the Data Integration
Service generates SQL queries, the service encloses delimited identifiers within this
character.
Support Mixed-case Enable if the database uses case-sensitive identifiers. When enabled, the Data
Identifiers Integration Service encloses all identifiers within the character selected for the SQL
Identifier Character property.
When the SQL Identifier Character property is set to none, the Support Mixed-case
Identifiers property is disabled.
ODBC Provider ODBC. The type of database to which ODBC connects. For pushdown optimization,
specify the database type to enable the Data Integration Service to generate native
database SQL. The options are:
- Other
- Sybase
- Microsoft_SQL_Server
Default is Other.
Note: The order of the connection properties might vary depending on the tool where you view them.
Property Description
Name Name of the connection. The name is not case sensitive and must be unique within the domain.
The name cannot exceed 128 characters, contain spaces, or contain the following special
characters:
~ ` ! $ % ^ & * ( ) - + = { [ } ] | \ : ; " ' < , > . ? /
ID String that the Data Integration Service uses to identify the connection. The ID is not case
sensitive. It must be 255 characters or less and must be unique in the domain. You cannot change
this property after you create the connection. Default value is the connection name.
Description The description of the connection. The description cannot exceed 255 characters.
Password A password for the specified user name or a valid PowerExchange passphrase.
A PowerExchange passphrase can be from 9 to 31 characters in length and can contain the
following characters:
- Uppercase and lowercase letters
- The numbers 0 to 9
- Spaces
- The following special characters:
’ - ; # \ , . / ! % & * ( ) _ + { } : @ | < > ?
Note: The first character is an apostrophe.
Passphrases cannot include single quotation marks (‘), double quotation marks (“), or currency
symbols.
To use passphrases, ensure that the PowerExchange Listener runs with a security setting of
SECURITY=(1,N) or higher in the DBMOVER member. For more information, see "SECURITY
Statement" in the PowerExchange Reference Manual.
Location Node name for the location of the PowerExchange Listener that connects to DB2. The node name is
defined in the first parameter of the NODE statement in the PowerExchange dbmover.cfg
configuration file.
Environment SQL commands to set the database environment when you connect to the database. The Data
SQL Integration Service executes the connection environment SQL each time it connects to the
database.
Database file Specifies the i5/OS database file override in the following format:
overrides
from_file/to_library/to_file/to_member
Where:
- from_file is the file to be overridden.
- to_library is the new library to use.
- to_file is the file in the new library to use.
- to_member is optional and is the member in the new library and file to use. *FIRST is used if
nothing is specified.
You can specify up to eight unique file overrides on a single connection. A single override applies
to a single source or target. When you specify more than one file override, enclose the string of file
overrides in double quotes (") and include a space between each file override.
Note: If you specify both Library List and Database File Overrides and a table exists in both, the
Database File Overrides value takes precedence.
Library list List of libraries that PowerExchange searches to qualify the table name for Select, Insert, Delete, or
Update statements. PowerExchange searches the list if the table name is unqualified.
Separate libraries with commas.
Note: If you specify both Library List and Database File Overrides and a table exists in both, the
Database File Overrides value takes precedence.
Code Page The code page used to read from a source database or write to a target database or file.
SQL Identifier Type of character that the database uses to enclose delimited identifiers in SQL queries. The
character to available characters depend on the database type.
use Select (None) if the database uses regular identifiers. When the Data Integration Service generates
SQL queries, the service does not place delimited characters around any identifiers.
Select a character if the database uses delimited identifiers. When the Data Integration Service
generates SQL queries, the service encloses delimited identifiers within this character.
Support mixed Enable if the database uses case-sensitive identifiers. When enabled, the Data Integration Service
case identifiers encloses all identifiers within the character selected for the SQL Identifier Character property.
When the SQL Identifier Character property is set to none, the Support Mixed-case Identifiers
property is disabled.
Isolation level Commit scope of the transaction. Select one of the following options:
- None
- CS. Cursor stability.
- RR. Repeatable Read.
- CHG. Change.
- ALL
Default is CS.
Encryption type Optional. The type of encryption that the Data Integration Service uses. Select one of the following
options:
- None
- AES
Default is None.
Note: Informatica recommends that you use Secure Sockets Layer (SSL) authentication instead of
configuring the Encryption type and Encryption level connection properties. SSL authentication
provides stricter security and is used by several Informatica products.
For more information about implementing SSL authentication in a PowerExchange network, see the
PowerExchange Reference Manual.
Encryption level If you select AES for Encryption Type, select one of the following options to indicate the encryption
level that the Data Integration Service uses:
- 1. Use a128-bit encryption key.
- 2. Use a 192-bit encryption key.
- 3. Use a 256-bit encryption key.
If you do not select AES for Encryption Type, this option is ignored.
Default is 1.
Pacing size Optional. Amount of data that the source system can pass to the PowerExchange Listener. Set the
pacing size if an external application, database, or the Data Integration Service node is a
bottleneck. User lower values for faster performance.
The minimum value and default value is 0. A value of 0 provides the best performance.
Interpret as Optional. Select this option to express the pacing size as a number of rows. Clear this option to
rows express the pacing size in kilobytes. By default, this option is not selected and the pacing size is in
kilobytes.
Compression Optional. Select this option to enable source data compression. By compressing data, you can
decrease the amount of data that Informatica applications send over the network. By default, this
option is not selected and compression is disabled.
Array size Optional. The number of records in the storage array for the worker threads. This option is
applicable when you set the Worker Threads option to a value greater than 0. Valid values are 25 to
5000. Default is 25.
Write mode Optional. Mode in which the Data Integration Service sends data to the PowerExchange Listener.
Select one of the following write modes:
- CONFIRMWRITEON. Sends data to the PowerExchange Listener and waits for a response before
sending more data. Select this option when error recovery is a priority. However, this option
might degrade performance.
- CONFIRMWRITEOFF. Sends data to the PowerExchange Listener without waiting for a response.
Use this option if you can reload the target table when an error occurs.
- ASYNCHRONOUSWITHFAULTTOLERANCE. Sends data to the PowerExchange Listener without
waiting for a response. This option also enables error detection. This option combines the speed
of CONFIRMWRITEOFF and the data integrity of CONFIRMWRITEON.
Default is CONFIRMWRITEON.
Reject file Overrides the default prefix of PWXR for the reject file. PowerExchange creates the reject file on
the target machine when the write mode is ASYNCHRONOUSWITHFAULTTOLERANCE. Enter
PWXDISABLE to prevent the creation of the reject files.
Note: The order of the connection properties might vary depending on the tool where you view them.
Property Description
Name Name of the connection. The name is not case sensitive and must be unique within the domain.
The name cannot exceed 128 characters, contain spaces, or contain the following special
characters:
~ ` ! $ % ^ & * ( ) - + = { [ } ] | \ : ; " ' < , > . ? /
ID String that the Data Integration Service uses to identify the connection. The ID is not case
sensitive. It must be 255 characters or less and must be unique in the domain. You cannot change
this property after you create the connection. Default value is the connection name.
Description Description of the connection. The description cannot exceed 255 characters.
Password Password for the specified user name or a valid PowerExchange passphrase.
A PowerExchange passphrase can be from 9 to 128 characters in length and can contain the
following characters:
- Uppercase and lowercase letters
- The numbers 0 to 9
- Spaces
- The following special characters:
’ - ; # \ , . / ! % & * ( ) _ + { } : @ | < > ?
Note: The first character is an apostrophe.
Passphrases cannot include single quotation marks (‘), double quotation marks (“), or currency
symbols.
To use passphrases, ensure that the PowerExchange Listener runs with a security setting of
SECURITY=(1,N) or higher in the DBMOVER member. For more information, see "SECURITY
Statement" in the PowerExchange Reference Manual.
The allowable characters in the IBM IRRPHREX exit do not affect the allowable characters in
PowerExchange passphrases.
Note: A valid RACF passphrase can be up to 100 characters in length. PowerExchange truncates
passphrases longer than 100 characters when passing them to RACF for validation.
Location Node name for the location of the PowerExchange Listener that connects to DB2. The node name is
defined in the first parameter of the NODE statement in the PowerExchange dbmover.cfg
configuration file.
Environment SQL commands to set the database environment when you connect to the database. The Data
SQL Integration Service executes the connection environment SQL each time it connects to the
database.
Correlation ID Value to be concatenated to prefix PWX to form the DB2 correlation ID for DB2 requests.
Code Page Code page used to read from a source database or write to a target database or file.
SQL identifier Type of character that the database uses to enclose delimited identifiers in SQL queries. The
character to available characters depend on the database type.
use Select (None) if the database uses regular identifiers. When the Data Integration Service generates
SQL queries, the service does not place delimited characters around any identifiers.
Select a character if the database uses delimited identifiers. When the Data Integration Service
generates SQL queries, the service encloses delimited identifiers within this character.
Support mixed Enable if the database uses case-sensitive identifiers. When enabled, the Data Integration Service
case identifiers encloses all identifiers within the character selected for the SQL Identifier Character property.
When the SQL Identifier Character property is set to none, the Support Mixed-case Identifiers
property is disabled.
Encryption type Optional. The type of encryption that the Data Integration Service uses. Select one of the following
options:
- None
- AES
Default is None.
Note: Informatica recommends that you use Secure Sockets Layer (SSL) authentication instead of
configuring the Encryption Type and Level connection properties. SSL authentication provides
stricter security and is used by several Informatica products.
For more information about implementing SSL authentication in a PowerExchange network, see the
PowerExchange Reference Manual.
Encryption level If you select AES for Encryption Type, select one of the following options to indicate the encryption
level that the Data Integration Service uses:
- 1. Use a128-bit encryption key.
- 2. Use a 192-bit encryption key.
- 3. Use a 256-bit encryption key.
If you do not select AES for Encryption Type, this option is ignored.
Default is 1.
Pacing size Optional. Amount of data that the source system can pass to the PowerExchange Listener. Set the
pacing size if an external application, database, or the Data Integration Service node is a
bottleneck. User lower values for faster performance.
The minimum value and default value is 0. A value of 0 provides the best performance.
Interpret as Optional. Select this option to express the pacing size as a number of rows. Clear this option to
rows express the pacing size in kilobytes. By default, this option is not selected and the pacing size is in
kilobytes.
Compression Optional. Select this option to enable source data compression. By compressing data, you can
decrease the amount of data that Informatica applications send over the network. By default, this
option is not selected and compression is disabled.
Offload Optional. Controls whether to offload some bulk data processing from the source machine to the
processing Data Integration Service machine. Select one of the following options:
- AUTO. The Data Integration Service determines whether to use offload processing.
- Yes. Use offload processing.
- No. Do not use offload processing.
Default is No.
Worker threads Optional. Number of threads that the Data Integration Service uses to process bulk data when
offload processing is enabled. For optimal performance, this value should not exceed the number
of available processors on the Data Integration Service machine. Valid values are 1 through 64.
Default is 0, which disables multithreading.
Array size Optional. The number of records in the storage array for the worker threads. This option is
applicable when you set the Worker Threads option to a value greater than 0. Valid values are 1 to
5000. Default is 25.
Write mode Mode in which the Data Integration Service sends data to the PowerExchange Listener. Configure
one of the following write modes:
- CONFIRMWRITEON. Sends data to the PowerExchange Listener and waits for a response before
sending more data. Select if error recovery is a priority. This option might decrease
performance.
- CONFIRMWRITEOFF. Sends data to the PowerExchange Listener without waiting for a response.
Use this option when you can reload the target table if an error occurs.
- ASYNCHRONOUSWITHFAULTTOLERANCE. Sends data to the PowerExchange Listener without
waiting for a response. This option also provides the ability to detect errors. This provides the
speed of Confirm Write Off with the data integrity of Confirm Write On.
Default is CONFIRMWRITEON.
Reject file Overrides the default prefix of PWXR for the reject file. PowerExchange creates the reject file on
the target machine when the write mode is ASYNCHRONOUSWITHFAULTTOLERANCE. Enter
PWXDISABLE to prevent the creation of reject files.
Option Description
Location Node name for the location of the PowerExchange Listener that connects to IMS. The node
name is defined in the first parameter of the NODE statement in the PowerExchange
dbmover.cfg configuration file.
Password Password for the specified database user name or a valid PowerExchange passphrase.
A PowerExchange passphrase can be from 9 to 128 characters in length and can contain the
following characters:
- Uppercase and lowercase letters
- The numbers 0 to 9
- Spaces
- The following special characters:
’ - ; # \ , . / ! % & * ( ) _ + { } : @ | < > ?
Note: The first character is an apostrophe.
Passphrases cannot include single quotation marks (‘), double quotation marks (“), or
currency symbols.
The allowable characters in the IBM IRRPHREX exit do not affect the allowable characters in
PowerExchange passphrases.
Note: A valid RACF passphrase can be up to 100 characters in length. PowerExchange
truncates passphrases longer than 100 characters when passing them to RACF for validation.
To use passphrases for IMS connections, ensure that the following requirements are met:
- The PowerExchange Listener must run with a security setting of SECURITY=(1,N) or higher
in the DBMOVER member. For more information, see "SECURITY Statement" in the
PowerExchange Reference Manual.
- You must configure ODBA access to IMS as described in the PowerExchange Navigator User
Guide.
- You must use IMS data maps that specify IMS ODBA as the access method. Do not use
data maps that specify the DL/1 BATCH access method because this access method
requires the use of netport jobs, which do not support passphrases.
- The IMS database must be online in the IMS control region to use ODBA access to IMS.
Code page Required. Name of the code page to use for reading from or writing to the data source.
Usually, this value is an ISO code page name, such as ISO-8859-6.
Encryption type The type of encryption that the Data Integration Service uses. Select one of the following
options:
- None
- AES
Default is None.
Note: Informatica recommends that you use Secure Sockets Layer (SSL) authentication
instead of configuring the Encryption Type and Level connection properties. SSL
authentication provides stricter security and is used by several Informatica products.
For more information about implementing SSL authentication in a PowerExchange network,
see the PowerExchange Reference Manual.
[Encryption] Level If you select AES for Encryption Type, select one of the following options to indicate the
encryption level that the Data Integration Service uses:
- 1. Use a128-bit encryption key.
- 2. Use a 192-bit encryption key.
- 3. Use a 256-bit encryption key.
If you do not select AES for Encryption Type, this option is ignored.
Default is 1.
Pacing size Optional. Amount of data that the source system can pass to the PowerExchange Listener.
Set the pacing size if an external application, database, or the Data Integration Service node
is a bottleneck. User lower values for faster performance.
The minimum value and default value is 0. A value of 0 provides the best performance.
Interpret as rows Optional. Select this option to express the pacing size as a number of rows. Clear this option
to express the pacing size in kilobytes. By default, this option is not selected and the pacing
size is in kilobytes.
Compression Optional. Select this option to enable source data compression. By compressing data, you
can decrease the amount of data that Informatica applications send over the network. By
default, this option is not selected and compression is disabled.
Offload processing Optional. Controls whether to offload some bulk data processing from the source machine to
the Data Integration Service machine. Select one of the following options:
- AUTO. The Data Integration Service determines whether to use offload processing.
- Yes. Use offload processing.
- No. Do not use offload processing.
Default is AUTO.
Worker threads Optional. Number of threads that the Data Integration Service uses to process bulk data when
offload processing is enabled. For optimal performance, this value should not exceed the
number of available processors on the Data Integration Service machine. Valid values are 1
through 64. Default is 0, which disables multithreading.
Array size Optional. The number of records in the storage array for the worker threads. This option is
applicable when you set the Worker Threads option to a value greater than 0. Valid values are
1 to 5000. Default is 25.
Write mode Optional. Mode in which Data Integration Service sends data to the PowerExchange Listener.
Select one of the following write modes:
- CONFIRMWRITEON. Sends data to the PowerExchange Listener and waits for a response
before sending more data. Select this option when error recovery is a priority. However, this
option might degrade performance.
- CONFIRMWRITEOFF. Sends data to the PowerExchange Listener without waiting for a
response. Use this option if you can reload the target table when an error occurs.
- ASYNCHRONOUSWITHFAULTTOLERANCE. Sends data to the PowerExchange Listener
without waiting for a response. This option also enables error detection. This option
combines the speed of CONFIRMWRITEOFF and the data integrity of CONFIRMWRITEON.
Default is CONFIRMWRITEON.
Note: The order of the connection properties might vary depending on the tool where you view them.
Property Description
Name Name of the connection. The name is not case sensitive and must be unique within the domain. The
name cannot exceed 128 characters, contain spaces, or contain the following special characters:
~ ` ! $ % ^ & * ( ) - + = { [ } ] | \ : ; " ' < , > . ? /
ID String that the Data Integration Service uses to identify the connection. The ID is not case sensitive.
It must be 255 characters or less and must be unique in the domain. You cannot change this property
after you create the connection. Default value is the connection name.
Description The description of the connection. The description cannot exceed 765 characters.
Connection Connection string to connect to the database. Use the following connection string:
String
jdbc:<subprotocol>:<subname>
For more information about the connection string to use with specific drivers, see the vendor
documentation.
Environment Optional. Enter SQL commands to set the database environment when you connect to the database.
SQL The Data Integration Service executes the connection environment SQL each time it connects to the
database.
Note: If you enable Sqoop, Sqoop ignores this property.
Transaction Optional. Enter SQL commands to set the database environment when you connect to the database.
SQL The Data Integration Service executes the transaction environment SQL at the beginning of each
transaction.
Note: If you enable Sqoop, Sqoop ignores this property.
SQL Identifier Type of character that the database uses to enclose delimited identifiers in SQL queries. The
Character available characters depend on the database type.
Select (None) if the database uses regular identifiers. When the Data Integration Service generates
SQL queries, the service does not place delimited characters around any identifiers.
Select a character if the database uses delimited identifiers. When the Data Integration Service
generates SQL queries, the service encloses delimited identifiers within this character.
Note: If you enable Sqoop, Sqoop ignores this property.
Support Enable if the database uses case-sensitive identifiers. When enabled, the Data Integration Service
Mixed-case encloses all identifiers within the character selected for the SQL Identifier Character property.
Identifiers When the SQL Identifier Character property is set to none, the Support Mixed-case Identifiers
property is disabled.
Note: If you enable Sqoop, Sqoop honors this property when you generate and execute a DDL script
to create or replace a target at run time. In all other scenarios, Sqoop ignores this property.
Use Sqoop Enables Sqoop connectivity for the data object that uses the JDBC connection. The Data Integration
Connector Service runs the mapping in the Hadoop run-time environment through Sqoop.
You can configure Sqoop connectivity for relational data objects, customized data objects, and
logical data objects that are based on a JDBC-compliant database.
Select Sqoop v1.x to enable Sqoop connectivity.
Default is None.
Sqoop Enter the arguments that Sqoop must use to connect to the database. Separate multiple arguments
Arguments with a space.
To run the mapping on the Blaze engine with the Teradata Connector for Hadoop (TDCH) specialized
connectors for Sqoop, you must define the TDCH connection factory class in the Sqoop arguments.
The connection factory class varies based on the TDCH Sqoop Connector that you want to use.
- To use Cloudera Connector Powered by Teradata, configure the following Sqoop argument:
-
Dsqoop.connection.factories=com.cloudera.connector.teradata.TeradataManager
Factory
- To use Hortonworks Connector for Teradata (powered by the Teradata Connector for Hadoop),
configure the following Sqoop argument:
-
Dsqoop.connection.factories=org.apache.sqoop.teradata.TeradataManagerFactor
y
To run the mapping on the Spark engine, you do not need to define the TDCH connection factory
class in the Sqoop arguments. The Data Integration Service invokes the Cloudera Connector Powered
by Teradata and Hortonworks Connector for Teradata (powered by the Teradata Connector for
Hadoop) by default.
Note: To run the mapping with a generic JDBC connector instead of the specialized Cloudera or
Hortonworks connector, you must define the --driver and --connection-manager Sqoop arguments in
the JDBC connection. If you define the --driver and --connection-manager arguments in the Read or
Write transformation of the mapping, Sqoop ignores the arguments.
If you do not enter Sqoop arguments, the Data Integration Service constructs the Sqoop command
based on the JDBC connection properties.
Note: The order of the connection properties might vary depending on the tool where you view them.
Property Description
Name The name of the connection. The name is not case sensitive and must be unique within the domain. You
can change this property after you create the connection. The name cannot exceed 128 characters,
contain spaces, or contain the following special characters:~ ` ! $ % ^ & * ( ) - + = { [ } ] | \ : ; " ' < , > . ? /
ID String that the Data Integration Service uses to identify the connection. The ID is not case sensitive. It
must be 255 characters or less and must be unique in the domain. You cannot change this property
after you create the connection. Default value is the connection name.
Description The description of the connection. The description cannot exceed 4,000 characters.
The Details tab contains the connection attributes of the JDBC V2 connection. The following table describes
the connection attributes:
Property Description
Support Enable if the database uses case-sensitive identifiers. When enabled, the Data Integration Service
Mixed-case encloses all identifiers within the character selected for the SQL Identifier Character property.
Identifiers For example, Aurora PostgreSQL database supports mixed-cased characters. You must enable this
property to connect to the Aurora PostgreSQL database.
When the SQL Identifier Character property is set to none, the Support Mixed-case Identifiers
property is disabled.
SQL Type of character that the database uses to enclose delimited identifiers in SQL queries. The
Identifier available characters depend on the database type.
Character Select (None) if the database uses regular identifiers. When the Data Integration Service generates
SQL queries, the service does not place delimited characters around any identifiers.
Select a character if the database uses delimited identifiers. When the Data Integration Service
generates SQL queries, the service encloses delimited identifiers within this character.
Note: Select SQL Identifier Character as None when you specify the SAP HANA Database subtype.
Property Description
Name The name of the connection. The name is not case sensitive and must be unique
within the domain. It cannot exceed 128 characters, contain spaces, or contain
the following special characters:
~ ` ! $ % ^ & * ( ) - + = { [ } ] | \ : ; " ' < , > . ? /
ID The string that the Data Integration Service uses to identify the connection. The
ID is not case sensitive. It must be 255 characters or less and must be unique in
the domain. You cannot change this property after you create the connection.
Default value is the connection name.
Description The description of the connection. The description cannot exceed 765
characters.
Location The Informatica domain where you want to create the connection.
Environment Name of the JD Edwards EnterpriseOne environment you want to connect to.
Driver Class Name The following list provides the driver class name that you can enter for the
applicable database type:
- DataDirect JDBC driver class name for Oracle:
com.informatica.jdbc.oracle.OracleDriver
- DataDirect JDBC driver class name for IBM DB2:
com.informatica.jdbc.db2.DB2Driver
- DataDirect JDBC driver class name for Microsoft SQL Server:
com.informatica.jdbc.sqlserver.SQLServerDriver
For more information about which driver class to use with specific databases,
see the vendor documentation.
Connection String The connection string to connect to the database. Use the following connection
string:
The JDBC connection string uses the following syntax:
- For Oracle: jdbc:informatica:oracle://<host
name>:<port>,ServiceName=<db service name>
- For DB2: jdbc:informatica:db2://<host
name>:<port>;databaseName=<db name>
- For Microsoft SQL: jdbc:informatica:sqlserver://<host
name>:<port>;databaseName=<db name>
When you configure a Kafka connection, you configure the following properties:
• The list of Kafka brokers that the connection reads from or writes to.
• The number of seconds the Integration Service attempts to reconnect to the database if the connection
fails.
• Version of the Kafka messaging broker. Configure the Kafka messaging broker version to Apache 0.10.1.1
& above.
Property Description
Name The name of the connection. The name is not case sensitive and must be unique within the domain. You
can change this property after you create the connection. The name cannot exceed 128 characters,
contain spaces, or contain the following special characters:
~ ` ! $ % ^ & * ( ) - + = { [ } ] | \ : ; " ' < , > . ? /
ID The string that the Data Integration Service uses to identify the connection. The ID is not case sensitive.
It must be 255 characters or less and must be unique in the domain. You cannot change this property
after you create the connection. Default value is the connection name.
Description The description of the connection. Enter a string that you can use to identify the connection. The
description cannot exceed 4,000 characters.
Location The domain where you want to create the connection. Select the domain name.
Property Description
Kafka Broker List Comma-separated list of Kafka brokers which maintain the
configuration of the Kafka messaging broker.
To specify a Kafka broker, use the following format:
<IP Address>:<port>
Retry Timeout Number of seconds after which the Integration Service attempts to
reconnect to the Kafka broker to read or write data. If the source or
target is not available for the time you specify, the mapping execution
stops to avoid any data loss.
Kafka Broker Version Configure the Kafka messaging broker version to Apache 0.10.1.1 &
above.
SSL Properties
The following table describes the SSL properties for the Kafka connection:
Property Description
SSL Mode Required. SSL mode indicates the encryption type to use for the connection.
You can choose a mode from the following SSL modes:
- Disabled
- One way
- Two way
SSL TrustStore File Required when One way SSL mode is selected.
Path Absolute path and file name of the SSL truststore file that contains certificates of the
trusted SSL server.
SSL KeyStore File Path Required when Two way SSL mode is selected.
Absolute path and file name of the SSL keystore file that contains private keys and
certificates for the SSL server.
sh infacmd.sh createConnection -dn <domain name> -un <domain user> -pd <domain password> -cn
<connection name> -cid <connection id> -ct Kafka -o
kfkBrkList=<host1:port1>,<host2:port2>,<host3:port3> kafkabrokerversion=<version>
additionalConnectionProperties=<additional properties>
For more information about the CreateConnection command, see the Informatica Command Reference.
Note: The order of the connection properties might vary depending on the tool where you view them.
You can create and manage a Kudu connection in the Administrator tool or the Developer tool. The following
table describes the Kudu connection properties:
Property Description
Name The name of the connection. The name is not case sensitive and must be unique within the domain. You
can change this property after you create the connection. The name cannot exceed 128 characters,
contain spaces, or contain the following special characters: ~ ` ! $ % ^ & * ( ) - + = { [ } ] | \ : ; " ' < , > . ? /
ID String that the Data Integration Service uses to identify the connection. The ID is not case sensitive. It
must be 255 characters or less and must be unique in the domain. You cannot change this property
after you create the connection.
Default value is the connection name.
Description The description of the connection. The description cannot exceed 4,000 characters.
Property Description
Cluster Configuration The Hadoop cluster that you use for the connection.
Property Description
Name The name of the connection. The name is not case sensitive and must be unique
within the domain. It cannot exceed 128 characters, contain spaces, or contain
the following special characters:
~ ` ! $ % ^ & * ( ) - + = { [ } ] | \ : ; " ' < , > . ? /
ID The string that the Data Integration Service uses to identify the connection. The
ID is not case sensitive. It must be 255 characters or less and must be unique in
the domain. You cannot change this property after you create the connection.
Default value is the connection name.
Description The description of the connection. The description cannot exceed 765
characters.
Location The Informatica domain where you want to create the connection.
Anonymous Connection Establishes an anonymous connection with the LDAP directory server. Select
anonymous connection to access a directory server as an anonymous user
without authentication.
Note: You cannot establish an anonymous connection with Active Directory.
User Name The LDAP user name to connect to the LDAP directory server.
Secure Connection Establishes a secure connection with the LDAP directory server through the TLS
protocol.
TrustStore File Name The file name of the truststore that contains the TLS certificate to establish a
secure connection with the LDAP directory server.
Default is infa_truststore.jks.
Required if you select Secure Connection. Contact the LDAP Administrator for
the truststore file name and password.
TrustStore Password The password for the truststore file that contains the SSL certificate.
KeyStore File Name The file name of the keystore that contains the keys and certificates required to
establish a secure communication with the LDAP directory server.
Required if you select Secure Connection. Contact the LDAP Administrator for
the keystore file name and password.
KeyStore Password The password for the keystore file required for secure communication.
Note: The order of the connection properties might vary depending on the tool where you view them.
You can create and manage a Microsoft Azure Blob Storage connection in the Administrator tool or the
Developer tool. The following table describes the Microsoft Azure Blob Storage connection properties:
Property Description
ID String that the Data Integration Service uses to identify the connection. The ID is not
case sensitive. It must be 255 characters or less and must be unique in the domain.
You cannot change this property after you create the connection. Default value is the
connection name.
Property Description
Authorization Type Authorization type. You can select any of the following authorization mechanisms:
- Shared Key Authorization
- Shared Access Signatures
Account Key Microsoft Azure Storage access key. Applies when you select shared key
authorization.
SAS Token SAS URI with SAS token that you generate on Microsoft Azure portal for your
account. Applies when you select shared access signature authorization type.
Note: You must provide a valid SAS URI with a valid SAS token.
Container Name The root container or sub-folders with the absolute path.
Note: To import complex files, specify only the root container.
Endpoint Suffix Type of Microsoft Azure end-points. You can select any of the following end-
points:
- core.windows.net: Default
- core.usgovcloudapi.net: To select the US government Microsoft Azure end-
points
- core.chinacloudapi.cn: Not applicable
The following table describes the Microsoft Azure Cosmos DB connection properties:
Property Description
ID String that the Data Integration Service uses to identify the connection. The ID is not case sensitive.
It must be 255 characters or less and must be unique in the domain. You cannot change this
property after you create the connection. Default value is the connection name.
Description Description of the connection. The description cannot exceed 765 characters.
Location The project or folder in the Model repository where you want to store the Cosmos DB connection.
Key The primary and secondary key to which provides you complete administrative access to the
resources within Microsoft Azure Cosmos DB account.
Database Name of the database that contains the collections from which you want to read or write JSON
documents.
Note: You can find the Cosmos DB URI and Key values in the Keys settings on Azure portal. Contact your
Azure administrator for more details.
Note: The order of the connection properties might vary depending on the tool where you view them.
You can create and manage a Microsoft Azure Data Lake Storage Gen1 connection in the Administrator tool
or the Developer tool. The following table describes the Microsoft Azure Data Lake Storage Gen1 connection
properties:
Property Description
Name The name of the connection. The name is not case sensitive and must be unique within the domain. You
can change this property after you create the connection. The name cannot exceed 128 characters,
contain spaces, or contain the following special characters: ~ ` ! $ % ^ & * ( ) - + = { [ } ] | \ : ; " ' < , > . ? /
ID String that the Data Integration Service uses to identify the connection. The ID is not case sensitive. It
must be 255 characters or less and must be unique in the domain. You cannot change this property
after you create the connection.
Default value is the connection name.
Description The description of the connection. The description cannot exceed 4,000 characters.
Type The connection type. Select Microsoft Azure Data Lake Storage Gen1.
Property Description
ADLS Account Name The name of the Microsoft Azure Data Lake Storage Gen1.
ClientID The ID of your application to complete the OAuth Authentication in the Active Directory.
Client Secret The client secret key to complete the OAuth Authentication in the Active Directory.
Directory The Microsoft Azure Data Lake Storage Gen1 directory that you use to read data or write data.
The default is root directory.
AuthEndpoint The OAuth 2.0 token endpoint from where access code is generated based on based on the
Client ID and Client secret is completed.
For more information about creating a client ID, client secret, and auth end point, contact the Azure
administrator or see Microsoft Azure Data Lake Storage Gen1 documentation.
Note: The order of the connection properties might vary depending on the tool where you view them.
You can create and manage a Microsoft Azure Data Lake Storage Gen2 connection in the Administrator tool
or the Developer tool. The following table describes the Microsoft Azure Data Lake Storage Gen2 connection
properties:
Property Description
Name The name of the connection. The name is not case sensitive and must be unique within the domain. You
can change this property after you create the connection. The name cannot exceed 128 characters,
contain spaces, or contain the following special characters: ~ ` ! $ % ^ & * ( ) - + = { [ } ] | \ : ; " ' < , > . ? /
ID String that the Data Integration Service uses to identify the connection. The ID is not case sensitive. It
must be 255 characters or less and must be unique in the domain. You cannot change this property
after you create the connection.
Default value is the connection name.
Description The description of the connection. The description cannot exceed 4,000 characters.
Type The connection type. Select Microsoft Azure Data Lake Storage Gen2.
Property Description
Account Name The Microsoft Azure Data Lake Storage Gen2 account name or the service name.
Authentication Type Authentication type to access the Microsoft Azure Data Lake Storage Gen2 account.
Select one of the following options:
- Service Principal Authentication. Select to use the client ID, client secret, and tenant ID to
connect to Microsoft Azure Data Lake Storage Gen2.
- Shared Key Authentication. Select to use the account key to connect to Microsoft Azure
Data Lake Storage Gen2.
Note: You cannot use shared key authentication in a streaming mapping.
Client ID Applicable for Service Principal Authentication. The ID of your application to complete the
OAuth Authentication in the Azure Active Directory (AD).
Client Secret Applicable for Service Principal Authentication. The client secret key to complete the OAuth
Authentication in the Azure Active Directory.
Tenant ID Applicable for Service Principal Authentication. The Directory ID of the Azure Active Directory.
Account Key Applicable for Shared Key Authentication. The account key for the Microsoft Azure Data Lake
Storage Gen2 account.
File System Name The name of an existing file system in the Microsoft Azure Data Lake Storage Gen2.
Directory Path The path of an existing directory without the file system name. There is no default directory.
You can select one of the following syntax:
- / for root directory.
- /dir1
- dir1/dir2
Adls Gen2 End-point Type of Microsoft Azure endpoints. You can select any of the following endpoints:
- core.windows.net: Default
- core.usgovcloudapi.net: To select the Azure Government endpoints
For more information about creating a client ID, client secret, tenant ID, and file system name, contact the
Azure administrator or see Microsoft Azure Data Lake Storage Gen2 documentation.
Note: The order of the connection properties might vary depending on the tool where you view them.
Property Description
Name The name of the connection. The name is not case sensitive and must be unique within the domain. You
can change this property after you create the connection. The name cannot exceed 128 characters,
contain spaces, or contain the following special characters: ~ ` ! $ % ^ & * ( ) - + = { [ } ] | \ : ; " ' < , > . ? /
ID String that the Data Integration Service uses to identify the connection. The ID is not case sensitive. It
must be 255 characters or less and must be unique in the domain. You cannot change this property
after you create the connection. Default value is the connection name.
Description The description of the connection. The description cannot exceed 4,000 characters.
Property Description
Azure DW JDBC Microsoft Azure Data Warehouse JDBC connection string. For example, you can enter the
URL following connection string:
jdbc:sqlserver:// <Server>.database.windows.net:1433;database=<Database>
The Administrator can download the URL from Microsoft Azure portal.
Azure DW JDBC User name to connect to the Microsoft Azure SQL Data Warehouse account. You must have
Username permission to read, write, and truncate data in Microsoft Azure SQL Data Warehouse.
Azure DW JDBC Password to connect to the Microsoft Azure SQL Data Warehouse account.
Password
Azure DW Schema Name of the schema in Microsoft Azure SQL Data Warehouse.
Name
Azure Storage Type of Azure storage to stage the files. You can select any of the following storage type:
Type - Azure Blob. Default. To use Microsoft Azure Blob Storage to stage the files.
- ADLS Gen2. To use Microsoft Azure Data Lake Storage Gen2 as storage to stage the files.
Azure Blob Name of the Microsoft Azure Storage account to stage the files.
Account Name
Azure Blob The key that authenticates the access to the Blob storage account.
Account Key
ADLS Gen2 Name of the Microsoft Azure Data Lake Storage Gen2 account to stage the files.
Storage Account
Name
ADLS Gen2 Microsoft Azure Data Lake Storage Gen2 access key to stage the files.
Account Key
VNet Rule Enable to connect to a Microsoft Azure SQL Data Warehouse endpoint residing in a virtual
network (VNet).
Note: The order of the connection properties might vary depending on the tool where you view them.
Property Description
Name Name of the connection. The name is not case sensitive and must be unique within the
domain. The name cannot exceed 128 characters, contain spaces, or contain the
following special characters:
~ ` ! $ % ^ & * ( ) - + = { [ } ] | \ : ; " ' < , > . ? /
ID String that the Data Integration Service uses to identify the connection. The ID is not
case sensitive. It must be 255 characters or less and must be unique in the domain.
You cannot change this property after you create the connection. Default value is the
connection name.
Description The description of the connection. The description cannot exceed 765 characters.
Use trusted connection Enables the application service to use Windows authentication to access the database.
The user name that starts the application service must be a valid Windows user with
access to the database. By default, this option is cleared.
Note: Windows and NTLM authentication is not certified for a Microsoft SQL Server
2017 version hosted on Linux.
User Name The database user name. Required if Microsoft SQL Server uses NTLMv1 or NTLMv2
authentication.
Password The password for the database user name. Required if Microsoft SQL Server uses
NTLMv1 or NTLMv2 authentication.
Pass-through security Enables pass-through security for the connection. When you enable pass-through
enabled security for a connection, the domain uses the client user name and password to log
into the corresponding database, instead of the credentials defined in the connection
object.
Metadata Access Connection string used to access metadata from the database.
Properties: Connection Use the following connection string:
String
jdbc:informatica:sqlserver://<host
name>:<port>;DatabaseName=<database name>
To test the connection with NTLM authentication, include the following parameters in
the connection string:
- AuthenticationMethod. The NTLM authentication version to use.
Note: UNIX supports NTLMv1 and NTLMv2 but not NTLM.
- Domain. The domain that the SQL server belongs to.
The following example shows the connection string for a SQL server that uses NTLMv2
authentication in a NT domain named Informatica.com:
jdbc:informatica:sqlserver://
host01:1433;DatabaseName=SQL1;AuthenticationMethod=ntlm2java;Domai
n=Informatica.com
If you connect with NTLM authentication, you can enable the Use trusted connection
option in the MS SQL Server connection properties. If you connect with NTLMv1 or
NTLMv2 authentication, you must provide the user name and password in the
connection properties.
AdvancedJDBCSecurityOpti Database parameters for metadata access to a secure database. Informatica treats the
ons value of the AdvancedJDBCSecurityOptions field as sensitive data and stores the
parameter string encrypted.
To connect to a secure database, include the following parameters:
- EncryptionMethod. Required. Indicates whether data is encrypted when transmitted
over the network. This parameter must be set to SSL.
- ValidateServerCertificate. Optional. Indicates whether Informatica validates the
certificate that is sent by the database server.
If this parameter is set to True, Informatica validates the certificate that is sent by
the database server. If you specify the HostNameInCertificate parameter,
Informatica also validates the host name in the certificate.
If this parameter is set to false, Informatica does not validate the certificate that is
sent by the database server. Informatica ignores any truststore information that you
specify.
- HostNameInCertificate. Optional. Host name of the machine that hosts the secure
database. If you specify a host name, Informatica validates the host name included
in the connection string against the host name in the SSL certificate.
- cryptoProtocolVersion. Optional. If you enable TLS for the Microsoft SQL Server
instance, set the cryptoProtocolVersion parameter as follows:
cryptoProtocolVersion=TLSv<version number>.
For example, cryptoProtocolVersion=TLSv1.2
Note: The version number must be the same as the TLS version you configured for
the server.
- TrustStore. Required. Path and file name of the truststore file.
Note: If you configure SSL or TLS and specify only the file name, you must copy the
truststore file to the following directory to test the connection: <Informatica
server installation directory>/tomcat/bin
- TrustStorePassword. Required. Password for the truststore file for the secure
database.
Not applicable for ODBC.
Note: Informatica appends the secure JDBC parameters to the connection string. If you
include the secure JDBC parameters directly to the connection string, do not enter any
parameters in the AdvancedJDBCSecurityOptions field.
Data Access Properties: The connection provider that you want to use to connect to the Microsoft SQL Server
Provider Type database.
You can select the following provider types:
- ODBC
- Oldeb(Deprecated)
Default is ODBC.
Note: Although the Microsoft SQL Server connection user interface shows the OLEDB
provider type as deprecated, Informatica supports the OLEDB provider type. For more
information about the OLEDB provider type support statement, see the following
Knowledge Base article KB 522895.
Use DSN Enables the Data Integration Service to use the Data Source Name for the connection.
If you select the Use DSN option, the Data Integration Service retrieves the database
and server names from the DSN.
If you do not select the Use DSN option, you must provide the database and server
names.
Connection String Use the following connection string if you do not enable DSN mode:
<server name>@<database name>
If you enable DSN mode, use the following connection strings:
<DSN Name>
Code Page The code page used to read from a source database or to write to a target database or
file.
Packet Size The packet size used to transmit data. Used to optimize the native drivers for
Microsoft SQL Server.
Schema Name The name of the schema in the database. You must specify the schema name for the
Profiling Warehouse if the schema name is different from the database user name. You
must specify the schema name for the data object cache database if the schema name
is different from the database user name and if you configure user-managed cache
tables.
Note: When you generate a table DDL through a dynamic mapping or through the
Generate and Execute DDL option, the DDL metadata does not include schema name
and owner name properties.
Environment SQL SQL commands to set the database environment when you connect to the database.
The Data Integration Service runs the connection environment SQL each time it
connects to the database.
Transaction SQL SQL commands to set the database environment when you connect to the database.
The Data Integration Service runs the transaction environment SQL at the beginning of
each transaction.
SQL Identifier Character Type of character that the database uses to enclose delimited identifiers in SQL
queries. The available characters depend on the database type.
Select (None) if the database uses regular identifiers. When the Data Integration
Service generates SQL queries, the service does not place delimited characters around
any identifiers.
Select a character if the database uses delimited identifiers. When the Data Integration
Service generates SQL queries, the service encloses delimited identifiers within this
character.
Support Mixed-case Enable if the database uses case-sensitive identifiers. When enabled, the Data
Identifiers Integration Service encloses all identifiers within the character selected for the SQL
Identifier Character property.
When the SQL Identifier Character property is set to none, the Support Mixed-case
Identifiers property is disabled.
ODBC Provider ODBC. The type of database to which ODBC connects. For pushdown optimization,
specify the database type to enable the Data Integration Service to generate native
database SQL. The options are:
- Other
- Sybase
- Microsoft_SQL_Server
Default is Other.
Note: The order of the connection properties might vary depending on the tool where you view them.
Property Description
Name Name of the connection. The name is not case sensitive and must be unique within the domain.
The name cannot exceed 128 characters, contain spaces, or contain the following special
characters:
~ ` ! $ % ^ & * ( ) - + = { [ } ] | \ : ; " ' < , > . ? /
ID String that the Data Integration Service uses to identify the connection. The ID is not case
sensitive. It must be 255 characters or less and must be unique in the domain. You cannot change
this property after you create the connection. Default value is the connection name.
Description Description of the connection. The description cannot exceed 765 characters.
User name User name with the appropriate permissions to access the Netezza database.
JDBC Url JDBC URL that the Developer tool must use when it connects to the Netezza database.
Use the following format:
jdbc:netezza://<hostname>:<port>/<database name>
Connection Name of the ODBC data source that you want to use to connect to the Netezza database.
String
Timeout Number of seconds that the Developer tool waits for a response from the Netezza database before
it closes the connection.
Note: The order of the connection properties might vary depending on the tool where you view them.
Property Description
Name Name of the connection. The name is not case sensitive and must be unique within the domain.
The name cannot exceed 128 characters, contain spaces, or contain the following special
characters:
~ ` ! $ % ^ & * ( ) - + = { [ } ] | \ : ; " ' < , > . ? /
ID String that the Data Integration Service uses to identify the connection. The ID is not case
sensitive. It must be 255 characters or less and must be unique in the domain. You cannot change
this property after you create the connection. Default value is the connection name.
Description Description of the connection. The description cannot exceed 4,000 characters.
User name Optional. User name with the appropriate permissions to read data from the OData resource.
URL OData service root URL that exposes the data that you want to read.
Security Type Optional. Security protocol that the Developer tool must use to establish a secure connection with
the OData server.
Select one of the following values:
- None
- SSL
- TLS
Default is None.
Note: The order of the connection properties might vary depending on the tool where you view them.
Property Description
Name Name of the connection. The name is not case sensitive and must be unique within the
domain. The name cannot exceed 128 characters, contain spaces, or contain the
following special characters:
~ ` ! $ % ^ & * ( ) - + = { [ } ] | \ : ; " ' < , > . ? /
ID String that the Data Integration Service uses to identify the connection. The ID is not
case sensitive. It must be 255 characters or less and must be unique in the domain.
You cannot change this property after you create the connection. Default value is the
connection name.
Description The description of the connection. The description cannot exceed 765 characters.
Pass-through security Enables pass-through security for the connection. When you enable pass-through
enabled security for a connection, the domain uses the client user name and password to log
into the corresponding database, instead of the credentials defined in the connection
object.
Data Access Properties: The ODBC connection URL used to access metadata from the database.
Connection String <data source name>
Code Page The code page used to read from a source database or to write to a target database or
file.
Environment SQL SQL commands to set the database environment when you connect to the database.
The Data Integration Service runs the connection environment SQL each time it
connects to the database.
Transaction SQL SQL commands to set the database environment when you connect to the database.
The Data Integration Service runs the transaction environment SQL at the beginning of
each transaction.
SQL Identifier Character Type of character that the database uses to enclose delimited identifiers in SQL
queries. The available characters depend on the database type.
Select (None) if the database uses regular identifiers. When the Data Integration
Service generates SQL queries, the service does not place delimited characters around
any identifiers.
Select a character if the database uses delimited identifiers. When the Data Integration
Service generates SQL queries, the service encloses delimited identifiers within this
character.
Support Mixed-case Enable if the database uses case-sensitive identifiers. When enabled, the Data
Identifiers Integration Service encloses all identifiers within the character selected for the SQL
Identifier Character property.
When the SQL Identifier Character property is set to none, the Support Mixed-case
Identifiers property is disabled.
ODBC Provider The type of database to which ODBC connects. For pushdown optimization, specify the
database type to enable the Data Integration Service to generate native database SQL.
The options are:
- Other
- Sybase
- Microsoft_SQL_Server
- Snowflake
Default is Other.
Note: Use an ODBC connection to connect to Microsoft SQL Server when the Data Integration Service runs on
UNIX or Linux. Use a native connection to Microsoft SQL Server when the Data Integration Service runs on
Windows.
Note: The order of the connection properties might vary depending on the tool where you view them.
Property Description
Name Name of the connection. The name is not case sensitive and must be unique within the
domain. The name cannot exceed 128 characters, contain spaces, or contain the
following special characters:
~ ` ! $ % ^ & * ( ) - + = { [ } ] | \ : ; " ' < , > . ? /
ID String that the Data Integration Service uses to identify the connection. The ID is not
case sensitive. It must be 255 characters or less and must be unique in the domain.
You cannot change this property after you create the connection. Default value is the
connection name.
Description The description of the connection. The description cannot exceed 765 characters.
Pass-through security Enables pass-through security for the connection. When you enable pass-through
enabled security for a connection, the domain uses the client user name and password to log
into the corresponding database, instead of the credentials defined in the connection
object.
Metadata Access Connection string used to access metadata from the database.
Properties: Connection Use the following connection string:
String
jdbc:informatica:oracle://<host_name>:<port>;SID=<database name>
Use the following connection string to connect to Oracle database through Oracle
Connection Manager:
jdbc:informatica:oracle:TNSNamesFile=<fully qualified path to the
tnsnames.ora file>;TNSServerName=<TNS server name>;
AdvancedJDBCSecurityOpti Database parameters for metadata access to a secure database. Informatica treats the
ons value of the AdvancedJDBCSecurityOptions field as sensitive data and stores the
parameter string encrypted.
To connect to a secure database, include the following parameters:
- EncryptionMethod. Required. Indicates whether data is encrypted when transmitted
over the network. This parameter must be set to SSL.
- ValidateServerCertificate. Optional. Indicates whether Informatica validates the
certificate that is sent by the database server.
If this parameter is set to True, Informatica validates the certificate that is sent by
the database server. If you specify the HostNameInCertificate parameter,
Informatica also validates the host name in the certificate.
If this parameter is set to false, Informatica does not validate the certificate that is
sent by the database server. Informatica ignores any truststore information that you
specify.
- HostNameInCertificate. Optional. Host name of the machine that hosts the secure
database. If you specify a host name, Informatica validates the host name included
in the connection string against the host name in the SSL certificate.
- cryptoProtocolVersion. Optional. If you enable TLS for the Oracle instance, set the
cryptoProtocolVersion parameter as follows:
cryptoProtocolVersion=TLSv<version number>.
For example, cryptoProtocolVersion=TLSv1.2
Note: The version number must be the same as the TLS version you configured for
the server.
- TrustStore. Required. Path and file name of the truststore file.
Note: If you configure SSL or TLS and specify only the file name, you must copy the
truststore file to the following directory to test the connection: <Informatica
server installation directory>/tomcat/bin
- TrustStorePassword. Required. Password for the truststore file for the secure
database.
- KeyStore. Required. Path and file name of the keystore file.
- KeyStorePassword. Required. Password for the keystore file for the secure
database.
Note: Informatica appends the secure JDBC parameters to the connection string. If you
include the secure JDBC parameters directly to the connection string, do not enter any
parameters in the AdvancedJDBCSecurityOptions field.
Code Page The code page used to read from a source database or to write to a target database or
file.
Environment SQL SQL commands to set the database environment when you connect to the database.
The Data Integration Service runs the connection environment SQL each time it
connects to the database.
Transaction SQL SQL commands to set the database environment when you connect to the database.
The Data Integration Service runs the transaction environment SQL at the beginning of
each transaction.
Enable Parallel Mode Enables parallel processing when loading data into a table in bulk mode. By default,
this option is cleared.
SQL Identifier Character Type of character that the database uses to enclose delimited identifiers in SQL
queries. The available characters depend on the database type.
Select (None) if the database uses regular identifiers. When the Data Integration
Service generates SQL queries, the service does not place delimited characters around
any identifiers.
Select a character if the database uses delimited identifiers. When the Data Integration
Service generates SQL queries, the service encloses delimited identifiers within this
character.
Support Mixed-case Enable if the database uses case-sensitive identifiers. When enabled, the Data
Identifiers Integration Service encloses all identifiers within the character selected for the SQL
Identifier Character property.
When the SQL Identifier Character property is set to none, the Support Mixed-case
Identifiers property is disabled.
Note: The order of the connection properties might vary depending on the tool where you view them.
Property Description
Name Name of the connection. The name is not case sensitive and must be unique
within the domain. It cannot exceed 128 characters, contain spaces, or contain
the following special characters:
~ ` ! $ % ^ & * ( ) - + = { [ } ] | \ : ; " ' < , > . ? /
ID String that the Data Integration Service uses to identify the connection. The ID is
not case sensitive. It must be 255 characters or less and must be unique in the
domain. You cannot change this property after you create the connection.
Default value is the connection name.
Description The description of the connection. The description cannot exceed 765
characters.
Location The Informatica domain where you want to create the connection.
Type The connection type. You can select the Standard connection type or OAuth
connection type.
User Name Applicable for Standard connection type. Salesforce user name.
User Password Applicable for Standard connection type. Password for the Salesforce user
name.
To access Salesforce outside the trusted network of your organization, you must
append a security token to your password to log in to the API or a desktop
client.
To receive or reset your security token, log in to Salesforce and click Setup > My
Personal Information > Reset My Security Token.
Password is case sensitive.
Refresh Token Applicable for OAuth connection type. Refresh Token of Salesforce.
Consumer Key Applicable for OAuth connection type. The Consumer Key obtained from
Salesforce, required to generate the Refresh Token. For more information about
how to generate the Consumer Key, see the Salesforce documentation.
Consumer Secret Applicable for OAuth connection type. The Consumer Secret obtained from
Salesforce, required to generate the Refresh Token. For more information about
how to generate the Consumer Secret, see the Salesforce documentation.
Note: The order of the connection properties might vary depending on the tool where you view them.
The following table describes the Salesforce Marketing Cloud connection properties:
Salesforce Marketing Cloud Url The URL that the Data Integration Service uses to connect to the
Salesforce Marketing Cloud WSDL.
The following URL is an example for OAuth 1.0 URL:
https://webservice.s7.exacttarget.com/etframework.wsdl
The following URL is an example for OAuth 2.0 URL:
https://<SUBDOMAIN>.soap.marketingcloudapis.com/
etframework.wsdl
Important: Salesforce is going to deprecate the OAuth 1.0 APIs by
September 30th , 2022. Informatica recommends that you upgrade to
OAuth 2.0 for new and existing packages.
Enable Logging When you enable logging you can see the session log for the tasks.
UTC Offset The Secure Agent uses the UTC offset connection property to read data
from and write data to Salesforce Marketing Cloud in UTC offset time
zone.
Batch Size Number of rows that the Secure Agent writes in a batch to the target.
When you insert or update data and specify the contact key, the data
associated with the specified contact ID is inserted or updated in a batch
to Salesforce Marketing Cloud. When you upsert data to Salesforce
Marketing Cloud, do not specify the contact key.
Enable Multiple BU Select this option if there are multiple business units in your Salesforce
Marketing Cloud account. You can use the Salesforce Marketing Cloud
connection to access data across all business units.
Note: The order of the connection properties might vary depending on the tool where you view them.
Property Description
Username Required. User name for the SAP source system that you want to access.
Message host Required when you create an SAP load balancing connection.
name Host name of the SAP message server.
Language Optional. Language that you want to use for mappings and workflows.
Must be compatible with the Developer tool code page.
If you leave this option blank, the Developer tool uses the default language of the SAP system.
Trace Optional. Use this option to track the JCo calls that the SAP system makes. SAP stores the
information about the JCo calls in a trace file.
Specify one of the following values:
- 0. Off
- 1. Full
Default is 0.
You can access the trace files from the following directories:
- <Informatica installation directory>/tomcat/bin directory on the machine where
you installed the Informatica services
- <Informatica installation directory>/clients/DeveloperClient directory on the
machine where you installed the Developer tool
Additional Optional. Enter any other connection parameter that you want to use.
parameters Use the following format:
<parameter name>=<value>
Staging directory Path in the SAP system where the stage file will be created.
Source directory Path that contains the source file. The path must be accessible to the Data Integration Service.
Retry period Number of seconds that the Data Integration Service attempts to reconnect to the FTP host if the
connection fails.
If the Data Integration Service cannot reconnect to the FTP host in the retry period, the mapping or
workflow fails.
Default is 0. A value of 0 indicates an infinite retry period.
Public key file Required when you enable SFTP and the SFTP server uses public key authentication.
name Public key file path and file name.
Private key file Required when you enable SFTP and the SFTP server uses public key authentication.
name Private key file path and file name.
Private key file Required when you enable SFTP, and the SFTP server uses public key authentication and the
name password private key is encrypted.
Password to decrypt the private key file.
Port Range HTTP port range that the Data Integration Service must use to read data from the SAP server in
streaming mode.
Enter the minimum and maximum port numbers with a hyphen as the separator. The minimum and
maximum port number can range between 10000 and 65535. You can also specify the port range
according to your organization.
Default is 10000-65535.
Use HTTPS Select this option to enable HTTPS streaming when you read data from SAP tables.
By default, the Use HTTPS check box is not selected.
For more information about configuring HTTPS for table reader mappings in streaming mode, see
the article
"HTTPS Configuration for Table Reader Mappings in Streaming Mode for PowerExchange for SAP
NetWeaver" on the Informatica Documentation Portal.
A sequential data source is a data source that PowerExchange can access by using a data map defined with
an access method of SEQ. The Data Integration Service connects to the data source through PowerExchange.
Note: The order of the connection properties might vary depending on the tool where you view them.
Option Description
Location Node name for the location of the PowerExchange Listener that connects to the sequential
data set. The node name is defined in the first parameter of the NODE statement in the
PowerExchange dbmover.cfg configuration file.
User name A user name that has the authority to access the sequential data set.
Password Password for the specified user name or a valid PowerExchange passphrase.
A PowerExchange passphrase can be from 9 to 128 characters in length and can contain
the following characters:
- Uppercase and lowercase letters
- The numbers 0 to 9
- Spaces
- The following special characters:
’ - ; # \ , . / ! % & * ( ) _ + { } : @ | < > ?
Note: The first character is an apostrophe.
Passphrases cannot include single quotation marks (‘), double quotation marks (“), or
currency symbols.
To use passphrases, ensure that the PowerExchange Listener runs with a security setting
of SECURITY=(1,N) or higher in the DBMOVER member. For more information, see
"SECURITY Statement" in the PowerExchange Reference Manual.
The allowable characters in the IBM IRRPHREX exit do not affect the allowable characters
in PowerExchange passphrases.
Note: A valid RACF passphrase can be up to 100 characters in length. PowerExchange
truncates passphrases longer than 100 characters when passing them to RACF for
validation.
Code page Required. Name of the code page to use for reading from or writing to the sequential data
set. Usually, this value is an ISO code page name, such as ISO-8859-6.
Encryption type Optional. The type of encryption that the Data Integration Service uses. Select one of the
following options:
- None
- AES
Default is None.
Note: Informatica recommends that you use Secure Sockets Layer (SSL) authentication
instead of configuring the Encryption Type and Level connection properties. SSL
authentication provides stricter security and is used by several Informatica products.
For more information about implementing SSL authentication in a PowerExchange network,
see the PowerExchange Reference Manual.
[Encryption] Level If you select AES for Encryption Type, select one of the following options to indicate the
encryption level that the Data Integration Service uses:
- 1. Use a128-bit encryption key.
- 2. Use a 192-bit encryption key.
- 3. Use a 256-bit encryption key.
If you do not select AES for Encryption Type, this option is ignored.
Default is 1.
Pacing size Optional. Amount of data that the source system can pass to the PowerExchange Listener.
Set the pacing size if an external application, database, or the Data Integration Service
node is a bottleneck. User lower values for faster performance.
Minimum value and default value is 0. A value of 0 provides the best performance.
Interpret as rows Optional Select this option to express the pacing size as a number of rows. Clear this
option to express the pacing size in kilobytes. By default, this option is not selected and
the pacing size is in kilobytes.
Compression Optional. Select this option to enable source data compression. By compressing data, you
can decrease the amount of data that Informatica applications send over the network. By
default, this option is not selected and compression is disabled.
Offload processing Optional. Controls whether to offload some bulk data processing from the source machine
to the Data Integration Service machine. Select one of the following options:
- AUTO. The Data Integration Service determines whether to use offload processing.
- Yes. Use offload processing.
- No. Do not use offload processing.
Default is AUTO.
Worker threads Optional. Number of threads that the Data Integration Service uses to process bulk data
when offload processing is enabled. For optimal performance, this value should not
exceed the number of available processors on the Data Integration Service machine. Valid
values are 1 through 64. Default is 0, which disables multithreading.
Array size Optional The number of records in the storage array for the worker threads. This option is
applicable when you set the Worker Threads option to a value greater than 0. Valid values
are 25 to 5000. Default is 25.
Write mode Optional. Mode in which the Data Integration Service sends data to the PowerExchange
Listener. Select one of the following write modes:
- CONFIRMWRITEON. Sends data to the PowerExchange Listener and waits for a response
before sending more data. Select this option when error recovery is a priority. However,
this option might degrade performance.
- CONFIRMWRITEOFF. Sends data to the PowerExchange Listener without waiting for a
response. Use this option if you can reload the target table when an error occurs.
- ASYNCHRONOUSWITHFAULTTOLERANCE. Sends data to the PowerExchange Listener
without waiting for a response. This option also enables error detection. This option
combines the speed of CONFIRMWRITEOFF and the data integrity of CONFIRMWRITEON.
Default is CONFIRMWRITEON.
Note: The order of the connection properties might vary depending on the tool where you view them.
Property Description
Name The name of the connection. The name is not case sensitive and must be unique within the domain.
You can change this property after you create the connection. The name cannot exceed 128
characters, contain spaces, or contain the following special characters:~ ` ! $ % ^ & * ( ) - + = { [ } ] |
\:;"'<,>.?/
ID String that the Data Integration Service uses to identify the connection.
The ID is not case sensitive. The ID must be 255 characters or less and must be unique in the domain.
You cannot change this property after you create the connection.
Default value is the connection name.
Description Optional. The description of the connection. The description cannot exceed 4,000 characters.
Additional Enter one or more JDBC connection parameters in the following format:
JDBC URL
<param1>=<value>&<param2>=<value>&<param3>=<value>....
Parameters
For example:
user=jon&warehouse=mywh&db=mydb&schema=public
To access Snowflake through Okta SSO authentication, enter the web-based IdP implementing SAML
2.0 protocol in the following format:
authenticator=https://<Your_Okta_Account_Name>.okta.com
Note: Microsoft ADFS is not supported.
For more information about configuring Okta authentication, see the following website:
https://docs.snowflake.net/manuals/user-guide/admin-security-fed-auth-configure-
snowflake.html#configuring-snowflake-to-use-federated-authentication
Note: The order of the connection properties might vary depending on the tool where you view them.
Property Description
Name Name of the connection. The name is not case sensitive and must be unique within the domain. The
name cannot exceed 128 characters, contain spaces, or contain the following special characters:
~ ` ! $ % ^ & * ( ) - + = { [ } ] | \ : ; " ' < , > . ? /
ID String that the Data Integration Service uses to identify the connection. The ID is not case sensitive.
It must be 255 characters or less and must be unique in the domain. You cannot change this
property after you create the connection. Default value is the connection name.
Description Description of the connection. The description cannot exceed 765 characters.
User Name Teradata database user name with the appropriate read and write permissions to access the
database.
Property Description
Tenacity Number of hours that Teradata PT API continues trying to log on when the maximum number of
operations run on the Teradata database.
Must be a positive, non-zero integer. Default is 4.
Max Sessions Maximum number of sessions that Teradata PT API establishes with the Teradata database.
Must be a positive, non-zero integer. Default is 4.
Min Sessions Minimum number of Teradata PT API sessions required for the Teradata PT API job to continue.
Must be a positive integer between 1 and the Max Sessions value. Default is 1.
Sleep Number of minutes that Teradata PT API pauses before it retries to log on when the maximum
number of operations run on the Teradata database.
Must be a positive, non-zero integer. Default is 6.
Use Metadata Indicates that the Teradata Connector for Hadoop (TDCH) must use the JDBC URL that you
JDBC URL for specified in the connection string under the metadata access properties.
TDCH Default is selected.
Clear this option to enter a different JDBC URL that TDCH must use when it runs the mapping.
TDCH JDBC Url Enter the JDBC URL that TDCH must use when it runs a Teradata mapping. Use the following
format:
jdbc:teradata://<hostname>/database=<database
name>,tmode=ANSI,charset=UTF8
This field is available only when you clear the Use Metadata JDBC URL for TDCH option.
Data Encryption Enables full security encryption of SQL requests, responses, and data on Windows.
Default is disabled.
Additional Sqoop This property is applicable if you use a Hortonworks or Cloudera cluster, and run a Teradata
Arguments mapping on the Blaze or Spark engine through Sqoop.
Enter the arguments that Sqoop must use to process the data. For example, enter --method
split.by.amp. Separate multiple arguments with a space.
See the Hortonworks for Teradata Connector and Cloudera Connector Powered by Teradata
documentation for a list of arguments that you can specify.
Note: If you use Hortonworks for Teradata Connector, the --split-by argument is required if you
add two or more source tables in the read operation. If you use Cloudera Connector Powered by
Teradata, the --split-by argument is required in the source connection if the source table does
not have a primary key defined.
Property Description
ID String that the Data Integration Service uses to identify the connection. The ID is not case sensitive. It
must be 255 characters or less and must be unique in the domain. You cannot change this property
after you create the connection. Default value is the connection name.
Description Description of the connection. The description cannot exceed 765 characters.
Location The Informatica domain where you want to create the connection.
Connection Description
Property
Tableau Product The name of the Tableau product to which you want to connect.
You can choose one of the following Tableau products to publish the TDE or TWBX file:
- Tableau Desktop. Creates a TDE file in the Data Integration Service machine. You can then
manually import the TDE file to Tableau Desktop.
- Tableau Server. Publishes the generated TDE or TWBX file to Tableau Server.
- Tableau Online. Publishes the generated TDE or TWBX file to Tableau Online.
Connection URL URL of Tableau Server or Tableau Online to which you want to publish the TDE or TWBX file. The
URL has the following format: http://<Host name of Tableau Server or Tableau
Online>:<port>
User Name User name of the Tableau Server or Tableau Online account.
Content URL The name of the site on Tableau Server or Tableau Online where you want to publish the TDE or
TWBX file.
Contact the Tableau administrator to provide the site name.
Property Description
ID String that the Data Integration Service uses to identify the connection. The ID is not case sensitive. It
must be 255 characters or less and must be unique in the domain. You cannot change this property
after you create the connection. Default value is the connection name.
Description Description of the connection. The description cannot exceed 765 characters.
Location The Informatica domain where you want to create the connection.
Connection Description
Property
Tableau Product The name of the Tableau product to which you want to connect.
You can choose one of the following Tableau products to publish the .hyper or TWBX file:
Tableau Desktop
Creates a .hyper file in the Data Integration Service machine. You can then manually import
the .hyper file to Tableau Desktop.
Tableau Server
Tableau Online
Publishes the generated .hyper or TWBX file to Tableau Online.
Connection URL The URL of Tableau Server or Tableau Online to which you want to publish the .hyper or TWBX
file.
Enter the URL in the following format: http://<Host name of Tableau Server or Tableau
Online>:<port>
User Name The user name of the Tableau Server or Tableau Online account.
Password The password for the Tableau Server or Tableau Online account.
Site ID The ID of the site on Tableau Server or Tableau Online where you want to publish the or TWBX file.
Note: Contact the Tableau administrator to provide the site ID.
Schema File The path to a sample .hyper file from where the Data Integration Service imports the Tableau
Path metadata.
Enter one of the following options for the schema file path:
- Absolute path to the .hyper file.
- Directory path for the .hyper files.
- Empty directory path.
The path you specify for the schema file becomes the default path for the target .hyper file. If
you do not specify a file path, the Data Integration Service uses the following default file path for
the target .hyper file:
<Data Integration Service installation directory>/apps/
Data_Integration_Server/<latest version>/bin/rtdm
Note: The order of the connection properties might vary depending on the tool where you view them.
Property Description
Name Name of the connection. The name is not case sensitive and must be unique within the domain. The
name cannot exceed 128 characters, contain spaces, or contain the following special characters:
~ ` ! $ % ^ & * ( ) - + = { [ } ] | \ : ; " ' < , > . ? /
ID String that the Data Integration Service uses to identify the connection. The ID is not case sensitive. It
must be 255 characters or less and must be unique in the domain. You cannot change this property
after you create the connection. Default value is the connection name.
Description The description of the connection. The description cannot exceed 765 characters.
The following table describes the properties for hose type and OAuth authentication:
Property Description
Hose Type Streaming API methods. You can specify one of the following methods:
- Filter. The Twitter statuses/filter method returns public statuses that match the search
criteria.
- Sample. The Twitter statuses/sample method returns a random sample of all public
statuses.
Consumer Key The consumer key that you get when you create the application in Twitter. Twitter uses the key
to identify the application.
Consumer Secret The consumer secret that you get when you create the Twitter application. Twitter uses the
secret to establish ownership of the consumer key.
Do you have OAuth Indicates whether you want to configure OAuth. Select one of the following values:
details? - Yes. Indicates that you have the access token and secret.
- No. Launches the OAuth Utility.
Access Token Access token that the OAuth Utility returns. Twitter uses the token instead of the user
credentials to access the protected resources.
Access Secret Access secret that the OAuth Utility returns. The secret establishes ownership of a token.
Note: The order of the connection properties might vary depending on the tool where you view them.
Option Description
Location Node name for the location of the PowerExchange Listener that connects to the VSAM
data set. The node name is defined in the first parameter of the NODE statement in the
PowerExchange dbmover.cfg configuration file.
User name A user name that has the authority to connect to the VSAM data set.
Code page Required. Name of the code page to use for reading from or writing to the VSAM data set.
Usually, this value is an ISO code page name, such as ISO-8859-6.
Encryption type Optional. The type of encryption that the Data Integration Service uses. Select one of the
following options:
- None
- AES
Default is None.
Note: Informatica recommends that you use Secure Sockets Layer (SSL) authentication
instead of configuring the Encryption Type and Level connection properties. SSL
authentication provides stricter security and is used by several Informatica products.
For more information about implementing SSL authentication in a PowerExchange network,
see the PowerExchange Reference Manual.
[Encryption] Level If you select AES for Encryption Type, select one of the following options to indicate the
encryption level that the Data Integration Service uses:
- 1. Use a128-bit encryption key.
- 2. Use a 192-bit encryption key.
- 3. Use a 256-bit encryption key.
If you do not select AES for Encryption Type, this option is ignored.
Default is 1.
Pacing size Optional. Amount of data that the source system can pass to the PowerExchange Listener.
Set the pacing size if an external application, database, or the Data Integration Service
node is a bottleneck. User lower values for faster performance.
Minimum value and default value is 0. A value of 0 provides the best performance.
Interpret as rows Optional. Select this option to express the pacing size as a number of rows. Clear this
option to express the pacing size in kilobytes. By default, this option is not selected and
the pacing size is in kilobytes.
Compression Optional. Select this option to enable source data compression. By compressing data, you
can decrease the amount of data that Informatica applications send over the network. By
default, this option is not selected and compression is disabled.
Offload processing Optional. Controls whether to offload some bulk data processing from the source machine
to the Data Integration Service machine. Select one of the following options:
- AUTO. The Data Integration Service determines whether to use offload processing.
- Yes. Use offload processing.
- No. Do not use offload processing.
Default is AUTO.
Worker threads Optional. Number of threads that the Data Integration Service uses to process bulk data
when offload processing is enabled. For optimal performance, this value should not
exceed the number of available processors on the Data Integration Service machine. Valid
values are 1 through 64. Default is 0, which disables multithreading.
Array size Optional. The number of records in the storage array for the worker threads. This option is
applicable when you set the Worker Threads option to a value greater than 0. Valid values
are 25 to 5000. Default is 25.
Write mode Optional. Mode in which Data Integration Service sends data to the PowerExchange
Listener. Select one of the following write modes:
- CONFIRMWRITEON. Sends data to the PowerExchange Listener and waits for a response
before sending more data. Select this option when error recovery is a priority. However,
this option might degrade performance.
- CONFIRMWRITEOFF. Sends data to the PowerExchange Listener without waiting for a
response. Use this option if you can reload the target table when an error occurs.
- ASYNCHRONOUSWITHFAULTTOLERANCE. Sends data to the PowerExchange Listener
without waiting for a response. This option also enables error detection. This option
combines the speed of CONFIRMWRITEOFF and the data integrity of CONFIRMWRITEON.
Default is CONFIRMWRITEON.
Property Description
Username User name to connect to the web service. Enter a user name if you enable HTTP authentication
or WS-Security.
If the Web Service Consumer transformation includes WS-Security ports, the transformation
receives a dynamic user name through an input port. The Data Integration Service overrides the
user name defined in the connection.
Password Password for the user name. Enter a password if you enable HTTP authentication or WS-
Security.
If the Web Service Consumer transformation includes WS-Security ports, the transformation
receives a dynamic password through an input port. The Data Integration Service overrides the
password defined in the connection.
End Point URL URL for the web service that you want to access. The Data Integration Service overrides the URL
defined in the WSDL file.
If the Web Service Consumer transformation includes an endpoint URL port, the transformation
dynamically receives the URL through an input port. The Data Integration Service overrides the
URL defined in the connection.
Timeout Number of seconds that the Data Integration Service waits for a response from the web service
provider before it closes the connection. Specify a timeout value between 1 and 10,000 seconds.
HTTP Type of user authentication over HTTP. Select one of the following values:
Authentication - None. No authentication.
Type - Automatic. The Data Integration Service chooses the authentication type of the web service
provider.
- Basic. Requires you to provide a user name and password for the domain of the web service
provider. The Data Integration Service sends the user name and the password to the web
service provider for authentication.
- Digest. Requires you to provide a user name and password for the domain of the web service
provider. The Data Integration Service generates an encrypted message digest from the user
name and password and sends it to the web service provider. The provider generates a
temporary value for the user name and password and stores it in the Active Directory on the
Domain Controller. It compares the value with the message digest. If they match, the web
service provider authenticates you.
- NTLM. Requires you to provide a domain name, server name, or default user name and
password. The web service provider authenticates you based on the domain you are
connected to. It gets the user name and password from the Windows Domain Controller and
compares it with the user name and password that you provide. If they match, the web service
provider authenticates you. NTLM authentication does not store encrypted passwords in the
Active Directory on the Domain Controller.
WS Security Type Type of WS-Security that you want to use. Select one of the following values:
- None. The Data Integration Service does not add a web service security header to the
generated SOAP request.
- PasswordText. The Data Integration Service adds a web service security header to the
generated SOAP request. The password is stored in the clear text format.
- PasswordDigest. The Data Integration Service adds a web service security header to the
generated SOAP request. The password is stored in a digest form which provides effective
protection against replay attacks over the network. The Data Integration Service combines the
password with a nonce and a time stamp. The Data Integration Service applies a SHA hash on
the password, encodes it in base64 encoding, and uses the encoded password in the SOAP
header.
Trust Certificates File containing the bundle of trusted certificates that the Data Integration Service uses when
File authenticating the SSL certificate of the web service. Enter the file name and full directory path.
Default is <Informatica installation directory>/services/shared/bin/ca-
bundle.crt.
Client Certificate Client certificate that a web service uses when authenticating a client. Specify the client
File Name certificate file if the web service needs to authenticate the Data Integration Service.
Client Certificate Password for the client certificate. Specify the client certificate password if the web service
Password needs to authenticate the Data Integration Service.
Client Certificate Format of the client certificate file. Select one of the following values:
Type - PEM. Files with the .pem extension.
- DER. Files with the .cer or .der extension.
Specify the client certificate type if the web service needs to authenticate the Data Integration
Service.
Private Key File Private key file for the client certificate. Specify the private key file if the web service needs to
Name authenticate the Data Integration Service.
Private Key Password for the private key of the client certificate. Specify the private key password if the web
Password service needs to authenticate the Data Integration Service.
Private Key Type Type of the private key. PEM is the supported type.
A database identifier is a database object name. Tables, views, columns, indexes, triggers, procedures,
constraints, and rules can have identifiers. You use the identifier to reference the object in SQL queries. A
database can have regular identifiers or delimited identifiers that must be enclosed within delimited
characters.
For example, the following SQL statement uses the regular identifiers MYTABLE and MYCOLUMN:
SELECT * FROM MYTABLE
WHERE MYCOLUMN = 10
Delimited Identifiers
Delimited identifiers must be enclosed within delimited characters because they do not comply with the
format rules for identifiers.
If an identifier uses a reserved keyword, you must enclose the identifier within delimited characters in an
SQL query. For example, the following SQL statement accesses a table named ORDER:
SELECT * FROM “ORDER”
WHERE MYCOLUMN = 10
Identifiers that use special characters
If an identifier uses special characters, you must enclose the identifier within delimited characters in an
SQL query. For example, the following SQL statement accesses a table named MYTABLE$@:
SELECT * FROM “MYTABLE$@”
WHERE MYCOLUMN = 10
Case-sensitive identifiers
By default, identifiers in IBM DB2, Microsoft SQL Server, and Oracle databases are not case sensitive.
Database object names are stored in uppercase, but SQL queries can use any case to refer to them. For
example, the following SQL statements access the table named MYTABLE:
SELECT * FROM mytable
SELECT * FROM MyTable
SELECT * FROM MYTABLE
To use case-sensitive identifiers, you must enclose the identifier within delimited characters in an SQL
query. For example, the following SQL statement accesses a table named MyTable:
SELECT * FROM “MyTable”
WHERE MYCOLUMN = 10
Identifier Properties
When you create most database connections, you must configure database identifier properties. The
identifier properties that you configure depend on whether the database uses regular identifiers, uses
keywords or special characters in identifiers, or uses case-sensitive identifiers.
Type of character that the database uses to enclose delimited identifiers in SQL queries. The available
characters depend on the database type.
Select (None) if the database uses regular identifiers. When the Data Integration Service generates SQL
queries, the service does not place delimited characters around any identifiers.
Enable if the database uses case-sensitive identifiers. When enabled, the Data Integration Service
encloses all identifiers within the character selected for the SQL Identifier Character property.
In the Informatica client tools, you must refer to the identifiers with the correct case. For example, when
you create the database connection, you must enter the database user name with the correct case.
When the SQL Identifier Character property is set to none, the Support Mixed-case Identifiers property
is disabled.
In the database connection, set the SQL Identifier Character property to (None). When SQL Identifier
Character is set to none, the Support Mixed-case Identifiers property is disabled.
When the Data Integration Service generates SQL queries, the service does not place delimited characters
around any identifiers.
1. Set the SQL Identifier Character property to the character that the database uses for delimited
identifiers.
This example sets the property to "" (quotes).
2. Clear the Support Mixed-case Identifiers property.
When the Data Integration Service generates SQL queries, the service places the selected character around
identifiers that use a reserved keyword or that use a special character. For example, the Data Integration
Service generates the following query:
SELECT * FROM "MYTABLE$@" /* identifier with special characters enclosed within
delimited
character */
WHERE MYCOLUMN = 10 /* regular identifier not enclosed within delimited character */
1. Set the SQL Identifier Character property to the character that the database uses for delimited
identifiers.
This example sets the property to "" (quotes).
2. Select the Support Mixed-case Identifiers property.
Schedules
This chapter includes the following topics:
Schedules Overview
You can create a schedule to indicate when to run deployed mappings and workflows. You can also schedule
the date and time to run profiles and scorecards.
When you create a schedule, you configure when the schedule runs. You can add jobs to the schedule when
you create it or you can save the schedule and add jobs to it later.
You might define a schedule to automate routine tasks or manage resource usage in the domain. For
example, you can schedule large mappings to run at different times to prevent the mappings from
overloading the nodes. You can also pause the schedule to prevent the jobs from running during holidays or
during domain maintenance.
You can select a time zone when you configure the date and time to run the schedule. By default, when you
enter a start time for a schedule, the time zone is the time zone of your client machine. If the Data Integration
Service runs on a machine in a different time zone from the client, you need to ensure that the job runs at a
specific time for that time zone. You can choose the time zone when you define the schedule.
You can schedule deployed mappings and workflows including mappings and workflows for profiles and
scorecards. Optionally, you can configure a parameter file or parameter set for the job. Configure a
parameter file or set when you want to run the deployed mapping or workflow with different settings. If the
Scheduler Service runs on multiple nodes, you must store the parameter file in a directory that all of the
nodes can access. Configure the directory in the Storage Properties for the Scheduler Service.
216
Optionally, you can configure the Run As User property for the job. You might configure the property if you
need to schedule an object that you do not have permission on. When you configure the Run As User
property, you run the job as a user in the domain who has permission on the object.
After you create a schedule, you can edit the schedule. When you edit a schedule, you can change the
recurrence options, schedule jobs, or change the job parameters.
Creating a Schedule
Create a schedule to run deployed mappings, deployed workflows, profiles, or scorecards at a specified time.
6. If you are creating a recurring schedule, specify whether to run the job on a Daily, Weekly, or Monthly
interval.
The following table describes the recurrence options that you can configure:
Option Description
Daily - Run after every Run the job every n minutes or hours
Monthly - Run every n day Run the job on a day of the month
Monthly - Run every Run the job on a day even n week of the month
Monthly - Run every last day of the month Run the job on the last day of the month
7. In the Start section, configure a start date and time for the schedule.
8. Select the time zone for the schedule from the list.
The default time zone is the time zone of the client machine.
9. If you are creating a recurring schedule, you can optionally configure an end date and time.
You can end the recurrence on a specific date or after n runs.
Editing a Schedule
You can edit a schedule in the Administrator tool. You can edit a schedule to change recurrence options,
schedule jobs, or run jobs with different parameters.
When you pause a schedule, the jobs that run on the schedule stop running until you resume the schedule.
Deleting a Schedule
You can delete a schedule in the Administrator tool.
When you delete a schedule, any jobs that are running finish running, and all future runs are canceled. If you
do not want the jobs to stop running, add them to a different schedule before you delete the schedule.
You might migrate domain objects from a development environment to a test or production environment.
To export and import domain objects, use the following infacmd isp commands:
ExportDomainObjects
Exports native users, native groups, roles, and connections to an XML file.
ImportDomainObjects
Imports native users, native groups, roles, and connections into an Informatica domain.
You can use an infacmd control file to filter the objects during the export or import.
You can also use the infacmd xrf generateReadableViewXML command to generate a readable XML file from
an export file. You can review the readable XML file to determine if you need to filter the objects that you
import.
Export Process
You can use the command line to export domain objects from a domain.
221
The command exports the domain objects to an export file. You can use this file to import the objects into
another domain.
• When you export a user, by default, you do not export the user password. If you do not export the
password, the administrator must reset the password for the user after the user is imported into the
domain. However, when you run the infacmd isp exportDomainObjects command, you can choose to
export an encrypted version of the password.
• When you export a user, you do not export the associated groups of the user. If applicable, assign the user
to the group after you import the user and group.
• When you export a group, you export all sub-groups and users in the group.
• You cannot export the Administrator user, the Administrator role, the Everyone group, or LDAP users or
groups. To replicate LDAP users and groups in an Informatica domain, import the LDAP users and groups
directly from the LDAP directory service.
• To export native users and groups from domains of different versions, use the infacmd isp
exportUsersAndGroups command.
• When you export a connection, by default, you do not export the connection password. If you do not
export the password, the administrator must reset the password for the connection after the connection is
imported into the domain. However, when you run the infacmd isp exportDomainObjects command, you
can choose to export an encrypted version of the password.
Run infacmd xrf generateReadableViewXML command, to create a readable XML from the export file.
User
UserInfo
Role
ServicePrivilegeDef
Privilege
Group
GroupRef
UserRef
ConnectInfo
ConnectionPoolAttributes
DB2iNativeConnection
DB2NativeConnection
DB2zNativeConnection
JDBCConnection
ODBCNativeConnection
OracleNativeConnection
PWXMetaConnection
SAPConnection
Import Process
You can use the command line to import domain objects from an export file into a domain.
1. Run the infacmd xrf generateReadableViewXML command to generate a readable XML file from an
export file. Review the domain objects in the readable XML file and determine the objects that you want
to import.
2. If you do not want to import all domain objects in the export file, create an import control file to filter the
objects that are imported.
3. Run the infacmd isp importDomainObjects command to import the domain objects into the specified
domain.
• When you import a group, you import all sub-groups and users in the group.
• To import native users and groups from domains of different versions, use the infacmd isp
importUsersAndGroups command.
• After you import a user or group, you cannot rename the user or group.
• You import roles independently of users and groups. Assign roles to users and groups after you import
the roles, users, and groups.
• You cannot import the Administrator group, the Administrator user, the Administrator role, the Everyone
group, or LDAP users or groups.
Conflict Resolution
A conflict occurs when you try to import an object with a name that exists for an object in the target domain.
Configure the conflict resolution to determine how to handle conflicts during the import.
You can define a conflict resolution strategy through the command line or control file when you import the
objects. The control file takes precedence if you define conflict resolution in the command line and control
file. The import fails if there is a conflict and you did not define a conflict resolution strategy.
Rename
Renames the source object. You can provide a name in the control file, or else the name is generated. A
generated name has a number appended to the end of the name.
Replace
Merge
Merges the source and target objects into one group. For example, if you merge groups with the same
name, users and sub-groups from both groups are merged into the group in the target domain.
You cannot define the merge conflict resolution strategy through the command line. Use a control file to
define the merge conflict resolution strategy. You must include the group object type section with merge
as the conflict resolution policy with reuse, replace, or rename for all conflicting users in the control file.
For example, specify the merge conflict resolution strategy for the following groups:
You get the following results in the group after merge in the target domain:
License Management
This chapter includes the following topics:
• Run application services, such as the Analyst Service, Data Integration Service, and PowerCenter
Repository Service.
• Use add-on options, such as partitioning for PowerCenter, grid, and high availability.
• Access particular types of connections, such as Oracle, Teradata, Microsoft SQL Server, and IBM MQ
Series.
• Use Metadata Exchange options, such as Metadata Exchange for Cognos and Metadata Exchange for
Rational Rose.
When you install Informatica, the installation program creates a license object in the domain based on the
license key that you used during installation.
You assign a license object to each application service to enable the service. For example, you must assign a
license to the PowerCenter Integration Service before you can use the PowerCenter Integration Service to run
a workflow.
You can create additional license objects in the domain. Based on your project requirements, you may need
multiple license objects. For example, you may have two license objects, where each license object allows
you to run services on a different operating system. You might also use multiple license objects to manage
multiple projects in the same domain. One project may require access to particular database types, while the
other project does not.
225
License Validation
The Service Manager validates application service processes when they start. The Service Manager validates
the following information for each service process:
• Product version. Verifies that you are running the appropriate version of the Informatica services.
• Platform. Verifies that the Informatica services are running on a licensed operating system.
• Expiration date. Verifies that the license is not expired. If the license expires, no application service
assigned to the license can start. You must assign a valid license to the Informatica services to start
them.
• PowerCenter options. Determines the options that the Informatica services have permission to use. For
example, the Service Manager verifies if the PowerCenter Integration Service can use the Session on Grid
option.
• Connectivity. Verifies connections that the Informatica services have permission to use. For example, the
Service Manager verifies that PowerCenter can connect to a IBM DB2 database.
• Metadata Exchange options. Determines the Metadata Exchange options that are available for use. For
example, the Service Manager verifies that you have access to the Metadata Exchange for Business
Objects Designer.
You must have permission on the domain to view the logs for Licensing events.
• Create the license in the Administrator tool. You use a license key to create a license in the Administrator
tool.
• Assign a license to each application service. Assign a license to each application service to enable the
service.
• Unassign a license from an application service. Unassign a license from an application service if you want
to discontinue the service or migrate the service from a development environment to a production
environment. After you unassign a license from a service, you cannot enable the service until you assign
another valid license to it.
• Update the license. Update the license to add PowerCenter options to the existing license.
• Remove the license. Remove a license if it is obsolete.
You create a license from a license key file. You apply license keys to the license to enable additional
options. Informatica uses the following types of license keys:
• Original keys. Informatica generates an original key based on your contract. Informatica may provide
multiple original keys depending on your contract.
• Incremental keys. Informatica generates incremental keys based on updates to an existing license, such
as an extended license period or an additional option.
Note: Informatica licenses typically change with each version. Use a license key file valid for the current
version to ensure that your installation includes all functionality.
Original Keys
Original keys identify the contract, product, and licensed features. Licensed features include the Informatica
edition, deployment type, number of authorized CPUs, and authorized Informatica options and connectivity.
You use the original keys to install Informatica and create licenses for services. You must have a license key
to install Informatica. The installation program creates a license object for the domain in the Administrator
tool. You can use other original keys to create more licenses in the same domain. You use a different original
license key for each license object.
Incremental Keys
You use incremental license keys to update an existing license. You add an incremental key to an existing
license to add or remove options, such as PowerCenter options, connectivity, and Metadata Exchange
options. For example, if an existing license does not allow high availability, you can add an incremental key
with the high availability option to the existing license.
The Service Manager updates the license expiration date if the expiration date of an incremental key is later
than the expiration date of an original key. The Service Manager uses the latest expiration date. A license
object can have different expiration dates for options in the license. For example, the IBM DB2 relational
connectivity option may expire on 12/01/2006, and the session on grid option may expire on 04/01/06.
The Service Manager validates the incremental key against the original key used to create the license. An
error appears if the keys are not compatible.
You can also use the infacmd isp AddLicense command to add a license to the domain.
• Use a valid license key file. The license key file must contain an original license key. The license key file
must not be expired.
• You cannot use the same license key file for multiple licenses. Each license must have a unique original
key.
• Enter a unique name for each license. You create a name for the license when you create the license. The
name must be unique among all objects in the domain.
• Put the license key file in a location that is accessible by the Administrator tool computer. When you
create the license object, you must specify the location of the license key file.
After you create the license, you can change the description. To change the description of a license, select
the license in Navigator of the Administrator tool, and then click Edit.
Option Description
Name Name of the license. The name is not case sensitive and must be unique within the domain. It
cannot exceed 128 characters or begin with @. It also cannot contain spaces or the following
special characters:
`~%^*+={}\;:'"/?.,<>|!()][
Description Description of the license. The description cannot exceed 765 characters.
Path Path of the domain in which you create the license. Read-only field. Optionally, click Browse
and select a domain in the Select Folder window. Optionally, click Create Folder to create a
folder for the domain.
License File File containing the original key. Click Browse to locate the file.
If you try to create a license using an incremental key, a message appears that states you cannot apply
an incremental key before you add an original key.
You must use an original key to create a license.
3. Click Create.
You can use the Administrator tool or the infacmd isp UnassignLicense command to unassign a license from
a service.
You must disable a service before you can unassign a license from it. After you unassign the license from the
service, you cannot enable the service. You must assign a valid license to the service to reenable it.
You must disable the service before you can unassign the license. If you try to unassign a license from an
enabled service, a message appears that states you cannot remove the service because it is running.
When you add an incremental key to a license, the Service Manager adds or removes licensed options and
updates the license expiration date.
You can also use the infacmd isp UpdateLicense command to add an incremental key to a license.
After you update the license you must restart Informatica services for the changes to take effect.
• Verify that the license key file is accessible by the Administrator tool computer. When you update the
license object, you must specify the location of the license key file.
• The incremental key must be compatible with the original key. An error appears if the keys are not
compatible.
The Service Manager validates the license key against the original key based on the following information:
• Serial number
• Deployment type
• Distributor
• Informatica edition
• Informatica version
1. In the Administrator tool, click the Manage tab > Services and Nodes view.
2. Select a license in the Domain Navigator.
3. Click Manage Actions > Add Incremental Key.
The Update License window appears.
4. Click Choose Files to browse for a license key file.
5. Click OK.
6. In the License Details section of the Properties view, click Edit to edit the description of the license.
7. Click OK.
Removing a License
You can remove a license from a domain using the Administrator tool or the infacmd isp RemoveLicense
command.
Before you remove a license, disable all services assigned to the license. If you do not disable the services,
all running service processes abort when you remove the license. When you remove a license, the Service
Manager unassigns the license from each assigned service and removes the license from the domain. To re-
enable a service, assign another license to it.
If you remove a license, you can still view License Usage logs in the Log Viewer for this license, but you
cannot run the License Report on this license.
License Properties
You can view license details using the Administrator tool or the infacmd isp ShowLicense command.
The license details are based on all license keys applied to the license. The Service Manager updates the
existing license details when you add a new incremental key to the license.
You might review license details to determine options that are available for use. You may also review the
license details and license usage logs when monitoring licenses.
For example, you can determine the number of CPUs your company is licensed to use for each operating
system.
The Administrator tool displays the license properties in the following sections:
• License Details. View license details on the Properties tab. Shows license attributes, such as the license
object name, description, and expiration date.
• Supported Platforms. View supported platforms on the Properties tab. Shows the operating systems and
how many CPUs are supported for each operating system.
• Repositories. View the licensed repositories on the Properties tab. Shows the maximum number of
licensed repositories.
• Assigned Services. View application services that are assigned to the license on the Assigned Services
tab.
• PowerCenter Options. View the PowerCenter options on the Options tab. Shows all licensed PowerCenter
options, such as session on grid, high availability, and pushdown optimization.
• Connections. View the licensed connections on the Options tab. Shows all licensed connections. The
license enables you to use connections, such as DB2 and Oracle database connections.
• Metadata Exchange Options. View the Metadata Exchange options on the Options tab. Shows a list of all
licensed Metadata Exchange options, such as Metadata Exchange for Business Objects Designer.
You can also run the License Management Report to monitor licenses.
License Details
You can use the license details to view high-level information about the license. Use this license information
when you audit the licensing usage.
The general properties for the license appear in the License Details section of the Properties tab.
Property Description
Serial Number Serial number of the license. The serial number identifies the customer or
project. If you have multiple PowerCenter installations, there is a separate serial
number for each project. The original and incremental keys for a license have
the same serial number.
You can also use the license event logs to view audit summary reports. You must have permission on the
domain to view the logs for license events.
Supported Platforms
You assign a license to each service. The service can run on any operating system supported by the license.
One product license can support multiple operating system platforms.
The supported platforms for the license appear in the Supported Platforms section of the Properties tab.
The following table describes the supported platform properties for a license:
Property Description
Logical CPUs Number of CPUs you can run on the operating system.
Repositories
The maximum number of active repositories for the license appear in the Repositories section of the
Properties tab.
Property Description
Issued On Date on which the license was issued for this option.
Service Options
The license enables you to use Informatica Service options such as data cleansing, data federation, and
pushdown optimization.
The options for the license appear in the Service Options section of the Options tab.
Connections
The license enables you to use connections such as DB2 and Oracle database connections. The license also
enables you to use connections for PowerExchange adapters such as PowerExchange for Facebook.
The connections for the license appear in the Connections section of the Options tab.
The Metadata Exchange options for the license appear in the Metadata Exchange Options section of the
Options tab.
Monitoring
This chapter includes the following topics:
Monitoring Overview
In Informatica Administrator, you can monitor statistics for Data Integration Service jobs on the Monitor tab.
After you configure the monitoring Model repository at the domain level, you can view monitoring statistics in
the Administrator tool. The Monitor tab displays current and historical information about multiple Data
Integration Service and integration objects. Use the Summary Statistics view to view graphical summaries of
object state and distribution across multiple Data Integration Service. You can also view graphs of the
memory and CPU that the Data Integration Services used to run the objects. Use the Execution Statistics
view to monitor properties, run-time statistics, and run-time reports for integration objects.
• Ad hoc jobs
• Applications
• Logical data objects
• SQL data services
234
• Web services
• Workflows
Monitoring is a domain function that the Service Manager performs. The Service Manager stores the
monitoring configuration in the monitoring Model repository. The Service Manager also persists, updates,
retrieves, and publishes run-time statistics for integration objects in the monitoring Model repository.
The Monitoring tool is a direct link to the Monitor tab of the Administrator tool. The Monitoring tool is
useful if you do not need access to other Administrator tool features. You must have at least one
monitoring privilege to access the Monitoring tool. You can access the Monitoring tool from the
following URL:
http://<Administrator tool host><Administrator tool port>/monitoring/
The Analyst tool
You can monitor objects on the Job Status tab in the Analyst tool. The Job Status tab shows the status
of Analyst tool jobs, such as profile jobs, scorecard jobs, and jobs that load mapping specification
results to the target.
You can open the Monitoring tool from the Developer tool. When you monitor from the Developer tool,
you can view jobs that users run from the Developer tool. The Monitoring tool shows the status of
Developer tool jobs, such as mapping jobs.
Configuring Monitoring
You can configure the monitoring Model Repository Service at the domain level. After you configure the
Monitoring Configuration parameters, the Monitor tab displays the statistics and reports about the objects in
the domain. Statistics and reports appear in the History view on the Manage tab and in the Summary
Statistics and Execution Statistics views on the Monitor tab.
The monitoring Model Repository Service stores statistics and reports for the Data Integration Service jobs.
The statistics include historical information about objects that the multiple Data Integration Services run. The
reports show key metrics about integration objects.
If you do not configure monitoring, then some views on the Manage and Monitor tabs do not have content.
The workflow graph is also empty, and notifications disappear when you refresh the page.
1. Configure monitoring settings. Configure a Model repository as a monitoring Model repository to store
run-time statistics for objects that the multiple Data Integration Services runs.
2. Configure reports and statistics views. Choose which statistics appear in the Statistics and Reports
views.
Note: In a domain that uses Kerberos authentication, users must have the Administrator role for the
monitoring Model Repository Service that stores statistics. If users do not have Administrator role, some
statistics might not appear.
Create the monitoring Model repository content before you configure the monitoring settings. If you create
the content after you configure the monitoring settings, you must recycle the monitoring Model Repository
Service after the content is created.
1. In the Administrator tool, click the Manage tab > Services and Nodes view.
2. In the Domain Navigator, select the domain.
3. In the Domain section, click the Monitoring Configuration view.
The current monitoring configuration appears.
4. Click Edit to change the monitoring configuration.
5. Edit the following options:
Option Description
Model Repository Name of the Model repository that stores the historical information. The Model
Service repository must not be integrated with a version control system.
Username User name to access the Model Repository Service. Does not appear in domains that
use Kerberos authentication.
Password Password of the user name to access the Model Repository Service. Does not appear in
domains that use Kerberos authentication.
Security Domain Name of the security domain that the Model repository user belongs to.
Preserve Summary Number of days that the Model repository saves averaged data. If purging is disabled,
Historical Data then the Model repository saves the data indefinitely.
Default is 180. Minimum is 0. Maximum is 366.
Preserve Detailed Number of days that the Model repository saves per-minute data. If purging is disabled,
Historical Data then the Model repository saves the data indefinitely.
Default is 14. Minimum is 1. Maximum is 14.
Purge Statistics Every Interval, in days, at which the Model Repository Service purges data that is older than
the values configured in the Preserve Historical Data option.
Default is 1 day.
Days At Time of day when the Model Repository Service purges statistics. Default is 1:00 a.m.
Maximum Number of Maximum number of records that can be sorted in the Monitor tab. If the number of
Sortable Records records on the Monitor tab is greater than this value, then you can only sort by Start
Time and End Time. Default is 3,000.
Maximum Delay for Maximum time, in seconds, that the Data Integration Service buffers statistics before it
Update Notifications stores them in the Model repository and displays them in the Monitor tab. If the Data
Integration Service shuts down unexpectedly before it stores the statistics in the Model
repository, then the statistics are lost. Default is 10.
Show Milliseconds in Include milliseconds for date and time fields in the Monitor tab.
Date Time Field
6. Click OK.
To apply the settings, you must restart all of the Data Integration Services.
Before you configure statistics and reports, you must specify a monitoring Model Repository Service in the
Monitoring Configuration tab and enable the monitoring Model Repository Service.
1. In the Administrator tool, click the Monitor > Execution Statistics tab.
2. Click Actions > Report and Statistic Settings.
3. In the Report and Statistic Settings dialog box, click the Statistics tab.
4. Configure the time ranges that you want to use for statistics, and select the frequency at which the
statistics assigned to each time range should be updated.
5. In the Default Time Range list, select a default time range to appear for all statistics.
6. Click the Reports tab.
7. Enable the time ranges that you want to use for reports, and select the frequency at which the reports
assigned to each time range should be updated.
8. In the Default Time Range list, select a default time range to appear for all reports.
9. Click Select Reports.
10. In the Select Reports dialog box, add the reports that you want to run to the Selected Reports box.
11. Organize the reports in the order in which you want to view them on the Monitor tab.
12. Click OK to close the Select Reports dialog box.
13. Click OK in the Report and Statistic Settings dialog box.
To optimize monitoring performance, enable the Purge Statistics option in the Monitoring Configuration.
When you enable statistics purging, you configure how often the monitoring Model Repository Service purges
data. Use the Preserve options in the Monitoring Configuration to configure how long Detailed and Summary
data are saved in the monitoring Model repository before they are purged.
• Create the monitoring Model repository to store monitoring data. Specify the monitoring Model Repository
Service when you configure monitoring in the domain.
• Configure the monitoring Model Repository Service on the machine where you configure the domain.
• Verify that the monitoring Model repository database is tuned as required for the following properties:
- Maximum Heap Size property
- Memory settings
Summary Statistics
The Summary Statistics view shows information about the Data Integration Services and the objects that a
Data Integration Service runs.
Use the Summary Statistics view to view summaries of object distribution, object state, and Data Integration
Service resource usage for a specified time range. You can view statistics for the domain, or by Data
Integration Service or application.
• Ad hoc jobs. Jobs that users ran from the Developer tool or the Analyst tool.
• Deployed mappings and workflows. Mappings and workflows that are deployed in an application.
• Requests and connections. Deployed SQL data services and web services.
• Resource Usage. CPU and memory usage of the Data Integration Service processes that are running in the
domain or on a node in the domain.
For example, you can view all of the jobs in the domain that failed over the last eight hours. You can review
resource usage for that time range to see if the jobs failed due to resource issues on the node.
When you select a source and time range you can choose one of the following options to view the data:
• Graphical distribution. Displays doughnut and line charts that summarize object distribution and resource
usage for a time range. The doughnut charts show distributions of objects by state, type, and the Data
Integration Service that ran them. The line charts compare resource usage for the Data Integration Service
that ran the jobs to resource usage for all of the processes that ran on the node during that time range.
• Tabular distribution. View the total completed, running, canceled, aborted, and failed jobs.
• Details. View a list of the jobs, requests, or connections that comprise the summary statistics. You can
click the Job ID to view the specific job in the Execution Statistics.
• Export data. Export the detail data for a specific object type to a .csv file.
The Summary Statistics view displays statistics using data that is stored in the monitoring Model repository.
You must configure a monitoring Model Repository Service in the Monitoring Configuration tab before you
can view Summary Statistics.
Option Description
When you select a Data Integration Service in the Navigator, the contents panel shows the following views:
• Properties view
• Reports view
When you select a Data Integration Service in the Navigator, you can view the general properties and run-time
statistics.
You can view general properties, such as the service name, object type, and description. The Persist
Statistics Enabled property indicates whether the Data Integration Service stores persisted statistics in
the monitoring Model repository. This option is true when you configure the global settings for the
domain.
You can also view information about objects that the Data Integration Service runs. To view information
about an object, select the object in the Navigator or contents panel. Depending on the object type,
details about the object appear in the contents panel or details panel.
You can view run-time statistics about objects that the Data Integration Service runs. Select the object
type and time period to display the statistics. You can view statistics about jobs, applications,
connections, requests, and workflows. For example, you can view the number of failed, canceled, and
completed profiling jobs in the last four hours.
When you select a Data Integration Service in Navigator of the Monitor tab, the Reports view shows reports
about jobs. For example, you can view the Most Active Users for Jobs report to determine users that ran the
most jobs during a specific time period. Click a link in the report to show more details about the objects
An ad hoc job is a preview, scorecard, profile, mapping, audit, or reference table process that a user runs
from the Developer tool or the Analyst tool. When a user runs a job, a Data Integration Service runs the job
process and the job appears in the Monitor tab.
You can run up to five jobs at a time from the Developer tool. All remaining jobs are queued, and do not
appear in the Monitor tab until they run.
By default, you can monitor jobs that you run. If you have the appropriate monitoring privilege, you can also
view jobs that other users run.
When you select Ad Hoc Jobs in the Navigator of the Execution Statistics view, a list of jobs appears in the
contents panel. The contents panel groups related jobs based on the job type. You can expand a job type to
view the related jobs under it. For example, when you run a profile job, the Data Integration Service converts
the job into a mapping. The mapping appears under the profile job in the contents panel.
When you select a job in the contents panel, you can view logs for the job, view context for the job, or cancel
the job. You can also view properties for the job in the details panel. Depending on the type of job, the details
panel might show general properties, mapping properties, or statistics.
When you select Ad Hoc jobs in the Navigator of the Execution Statistics view, a list of jobs appears in the
contents panel. The contents panel groups related jobs based on the job type. You can expand a job type to
view the related jobs under it.
The following list describes the types of properties and statistics that can appear in the details panel:
Properties
Shows the general properties about the selected job, such as the name, job type, user who ran the job,
and start time of the job. If the job ran on a grid, the details panel shows the node that the job ran on.
Mapping Properties
You can view mapping properties when you select a profile or scorecard job in the contents panel. These
jobs have an associated mapping. You can view mapping properties such as the request ID, the mapping
name, and the log file name. You can also view throughput and resource usage statistics for the
associated mappings.
You can view a Blaze execution plan when you run a mapping with the Blaze engine in the Hadoop
environment. The Blaze execution plan displays the Blaze engine script that the Data Integration Service
generates based on the mapping logic, the unique identifier for the script, and the tasks that the script
depends on.
Summary Statistics
You can view summary statistics you select an ad hoc mapping job, deployed mapping job, or mapping
object in a workflow in the contents panel. The Summary Statistics view displays throughput and
resource usage statistics for the job run.
You can also view the summary statistics for the source or target on the Summary Statistics tab, such
as processed rows and bytes.
Note: Only the number of rows processed appears in the Summary Statistics for the Hive source and
target. The remaining property values appears as 0 or N/A for the Hive source or target.
Detailed Statistics
You can view detailed statistics for an ad hoc mapping job, deployed mapping job, or mapping object in a
workflow in the contents panel. Detailed statistics appear for jobs that run longer than one minute in
separate local processes. The Detailed Statistics view displays graphs of the throughput and resource
usage statistics for the job run.
The following image shows the Detailed Statistics view for a mapping job in a workflow:
Historical Statistics
You can view historical statistics when you select an ad hoc mapping job, deployed mapping job, or
mapping object in a workflow in the Contents panel. The Historical Statistics view shows averaged data
from the last 500 runs for a specific job. For example, you can view the minimum, maximum, and
average duration of the mapping job. You can view the average amount of CPU that the job consumes
The following image shows the Historical Statistics view for a mapping job that completed three times
and failed one time:
You can get aggregated cluster logs for deployed Hadoop mappings, Databricks mappings, auto-deploy
cluster jobs, local files, and Spark data preview jobs based on the job ID. You can get a .zip or tar.gz file
of the aggregated cluster logs for a job and write the compressed aggregated log file to a target
directory. You can also use the infacmd ms fetchAggregatedClusterLogs command or use the REST
API to collect the aggregated cluster logs for a job.
For more information about the infacmd ms fetchAggregatedClusterLogs command, see the
Informatica 10.4.0 Command Reference.
Task Description
Enlarge a graph Move the cursor over a graph, and then click the magnifying glass
icon.
Enlarge a section of an enlarged graph Drag the cursor to select an area to enlarge.
Switch between rows and bytes in the Click the Bytes option or the Rows option.
throughput graph
Choose which statistics are plotted on the In the throughput field, select the sources and targets that you want
throughput graph to view.
If an audit configuration contains both rules that run before the mapping and rules that run after the mapping,
separate audit jobs run for the pre-mapping and post-mapping jobs.
Monitor Applications
You can monitor applications on the Monitor tab.
When you select an application in the Navigator of the Execution Statistics view, the contents panel shows
the following views:
• Properties view
• Reports view
You can expand an application in the Navigator to monitor the application components.
When you select an application in the Navigator of the Execution Statistics view, you can view general
properties and run-time statistics.
You can view general properties, such as the name and description of the application. You can also view
additional information about the objects in an application. To view information about an object, select
the folder in the Navigator and the object in the contents panel. Details about the object appear in the
details panel.
Statistics
You can view run-time statistics about an application and about the jobs, connections, requests, and
workflows associated with the application. For example, you can view the number of enabled and
disabled applications, number of aborted connections, and number of completed, failed, and canceled
jobs and workflows.
When you select an application in the Navigator of the Execution Statistics view, the Reports view shows
reports about the application components.
• Monitoring tool. In the Developer tool, click the Menu button in the Progress view and select Monitor
Jobs. Select the Data Integration Service that runs the mapping and click OK. The Monitoring tool opens.
Expand an application in the Navigator and select the Deployed Mapping Jobs folder. A list of deployed
mapping jobs appears in the contents panel.
You can view mapping execution statistics in the Monitoring tool. The REST Operations Hub generates
statistics based on the API configuration.
• Administrator tool. Expand an application in the Navigator and select the Deployed Mapping Jobs folder.
A list of deployed mapping jobs appears in the contents panel.
The contents panel shows properties about each deployed mapping job, such as Job ID, name of the
mapping, state of the job, and start time of the job. If you run the job on a grid, the contents panel also shows
the node that the Data Integration Service that runs the process is running on.
Select a deployed mapping job in the contents panel to view logs for the job, reissue the job, cancel the job,
or view statistics about the job. You can view statistics about throughput and resource usage for the job run.
Note: The log contents for a deployed mapping job depend on how the Data Integration Service is configured.
For more information about logs when a Data Integration Service grid is configured to run jobs in separate
remote processes, see the Informatica Application Service Guide.
Task Description
Enlarge a graph Move the cursor over a graph, and then click the magnifying glass
icon.
Enlarge a section of an enlarged graph Drag the cursor to select an area to enlarge.
Switch between rows and bytes in the Click the Bytes option or the Rows option.
throughput graph
Choose which statistics are plotted on the In the throughput field, select the sources and targets that you want
throughput graph to view.
You can view information about logical data objects included in an application. To monitor a logical data
object, expand a Data Integration Service in the Navigator. Expand an application, and then select the Logical
Data Objects folder. A list of logical data objects appears in the contents panel. The contents panel shows
properties about each logical data object.
Select a logical data object in the contents panel to download the logs for a data object.
When you select a logical data object in the contents panel, the details panel shows the following views:
• Properties view
• Cache Refresh Runs view
You can view properties such as the data object name, logical data object model, folder path, cache state,
and last cache refresh information.
The Cache Refresh Runs view shows cache refresh details such as the cache run ID, request count, and row
count.
You can view information about the SQL data services included in an application. To monitor an SQL data
service, expand an application in the Navigator and select the SQL Data Services folder. A list of SQL data
services appears in the contents panel. The contents panel shows properties about each SQL data service,
such as the name, description, and state.
When you select an SQL data service in the contents panel, the contents panel shows the following views:
• Properties view
• Connections view
• Requests view
• Virtual Tables view
• Reports view
When you select an SQL data service in the contents panel of the Properties view, you can view the general
properties and run-time statistics.
You can view general properties, such as the SQL data service name and the description.
You can view run-time statistics about connections and requests for the SQL data service. Sample
statistics include the number of connections to the SQL data service, the number of requests, and the
number of aborted connections.
When you select a connection in the contents panel, you can abort the connection or access the Properties
view and Requests view in the details panel.
Properties View
The Properties view in the details panel shows the user who is using the connection, the state of the
connection, and the connect time.
Requests View
The Requests view in the details panel shows information about the requests for the SQL connection.
Each connection can have more than one request. The view shows request properties such as request
ID, user name, state of the request, start time, elapsed time, and end time.
The Requests view shows properties about the requests for the SQL connection. Each connection can have
more than one request. The view shows request properties such as request ID, connection ID, user name,
state of the request, start time, elapsed time, and end time.
Select a request in the contents panel to view additional information about the request in the details panel.
The view shows properties about the virtual tables, such as the name and description. When you select a
virtual table in the contents panel, you can view the Properties view and Cache Refresh Runs view in the
details panel.
Properties View
The Properties view displays general information and run-time statistics about the selected virtual table.
General properties include the virtual table name and the schema name. Monitoring statistics include the
number of request, the number of rows cached, and the last cache refresh time.
The Cache Refresh Runs view displays cache information for the selected virtual table. The view
includes the cache run ID, the request count, row count, and the cache hit rate. The cache hit rate is the
total number of requests on the cache divided by the total number of requests for the data object.
When you monitor an SQL data service, the Reports view shows reports about the SQL data service. For
example, you can view the Most Active SQL Connections report to determine the SQL connections that
received the most connection requests during a specific time period.
You can view information about web services included in an application. To monitor a web service, expand an
application in the Navigator and select the Web Services folder. A list of web services appears in the
contents panel. The contents panel shows properties about each web service, such as the name, description,
and state of each web service.
When you select the link for a web service in the contents panel, the contents panel shows the following
views:
• Properties view
• Reports view
• Operations view
• Requests view
When you select a web service in the contents panel of the Properties view, you can view the general
properties and monitoring statistics.
You can view general properties about the web service, such as the name and type of object.
You can view run-time statistics about web service requests during a specific time period. The Statistics
section shows the number of completed, failed, and total web service requests.
When you monitor a web service, the Reports view shows reports about the web service. For example, you
can view the Most Active WebService Client IP report to determine the IP addresses that received the most
number of web service requests during a specific time period.
When you select a web service operation in the contents panel, the details panel shows the Properties view,
Requests view, and Reports view.
Properties View
The Properties view shows general properties and statistics about the selected web service operation or
resource. General properties include the operation or resource name and type of object. The view also
Requests View
The Requests view shows properties about each web service operation, such as request ID, user name,
state, start time, elapsed time, and end time. You can filter the list of requests. You can also view logs
for the selected web service request.
The Reports view shows reports about SOAP web service operations.
When you select a web service request in the contents panel, you can view logs about the request in the
details panel. The details panel shows general properties and statistics about the selected web service
request. Statistics include the number of completed, failed, and total web service requests.
You can also abort a web service request from the Requests view. To abort a web service request, select the
workflow request and click Actions > Abort Selected Request in the contents panel.
Monitor Workflows
You can monitor workflows on the Execution Statistics view on the Monitor tab.
You can view information about workflow instances that run from a workflow in a deployed application. To
monitor a workflow, expand an application in the Navigator and select the Workflows folder. A list of
workflow instances appears in the contents panel. The contents panel shows properties about each workflow
instance, such as the name, state, start time, and recovery properties of each workflow instance. If the
workflow instance ran on a grid, the contents panel shows the node that ran each mapping in the workflow
instance.
Select a workflow instance in the contents panel to perform the following tasks:
Workflow Graph
You can view the details of a workflow that you run in the Monitoring tool in a graphical form.
After you run a workflow, you can see the graphical view of the workflow in the Monitoring tool. In the
workflow graph, you can see the sequential run of the mapping tasks in the workflow. The workflow graph
enables you to view the failure points in a workflow at a glance.
You can perform the following tasks from the workflow graph:
Workflow objects include events, tasks, and gateways. When you monitor workflows, you can monitor the
tasks that run in a workflow instance. The Monitor tab does not display information about events or
gateways in the workflow instance.
If an expression in a conditional sequence flow evaluates to false, the Data Integration Service does not run
the next object or any of the subsequent objects in that branch. The Monitor tab does not list objects that do
not run in the workflow instance. When a workflow instance includes objects that do not run, the instance can
still successfully complete.
You can expand a task in the contents panel to view information about the work item run by the task. For
example, if the workflow contains a mapping task, you can view throughput and resource usage statistics for
the mapping run.
Task Description
Enlarge a graph Move the cursor over a graph, and then click the magnifying glass
icon.
Enlarge a section of an enlarged graph Drag the cursor to select an area to enlarge.
Switch between rows and bytes in the Click the Bytes option or the Rows option.
throughput graph
Choose which statistics are plotted on the In the throughput field, select the sources and targets that you want
throughput graph to view.
Workflow States
When you monitor a workflow instance, you can view the state of the workflow instance. If a workflow
instance recovers after a task is interrupted, the Monitor adds an entry for the task instance that runs in the
recovered workflow.
Aborted
A workflow instance aborts when you choose to abort the workflow instance from the Monitoring tool or
using the infacmd wfs abortWorkflow command. You can also choose to abort a running workflow
instance when you stop the application that contains the workflow or when you disable the workflow in
the application.
Canceled
You choose to cancel the workflow instance from the Monitor tab or by using the infacmd wfs
cancelWorkflow command.
The workflow can also enter a canceled state if the Data Integration Service shuts down unexpectedly. If
the workflow is not configured for automatic recovery, the service process changes the workflow
instance state to Canceled when the service process restarts. Before the Data Integration Service
restarts, the workflow state and the task state appear as Running, although the workflow and the task
are no longer running. If the workflow is configured for automatic recovery, the service process recovers
the workflow instance and reruns the interrupted task when the service process restarts. The service
process sets the workflow instance state to Running.
Completed
The Data Integration Service successfully completes the workflow instance. A completed workflow
instance might indicate that all tasks, gateways, and sequence flow evaluations either successfully
completed or were in a branch that did not run.
A workflow can also enter a Completed state if a Command, Mapping, Notification, or Human task
encounters a recoverable error or nonrecoverable error. When the task encounters the error, the Data
Integration Service fails the task. The Data Integration Service runs subsequent workflow objects if
expressions in the conditional sequence flows evaluate to true or if the sequence flows do not include
conditions. If the workflow instance finishes running without another interruption, the Data Integration
Service updates the workflow state to Completed.
When the task fails, the Data Integration Service continues to run additional objects in the workflow
instance if expressions in the conditional sequence flows evaluate to true or if the sequence flows do
not include conditions. If the workflow instance finishes running without another interruption, the Data
Integration Service updates the workflow state to completed. A completed workflow instance can
contain both failed and completed tasks.
Failed
A workflow instance fails when a workflow error occurs. Workflow errors can occur when the Data
Integration Service reads the parameter file at the start of the workflow run, copies workflow parameter
and variable values to task input, or evaluates expressions in conditional sequence flows. In addition, a
workflow error occurs if an Assignment task or a gateway fails.
When a workflow error occurs, the Data Integration Service stops processing additional objects and fails
the workflow instance immediately. Workflow errors are nonrecoverable.
Running
Completed
Failed
• Any task in a workflow not enabled for recovery encounters any type of error.
• An Assignment task in a workflow enabled for recovery encounters any type of error.
• A Command, Mapping, Notification, or Human task with a restart recovery strategy in a workflow
enabled for recovery encounters a non recoverable error.
• A Mapping task with a skip recovery strategy in a workflow enabled for recovery encounters any type
of error.
Note: A workflow can complete if a task fails. The Data Integration Service runs subsequent workflow
objects if expressions in the conditional sequence flows evaluate to true or if the sequence flows do not
include conditions. If the workflow instance finishes running without another interruption, the Data
Integration Service updates the workflow state to Completed.
Running
You can also view the state of the mapping run from the workflow graph of the workflow that contains the
mapping task.
Mappings run by a Mapping task can have one of the following states:
Aborted
The Mapping task aborts while the mapping is running because you choose to abort the workflow
instance.
Completed
Failed
The mapping encounters an error. The mapping and the Mapping task appear as Failed in the Monitor.
The states do not depend on the Mapping task recovery strategy.
When you cancel a workflow instance, the Data Integration Service finishes processing any running task and
then stops processing the workflow instance. The service does not start running any subsequent workflow
objects.
When you abort a workflow instance, the Data Integration Service attempts to kill the process on any running
task. If an Assignment task or a gateway is running, the Data Integration Service completes the task or
gateway. After the task aborts or completes, the service aborts the workflow instance. The service does not
start running any subsequent workflow objects.
You can also cancel or abort a workflow from the workflow graph.
Workflow Recovery
Workflow recovery is the completion of a workflow instance from the point of interruption.
When a workflow is enabled for recovery, you can recover a workflow instance if a task encounters a
recoverable error, if you cancel the workflow instance, or if the Data Integration Service process shuts down
unexpectedly.
View the workflow log to identify the cause of the interruption. After fixing any recoverable errors, you can
recover the interrupted workflow instance if it is enabled for recovery.
You cannot change a workflow definition between the interrupted run and the recovery run. If a workflow
instance has a recoverable state and you change the workflow metadata in the Developer tool and redeploy
the application that contains the workflow, then the workflow instance is no longer recoverable.
The Data Integration Service tries to recover the previous workflow state if the service restarts after an
unexpected shutdown. By default, the Data Integration Service does not recover a workflow instance that
stopped during a Command task, Mapping task, or Notification task. In addition, the Data Integration Service
cannot recover a workflow instance by default if you cancel the workflow instance or cancel a running task in
the workflow instance. You can configure the recovery options on the workflow to enable the Data Integration
Service to recover a workflow instance in such cases.
When you configure the workflow options, you can configure the workflow for manual recovery or automatic
recovery. If you configure automatic recovery, the Data Integration Service restarts the workflow from the
point of interruption without any human interaction. If you configure manual recovery, you can restart the
workflow.
When a workflow instance recovers or when you recover a workflow instance, the Data Integration Service
restarts the task. The service continues processing the subsequent workflow objects. If a workflow instance
recovers after a task is interrupted, the Monitor adds an entry for the task instance that runs in the recovered
Recovery Properties
The read-only recovery properties display for each workflow instance. You configure the recovery properties
for the workflow definition in the Developer tool. You cannot change the values of the properties for the
workflow instance.
The following table describes the read-only recovery properties for a workflow instance:
Property Description
Automatically Recover Indicates that the Data Integration Service process tries to automatically recover
Workflows workflow instances that were interrupted. The workflow recovery starts after the Data
Integration Service process restarts.
Recovering a Workflow
You can recover interrupted workflow instances that are enabled for recovery.
Workflow Logs
The Data Integration Service generates log events when you run a workflow. Log events include information
about workflow errors, task progress, and the setting of workflow variables. Log events also include the
analyses of the links that the Data Integration Service evaluates in a sequence flow.
If a workflow instance includes a Mapping task, the Data Integration Service generates a separate log file for
the mapping. The mapping log file includes any errors encountered during the mapping run and load
summary and transformation statistics.
You can view the workflow and mapping logs from the Monitor tab.
When you recover an interrupted workflow instance, the Data Integration Service appends log events to the
current workflow log. When the recovered workflow instance includes a Mapping task that is restarted, the
Data Integration Service creates a mapping log.
If the workflow runs on a grid, the recovery of the workflow instance might run on a different node than the
original workflow instance run. If the recovery runs on a different node and the log directory is not in a shared
location, the Data Integration Service creates a log file with the same name on the current node.
The Data Integration Service writes information to the workflow log when the following types of event occur:
• The Data Integration Service starts to run a task or another object in the workflow.
• A task or another object in the workflow is in progress.
• The Data Integration Service finishes running a task or another object in the workflow.
• The Data Integration Service sets or updates a workflow variable.
• The Data Integration Service evaluates the links in a sequence flow and identifies the correct path for the
workflow process.
• The workflow encounters a workflow error.
If the status of a job is not known, the Monitoring tool reports the status as UNKNOWN. When the monitoring
Model Repository Service restarts or fails over to a backup node, it updates the Monitoring tool with the
latest status of each job if the job is still running. The status of a job that completes before the monitoring
Model Repository Service becomes available remains as UNKNOWN.
For example, a Data Integration Service runs on a grid. Two mappings are running when the monitoring Model
Repository Service fails. The Monitoring tool does not have the latest status of these mappings. One
mapping completes successfully before the monitoring Model Repository Service becomes available. The
other mapping continues to run after the monitoring Model repository Service becomes available. The
Monitoring tool reports the status of the first mapping as UNKNOWN. It shows the status of the second
mapping as RUNNING.
You can apply a filter to limit the number of objects that appear in the contents panel. You can create custom
filters based on a time range. Custom filters allow you to select particular dates and times for job start times,
end times, and elapsed times. Custom filters also allow you to filter results based on multiple filter criteria.
For example, you notice that your deployed mapping failed. When you view the context of the deployed
mapping, an unfiltered list of deployed mappings appears in a separate working view, showing you all
deployed mappings that started around the same time as your deployed mapping. You notice that the other
deployed mappings also failed. You determine that the cause of the problem is that the Data Integration
Service was unavailable.
1. Select Custom as the filter option for the Start Time or End Time column.
The Custom Filter: Date and Time dialog box appears.
2. Enter the date range using the specified date and time formats.
3. Click OK.
1. Select Custom as the filter option for the Elapsed Time column.
The Custom Filter: Elapsed Time dialog box appears.
2. Enter the time range.
3. Click OK.
Log Management
This chapter includes the following topics:
The Log Manager runs on the master gateway node. It collects and processes log events for Service Manager
domain operations, application services, and user activity. The log events contain operational and error
messages for a domain. The Service Manager and the application services send log events to the Log
Manager. When the Log Manager receives log events, it generates log event files. You can view service log
events in the Administrator tool based on criteria that you provide.
The Log Agent runs on all nodes in the domain. The Log Agent retrieves the workflow and session log events
that the PowerCenter Integration Service writes and displays them in the Workflow Monitor. Workflow log
events include information about workflow processing, workflow errors, and tasks that the PowerCenter
Integration Service performs. Session log events include information about the tasks performed by the
PowerCenter Integration Service, session errors, and load summary and transformation statistics for the
session. You can view log events for the last workflow run with the Log Events window in the Workflow
Monitor.
The Log Agent also collects and processes log events for jobs that the Data Integration Service runs. These
include profile jobs, scorecard jobs, preview jobs, mapping jobs, and SQL data services. You can view log
events for these jobs on the Monitoring tab.
265
Log event files are binary files that the Administrator tool Logs Viewer uses to display log events. When you
view log events in the Administrator tool, the Log Manager uses the log event files to display the log events
for the domain, application services, and user activity.
Domain logs include domain, application service, and user activity logs. You can view them in the
Administrator tool. System logs are for use only by Informatica Support to address open support issues.
You can use the Administrator tool to perform the following tasks with the Log Manager:
• Configure the log location. Configure the node that runs the Log Manager, the directory path for log event
files, purge options, and time zone for log events.
• Configure log management. Configure the Log Manager to purge logs or purge logs manually. Save log
events to XML, text, or binary files. Configure the time zone for the time stamp in the log event files.
• View log events. View domain function, application service, and user activity log events on the Logs tab.
Filter log events by domain, application service type, and user.
• Log event files. Stores log events in binary format. The Log Manager creates log event files to display log
events in the Logs tab. When you view events in the Administrator tool, the Log Manager retrieves the log
events from the event nodes.
The Log Manager stores the files by date and by node. Set the directory path with the infasetup tools
defineDomain command -ld option.
• Guaranteed Message Delivery files. Stores domain, application service, and user activity log events. The
Service Manager writes the log events to temporary Guaranteed Message Delivery files and sends the log
events to the Log Manager.
If the Log Manager becomes unavailable, the Guaranteed Message Delivery files stay in the default log
directory on the node where the service runs. By default, the directory path is
<Informatica_installation_directory>/logs/<Node_Name>. When the Log Manager becomes
available, the Service Manager for the node reads the log events in the temporary files, sends the log
events to the Log Manager, and deletes the temporary files.
The Log Manager performs the following tasks to process PowerCenter session and workflow log events:
1. During a session or workflow, the PowerCenter Integration Service writes binary log files on the node. It
sends information about the logs to the Log Manager.
2. The Log Manager stores information about workflow and session logs in the domain database. The
domain database stores information such as the path to the log file location, the node that contains the
log, and the PowerCenter Integration Service that created the log.
The Data Integration Service and the Log Manager perform the following tasks to process job log events for
the Data Integration Service:
1. When the Data Integration Service runs a job, it writes log events to text files on the node. The Data
Integration Service sends information about the logs to the Log Manager.
2. The Log Manager stores log information in the Model repository database. The Model repository
database stores information such as the path to the log file location, the node that contains the log, and
the Data Integration Service that created the log.
3. When you view a job log on the Monitor tab of the Administrator tool, the Log Manager retrieves the
information from the Model repository database. The Log Manager uses the information to determine
the location of the logs.
4. The Log manager dispatches a Log Agent to retrieve the log events on each node to display the log.
The Service Manager, the application services, and the Log Manager perform the following tasks:
1. An application service process writes log events to a Guaranteed Message Delivery file.
2. The application service process sends the log events to the Service Manager on the gateway node for
the domain.
3. The Log Manager processes the log events and writes log event files. The application service process
deletes the temporary file.
4. If the Log Manager is unavailable, the Guaranteed Message Delivery files stay on the node running the
service process. The Service Manager for the node sends the log events in the Guaranteed Message
Delivery files when the Log Manager becomes available, and the Log Manager writes log event files.
Use the Service Manager log files to troubleshoot issues when the Log Manager cannot process log events.
You will also need to use these files to troubleshoot issues when you contact Informatica Global Customer
Support.
Log Location
The Service Manager on the master gateway node writes log event files to the log file directory. When you
configure a node to serve as a gateway, you must configure the directory where the Service Manager on this
node writes the log event files. Each gateway node must have access to the directory path.
You configure the log location in the Properties view for the domain. Configure a directory location that is
accessible to the gateway node during installation or when you define the domain. Store the logs on a shared
disk when you have more than one gateway node. If the Log Manager is unable to write to the directory path,
it writes log events to node.log on the master gateway node.
When you configure the log location, the Administrator tool validates the directory as you update the
configuration. If the directory is invalid, the update fails. The Log Manager verifies that the log directory has
read/write permissions on startup. Log files might contain inconsistencies if the log directory is not shared in
a highly available environment.
You can change the directory path for domain logs in the Administrator tool or with the log service directory
parameter, -ld. You can use the -ld parameter with any of the following commands:
System Logs
System logs contain information that Informatica Support views to help solve issues that you raise with
Support. Ordinarily, you have no need to view these logs.
• infasetup DefineDomain
• infasetup DefineGatewayNode
• infasetup DefineWorkerNode
• infasetup UpdateGatewayNode
• infasetup UpdateWorkerNode
When you create a custom location, you can use a local location or a location that all domain nodes share.
The Service Manager adds the node name to the path and creates separate log directories for each node.
When you update the gateway node or worker node with a new default location for system logs, existing logs
remain intact. The server creates future logs at the new location, and abandons logs at the old location.
If you have multiple Informatica domains, you must configure a different directory path for the Log Manager
in each domain. Multiple domains cannot use the same shared directory path.
Note: When you change the directory path, you must restart Informatica services on the node you changed.
You can purge logs events periodically to manage the amount of log events stored by the Log Manager. You
can export logs before you purge them to keep a backup of the log events.
When the number of days or the size of the log directory exceeds the limit, the Log Manager deletes the log
event files, starting with the oldest log events. The Log Manager periodically verifies the purge options and
purges log events. The Log Manager does not purge the current day log event files and folder.
Option Description
Preserve logs for number of days Number of days to preserve logs. Default is 30.
Maximum size for logs in MB Number of megabytes of disk space to store logs. Default is 200.
Note: The Log Manager does not purge PowerCenter session and workflow log files.
Optionally, you can use the infacmd PurgeLog command to purge log events.
Option Description
Log Type Type of log events to purge. You can purge domain, service, user activity or all log events.
Service Type When you purge application service log events, you can purge log events for a particular
application service type or all application service types.
Purge Entries Date range of log events you want to purge. You can select the following options:
- All Entries. Purges all log events.
- Before Date. Purges log events that occurred before this date.
Use the yyyy-mm-dd format when you enter a date. Optionally, you can use the calendar to choose
the date. To use the calendar, click the date field.
Time Zone
When the Log Manager creates log event files, it generates a time stamp based on the time zone for each log
event. The log event time stamp includes date, time, and time zone information. When the Log Manager
creates log folders, it labels folders according to a time stamp. When you export or purge log event files, the
Log Manager uses this property to calculate which log event files to purge or export. Set the time zone to the
location of the machine that stores the log event files.
Verify that you do not lose log event files when you configure the time zone for the Log Manager. If the
application service that sends log events to the Log Manager is in a different time zone than the master
gateway node, you may lose log event files you did not intend to delete. Configure the same time zone for
each gateway node.
Note: When you change the time zone, you must restart Informatica Services on the node that you changed.
• View log events and the Administrator tool operational errors. View log events for the domain, an
application service, or user activity.
• Filter log event results. After you display the log events, you can display log events that match filter
criteria.
• Configure columns. Configure the columns you want the Logs tab to display.
• Save log events. You can save log events in XML, text, and binary format.
• Purge log events. You can manually purge log events.
• Copy log event rows. You can copy log event rows.
To view more information about a log event, click the log event in the search results.
On AIX and Linux, if the Log Manager receives an internal error message from the PowerCenter Integration
Service, it writes a stack trace to the log event window.
You can view logs to get more information about errors that you receive while working in the Administrator
tool.
Service Service Name Name of the application service for which you want to view log events. You
can choose a single application service name or all application services.
Domain, Severity The Log Manager returns log events with this severity level.
Service
User Activity User User name for the Administrator tool user.
User Activity Security Domain Security domain to which the user belongs.
Domain, Timestamp Date range for the log events that you want to view. You can choose the
Service, User following options:
Activity - Blank. View all log events.
- Within Last Day
- Within Last Month
- Custom. Specify the start and end date.
Default is Within Last Day.
Domain, Thread Filter criteria for text that appears in the thread data. You can use
Service wildcards (*) in this text field.
Domain, Message Code Filter criteria for text that appears in the message code. You can also use
Service wildcards (*) in this text field.
Domain, Message Filter criteria for text that appears in the message. You can also use
Service wildcards (*) in this text field.
Domain, Node Name of the node for which you want to view log events.
Service
Domain, Process Process identification number for the Windows or UNIX service process
Service that generated the log event. You can use the process identification
number to identify log events from a process when an application service
runs multiple processes on the same node.
User Activity Activity Code Filter criteria for text that appears in the activity code. You can also use
wildcards (*) in this text field.
User Activity Activity Filter criteria for text that appears in the activity. You can also use
wildcards (*) in this text field.
• Category
• Service Type
• Service Name
• Severity
• User
• Security Domain
Note: The columns appear based on the query options that you choose. For example, when you display a
service type, the service name appears in the Logs tab.
The Log Manager does not delete the log events when you save them. The Administrator Tool prompts you to
save or open the saved log events file.
Optionally, you can use the infacmd isp GetLog command to retrieve log events.
The format you choose to save log events to depends on how you plan to use the exported log events file:
• XML file. Use XML format if you want to analyze the log events in an external tool that uses XML or if you
want to use XML tools, such as XSLT.
• Text file. Use a text file if you want to analyze the log events in a text editor.
• Binary file. Use binary format to back up the log events in binary format. You might need to use this
format to send log events to Informatica Global Customer Support.
When you export log events, you can choose which logs you want to save. When you choose Service logs, you
can export logs for a particular service type. You can choose the sort order of the log events in the export file.
The Log Manager does not delete the log events when you export them. The Administrator tool prompts you
to save or open the exported log events file.
Optionally, you can use the infacmd GetLog command to retrieve log events.
• XML file. Use XML format if you want to analyze the log events in an external tool that uses XML or if you
want to use XML tools, such as XSLT.
• Text file. Use a text file if you want to analyze the log events in a text editor.
• Binary file. Use binary format to back up the log events in binary format. You might need to use this
format to send log events to Informatica Global Customer Support.
The following table describes the export log options for each log type:
Service Type Service Type of application service for which to export log events. You can also
export log events for all service types.
Export Entries Domain, Date range of log events you want to export. You can select the
Service, following options:
User - All Entries. Exports all log events.
Activity - Before Date. Exports log events that occurred before this date.
Use the yyyy-mm-dd format when you enter a date. Optionally, you can
use the calendar to choose the date. To use the calendar, click the date
field.
Export logs in descending Domain, Exports log events starting with the most recent log events.
chronological order Service,
User
Activity
XML Format
When you export log events to an XML file, the Log Manager exports each log event as a separate element in
the XML file. The following example shows an excerpt from a log events XML file:
<log xmlns:xsd="http://www.w3.org/2001/XMLSchema" xmlns:common="http://
www.informatica.com/pcsf/common" xmlns:metadata="http://www.informatica.com/pcsf/
metadata" xmlns:domainservice="http://www.informatica.com/pcsf/domainservice"
xmlns:logservice="http://www.informatica.com/pcsf/logservice" xmlns:xsi="http://
www.w3.org/2001/XMLSchema-instance">
<logEvent xsi:type="logservice:LogEvent" objVersion="1.0.0" timestamp="1129098642698"
severity="3" messageCode="AUTHEN_USER_LOGIN_SUCCEEDED" message="User Admin successfully
logged in." user="Admin" stacktrace="" service="authenticationservice"
serviceType="PCSF" clientNode="sapphire" pid="0" threadName="http-8080-Processor24"
context="" />
<logEvent xsi:type="logservice:LogEvent" objVersion="1.0.0" timestamp="1129098517000"
severity="3" messageCode="LM_36854" message="Connected to node [garnet] on outbound
connection [id = 2]." user="" stacktrace="" service="Copper" serviceType="IS"
clientNode="sapphire" pid="4484" threadName="4528" context="" />
Text Format
When you export log events to a text file, the Log Manager exports the log events in Information and Content
Exchange (ICE) Protocol. The following example shows an excerpt from a log events text file:
2006-02-27 12:29:41 : INFO : (2628 | 2768) : (IS | Copper) : sapphire : LM_36522 :
Started process [pid = 2852] for task instance Session task instance
[s_DP_m_DP_AP_T_DISTRIBUTORS4]:Executor - Master.
Binary Format
When you export log events to a binary file, the Log Manager exports the log events to a file that Informatica
Global Customer Support can import. You cannot view the file unless you convert it to text. You can use the
infacmd ConvertLogFile command to convert binary log files to text files, XML files, or readable text on the
screen.
Log Events
The Service Manager and application services send log events to the Log Manager. The Log Manager
generates log events for each service type.
Log events include a timestamp, in milliseconds, and a thread name that identifies the event.
You can view the following log event types on the Logs tab:
• Domain log events. Log events generated from the Service Manager functions.
• Analyst Service log events. Log events about each Analyst Service running in the domain.
• Content Management Service log events. Log events about each Content Management Service running in
the domain.
• Data Integration Service log events. Log events about each Data Integration Service running in the
domain.
• Metadata Manager Service log events. Log events about each Metadata Manager Service running in the
domain.
• Model Repository log events. Log events about each Model Repository Service running in the domain.
• PowerCenter Integration Service log events. Log events about each PowerCenter Integration Service
running in the domain.
• PowerCenter Repository Service log events. Log events from each PowerCenter Repository Service
running in the domain.
• Resource Manager Service log events. Log events about the Resource Manager Service running in the
domain.
• SAP BW Service log events. Log events about the interaction between the PowerCenter and the SAP
NetWeaver BI system.
• Service type, category, or user. The Logs tab categorizes events by domain category, service type, or user.
If you view application service logs, the Logs tab displays the application service names. When you view
domain logs, the Logs tab displays the domain categories in the log. When you view user activity logs, the
Logs tab displays the users in the log.
• Message or activity. Message or activity text for the log event. Use the message text to get more
information about the log events for domain and application services. Use the activity text to get more
information about log events for user activity. Some log events contain embedded log event in the
message texts. For example, the following log events contains an embedded log event:
Client application [PmDTM], connection [59]: recv failed.
In this log event, the following log event is the embedded log event:
[PmDTM], connection [59]: recv failed.
When the Log Manager displays the log event, the Log Manager displays the severity level for the
embedded log event.
• Security domain. When you view user activity logs, the Logs tab displays the security domain for each
user.
• Message or activity code. Log event code. If the message type is error or fatal, click on the message code
to open the Informatica Knowledge Base search for the message. You must configure the support portal
credentials in the user account to do the search.
• Process. The process identification number for the Windows or UNIX service process that generated the
log event. You can use the process identification number to identify log events from a process when an
application service runs multiple processes on the same node.
• Node. Name of the node running the process that generated the log event.
• Thread. Identification number or name of a thread started by a service process.
• Time stamp. Date, time, and time zone of the log event that occurred.
• Severity. The severity level for the log event. When you view log events, you can configure the Logs tab to
display log events for a specific severity level.
Use the domain log events to view information about the domain and troubleshoot issues. You can use the
domain log events to troubleshoot issues related to the startup and initialization of nodes and application
services for the domain.
• Authorization. Log events that occur when the Service Manager authorizes user requests for services.
Requests can come from the Administrator tool.
Note: Sensitive information might not appear in the domain log. For example, in case of a user
authentication failure, the user and security domain for failed login attempts do not appear. The User
Activity log contains full information.
• Container Management. Log events that occur when the Service Manager manages containers on nodes
with the compute role.
• Domain Configuration. Log events that occur when the Service Manager manages the domain
configuration metadata.
• Licensing. Log events that occur when the Service Manager registers license information.
• License Usage. Log events that occur when the Service Manager verifies license information from
application services.
• Log Manager. Log events from the Log Manager. The Log Manager runs on the master gateway node. It
collects and processes log events for Service Manager domain operations and application services.
• Log Agent. Log events from the Log Agent. The Log Agent runs on all nodes in the domain. It retrieves
PowerCenter workflow and session log events to display in the Workflow Monitor.
• Monitoring. Log events about Domain Functions.
• Node Configuration. Log events that occur as the Service Manager manages node configuration metadata
in the domain.
• User Management. Log events that occur when the Service Manager manages users, groups, roles, and
privileges.
• Service Manager. Log events from the Service Manager and signal exceptions from DTM processes. The
Service Manager manages all domain operations. If the error severity level of a node is set to Debug, when
a service starts the log events include the environment variables used by the service.
• Managing projects. Log events about managing projects in the Informatica Analyst, such as creating
objects, folders, and projects. Log events about creating profiles, scorecards, and reference tables.
• Running jobs. Log events about running profiles and scorecards. Logs about previewing data.
• User permissions. Log events about managing user permissions on projects.
• Configuration. Log events about system or service configuration changes, application deployment or
removal, and logs about the associated profiling warehouse.
• Data Integration Service processes. Log events about application deployment, data object cache refresh,
and user requests to run mappings, jobs, or workflows.
• Service failures. Log events about failures that cause the Data Integration Service to be unavailable, such
as Model repository connection failures or the service's failure to start.
• Client communication. Log events for communication between a PowerCenter or PowerExchange client
and a data source.
• Listener service. Log events about the Listener service, including configuring, enabling, and disabling the
service.
• Listener service operations. Log events for operations such as managing bulk data movement and change
data capture.
• Connections. Log events about connections between the Logger Service and the source databases.
• Logger service. Log events about the Logger Service, including configuring, enabling, and disabling the
service.
• Logger service operations. Log events for operations such as capturing changed data and writing the data
to PowerExchange Logger files.
• Model repository connections. Log events for connections to the repository from Informatica Developer,
Informatica Analyst, and the Data Integration Service.
• Model Repository Service. Log events about the Model Repository Service, including enabling, disabling,
starting, and stopping the service.
• Repository operations. Log events for repository operations such as creating and deleting repository
content, and adding deployed applications.
• User permissions. Log events about managing user permissions on the repository.
• Repository operations. Log events for accessing metadata in the Metadata Manager repository.
• Configuration. Log events about the configuration of the Metadata Manager Service.
• Run-time processes. Log events for running a Metadata Manager Service, such as missing native library
files.
• PowerCenter Integration Service log events. Session and workflow status for sessions and workflows that
use a PowerCenter Integration Service process to load data to the Metadata Manager warehouse or to
extract source metadata.
• PowerCenter Integration Service processes. Log events about the PowerCenter Integration Service
processes, including service ports, code page, operating mode, service name, and the associated
repository and PowerCenter Repository Service status.
• Licensing. Log events for license verification for the PowerCenter Integration Service by the Service
Manager.
• PowerCenter Repository connections. Log events for connections to the repository from PowerCenter
client applications, including user name and the host name and port number for the client application.
• PowerCenter Repository objects. Log events for repository objects locked, fetched, inserted, or updated by
the PowerCenter Repository Service.
• PowerCenter Repository Service processes. Log events about PowerCenter Repository Service processes,
including starting and stopping the PowerCenter Repository Service and information about repository
databases used by the PowerCenter Repository Service processes. Also includes repository operating
mode, the nodes where the PowerCenter Repository Service process runs, initialization information, and
internal functions used.
• Repository operations. Log events for repository operations, including creating, deleting, restoring, and
upgrading repository content, copying repository contents, and registering and unregistering local
repositories.
• Licensing. Log events about PowerCenter Repository Service license verification.
• Resource Manager Service. Log events about the Resource Manager Service, including enabling, disabling,
starting, and stopping the service.
• Compute nodes. Log events about nodes with the compute role registering with the Resource Manager
Service.
• SAP NetWeaver BI system log events. Requests from the SAP NetWeaver BI system to start a workflow
and status information from the ZPMSENDSTATUS ABAP program in the process chain.
• PowerCenter Integration Service log events. Session and workflow status for sessions and workflows that
use a PowerCenter Integration Service process to load data to or extract data from SAP NetWeaver BI.
To view log events about how the PowerCenter Integration Service processes an SAP NetWeaver BI workflow,
you must view the session or workflow log.
• Scheduler Service events. Log events about the Scheduler Service, including enabling, disabling, starting,
and stopping the service.
• Scheduled object events. Log events about starting scheduled object runs.
Web Services Hub log events contain the following log events:
• Web Services processes. Log events about web service processes, including starting and stopping Web
Services Hub, web services requests, the status of the requests, and error messages for web service calls.
Log events include information about which service workflows are fetched from the repository.
• PowerCenter Integration Service log events. Workflow and session status for service workflows including
invalid workflow errors.
Use the user activity log events to determine when a user created, updated, or removed services, nodes,
users, groups, or roles.
The Service Manager writes user activity log events when the Service Manager needs to authorize a user to
perform one of the following domain actions:
The Service Manager also writes user activity log events each time a user adds, updates, or removes a user,
group, operating system profile, or role.
The user activity log displays information about the user who performed the security action or who failed to
log in. When the Service Manager logs an unsuccessful login attempt, the logs display an error message that
contains the following information:
The Service Manager writes a user activity log event each time a user account is locked or unlocked. The
Service Manager also writes a user activity log event each time a user tries to log in to the domain with a
client application.
To include security audit trails in the user activity log events, you must enable the SecurityAuditTrail property
for the PowerCenter Repository Service in the Administrator tool.
When you import one or more repository objects, you can generate audit logs.
The audit logs contain the following information about the .xml file imported:
• Host name and IP address of the client machine from which the .xml file was imported
• Full local path of the .xml import file
• The file name
• The file size in bytes
• Logged in user name
• Number of objects imported
• Time stamp of the import operation
Log Aggregator
You can aggregate the log files of an application service that stops responding or shuts down unexpectedly.
You might need to analyze multiple log files to figure out issues with an application service.
You can use the log aggregator to aggregate all the log files associated with an application service and
compress required log files into a .zip file. You can download the .zip file and analyze the log files, or upload
the .zip file to Informatica Global Customer Support for analysis.
You cannot store the history of aggregated logs. You must download or send the file to Informatica Global
Customer Support after you aggregate the log files.
You can aggregate the hang and crash logs of the following application services:
• Analyst Service
• Data Integration Service
• Model Repository Service
• PowerCenter Integration Service
• PowerCenter Repository Service
In addition to the application service logs, the log aggregator captures debug information for the nodes in the
domain. The log aggregator aggregates the log files of the associated application services when you
The log collection directory in the master gateway node stores the application service logs when you
aggregate the logs. All the node processes in the domain must have read/write access on the log collection
directory. If the node processes cannot access the log collection directory, the aggregated logs do not
appear in the aggregated logs listgrid. The core dump directory stores the core dump files of the nodes in the
domain. Configure the log collection directory in the master gateway node and the core dump directory for
each node in the domain.
When you process the aggregated logs you can choose the collectors from which you want to collect log
information. The collectors are application services and nodes associated with the application service.
The Data Integration service writes a new log file for each Mapping task run. The log file contains information
about the events in the Mapping task. Log events are lines of text that contain a timestamp, thread identifier,
a severity code, and the log message. The message can contain general information or it can contain an error
message.
The following text shows the Mapping task log message format:
2022-03-03 12:28:05,409 IST <MappingCompiler-pool-4-thread-5> INFO: [MPSVCCMN_10081]
Mapping service is running [flatFile_allTransformations_mapping/
flatFile_allTransformations_mapping] deployed in [QueueApplication]
2022-03-03 12:28:05,409 IST <MappingCompiler-pool-4-thread-5> INFO: [MPSVCCMN_10083] The
Mapping Service Module submitted a job to the Integration Service. Job ID : [wyMK5pq-
EeyQm_HZtnihOw]
When you set the tracing level to verboseData, the Mapping task log shows the parameters and parameter
values for the mapping run.
The following text shows some Mapping task log messages that contain parameter values:
Integration Service will use override value [C:\Source] for parameter [ff_SrcDir] in
transformation [map_AllTx\read_src1].
Integration Service will use override value [8] for parameter [exp_Int] in
transformation [map_AllTx\Expression].
Integration Service will use override value [Mapping_New] for parameter [exp_String] in
transformation [map_AllTx\Expression].
Integration Service will use override value [C:\Source] for parameter [ldo_SrcDir] in
mapping \ mapplet [map_AllTx\DO_Lookup\DO_FF_REL_SRC_Read_Mapping].
After you run a mapping on the Spark engine on a Hadoop cluster, you can view the total number of cluster
nodes used to execute the mapping in the mapping task log. On the Blaze engine, you can view the number of
healthy cluster nodes used by the Grid Manager in the mapping task log.
Domain Reports
This chapter includes the following topics:
• License Management Report. Monitors the number of software options purchased for a license and the
number of times a license exceeds usage limits. The License Management Report displays the license
usage information such as CPU and repository usage and the node configuration details.
• Web Services Report. Monitors activities of the web services running on a Web Services Hub. The Web
Services Report displays run-time information such as the number of successful or failed requests and
average service time. You can also view historical statistics for a specific period of time.
Note: If the master gateway node runs on a UNIX machine and the UNIX machine does not have a graphics
display server, you must install X Virtual Frame Buffer on the UNIX machine to view the report charts in the
License Report or the Web Services Report. If you have multiple gateway nodes running on UNIX machines,
install X Virtual Frame Buffer on each UNIX machine.
You can save the License Management Report as a PDF on your local machine. You can also email a PDF
version of the report to someone.
Run the License Management Report to monitor the following license usage information:
• Licensing details. Shows general properties for every license assigned in the domain.
• CPU usage. Shows the number of logical CPUs used to run application services in the domain. The
License Management Report counts logical CPUs instead of physical CPUs for license enforcement. If the
284
number of logical CPUs exceeds the number of authorized CPUs, then the License Management Report
shows that the domain exceeded the CPU limit.
• Repository usage. Shows the number of PowerCenter Repository Services in the domain.
• User information. Shows information about users in the domain.
• Hardware configuration. Shows details about the machines used in the domain.
• Node configuration. Shows details about each node in the domain.
• Licensed options. Shows a list of PowerCenter and other Informatica options purchased for each license.
Licensing
The Licensing section of the License Management Report shows information about each license in the
domain.
The following table describes the licensing information in the License Management Report:
Property Description
Serial Number Serial number of the license. The serial number identifies the customer or project. If the
customer has multiple PowerCenter installations, there is a separate serial number for each
project. The original and incremental keys for a license have the same serial number.
Operating System / Operating system and bitmode for the license. Indicates whether the license is installed on a
BitMode 32-bit or 64-bit operating system.
AT Named Users Maximum number of users who are assigned the License Access for Informatica Analyst
privilege.
Product Bitmode Bitmode of the server binaries that are installed. Values are 32-bit or 64-bit.
CPU Summary
The CPU Summary section of the License Management Report shows the maximum number of logical CPUs
used to run application services in the domain. Use the CPU summary information to determine if the CPU
usage exceeded the license limits. If the number of logical CPUs is greater than the total number of CPUs
authorized by the license, the License Management Report indicates that the CPU limit is exceeded.
The License Management Report determines the number of logical CPUs based on the number of processors,
cores, and threads. Use the following formula to calculate the number of logical CPUs:
N*C*T, where
For example, a machine contains 4 processors. Each processor has 2 cores. The machine contains 8 (4*2)
physical cores. Hyperthreading is enabled, where each core contains 3 threads. The number of logical CPUs
is 24 (4*2*3).
Note: Although the License Management Report includes threads in the calculation of logical CPUs,
Informatica license compliance is based on the number of physical cores, not threads. To be compliant, the
number of physical cores must be less than or equal to the maximum number of licensed CPUs. If the
License Management Report shows that you have exceeded the license limit but the number of physical
cores is less than or equal to the maximum number of licensed CPUs, you can ignore the message. If you
have a concern about license compliance, contact your Informatica account manager.
The following table describes the CPU summary information in the License Management Report:
Property Description
Current Usage Maximum number of logical CPUs used concurrently on the day the report runs.
Peak Usage Maximum number of logical CPUs used concurrently during the last 12 months.
Peak Usage Date Date when the maximum number of logical CPUs were used concurrently during
the last 12 months.
Days Exceeded License Limit Number of days that the CPU usage exceeded the license limits. The domain
exceeds the CPU license limit when the number of concurrent logical CPUs
exceeds the number of authorized CPUs.
CPU Detail
The CPU Detail section of the License Management Report provides CPU usage information for each host in
the domain. The CPU Detail section shows the maximum number of logical CPUs used each day in a selected
time period.
The report counts the number of logical CPUs on each host that runs application services in the domain. The
report groups logical CPU totals by node.
The following table describes the CPU detail information in the License Management Report:
Property Description
Current Usage Maximum number of logical CPUs that the host used concurrently on the day the report runs.
Peak Usage Maximum number of logical CPUs that the host used concurrently during the last 12 months.
Peak Usage Date Date in the last 12 months when the host concurrently used the maximum number of logical
CPUs.
Assigned Licenses Name of all licenses assigned to services that run on the node.
The following table describes the repository summary information in the License Management Report:
Property Description
Current Usage Maximum number of repositories used concurrently in the domain on the day the
report runs.
Peak Usage Maximum number of repositories used concurrently in the domain during the last
12 months.
Peak Usage Date Date in the last 12 months when the maximum number of repositories were used
concurrently.
Days Exceeded License Limit Number of days that the repository usage exceeded the license limits.
User Summary
The User Summary section of the License Management Report provides information about Analyst tool users
in the domain.
The following table describes the user summary information in the License Management Report:
Property Description
Current Named Users Maximum number of users who are assigned the License Access for Informatica
Analyst privilege on the day the report runs.
Peaked Name Users Maximum number of users who are assigned the License Access for Informatica
Analyst privilege during the last 12 months.
Peak Named Users Date Date during the last 12 months when the maximum number of concurrent users
were assigned the License Access for Informatica Analyst privilege.
User Detail
The User Detail section of the License Management Report provides information about each Analyst tool user
in the domain.
The following table describes the user detail information in the License Management Report:
Property Description
Days Logged In Number of days the user logged in to the Analyst tool and performed profiling during the
last 12 months.
Peak Unique IP Maximum number of machines that the user was logged in to and performed profiling on
Addresses in a Day during a single day of the last 12 months.
Average Unique IP Daily average number of machines that the user was logged in to and running profiling on
Addresses during the last 12 months.
Peak IP Address Date Date when the user logged in to and performed profiling on the maximum number of
machines during a single day of the last 12 months.
Peak Daily Sessions Maximum number of times in a single day of the last 12 months that the user logged in to
any Analyst tool and performed profiling.
Average Daily Sessions Average number of times per day in the last 12 months that the user logged in to any
Analyst tool and performed profiling.
Peak Session Date Date in the last 12 months when the user had the most daily sessions in the Analyst tool.
Hardware Configuration
The Hardware Configuration section of the License Management Report provides details about machines
used in the domain.
The following table describes the hardware configuration information in the License Management Report:
Property Description
Logical CPUs Number of logical CPUs used to run application services in the domain.
Cores per socket Number of cores for each socket on the machine.
Node Configuration
The Node Configuration section of the License Management Report provides details about each node in the
domain.
Property Description
Node Name Name of the node or nodes assigned to a machine for a license.
Operating System Operating system of the machine on which the node runs.
Service Type Type of the application service configured to run on the node.
Service Name Name of the application service configured to run on the node.
Licensed Options
The Licensed Options section of the License Management Report provides details about each option for
every license assigned to the domain.
The following table describes the licensed option information in the License Management Report:
Property Description
The domain administrator can send the License Management Report in an email from Send License
Management Report page in the Administrator tool.
Property Description
To Email Email address to which you send the License Management Report.
Request ID Request ID that identifies the project for which the license was purchased.
Contact Email Email address of the contact person at the customer site.
2. Click OK.
The Administrator tool sends the License Management Report in an email.
The Web Services Report provides run-time and historical information on the web service requests handled
by the Web Services Hub. The report displays aggregated information for all web services in the Web
Services Hub and information for each web service running on the Web Services Hub. The Web Services
Report also provides historical information.
Time Interval
By default, the Web Services Report displays activity information for a five-minute interval. You can select
one of the following time intervals to display activity information for a web service or Web Services Hub:
• 5 seconds
• 1 minute
• 5 minutes
• 1 hour
• 24 hours
The Web Services Report displays activity information for the interval ending at the time you run the report.
For example, if you run the Web Services Report at 8:05 a.m. for an interval of one hour, the Web Services
Report displays the Web Services Hub activity from 7:05 a.m. and 8:05 a.m.
Caching
The Web Services Hub caches 24 hours of activity data. The cache is reinitialized every time the Web
Services Hub is restarted. The Web Services Report displays statistics from the cache for the time interval
that you run the report.
• Properties view. Displays General Properties, Web Services Hub Summary, and Historical Statistics for the
web services hub.
• Web Services view. Lists the web services in the web services hub. When you select a web service, you
can view Properties, Top IP Addresses, and Historical Statistics for the web service.
Property Description
Service type Type of Service. For a Web Services Hub, the service type is
ServiceWSHubService.
The following table describes the Web Services Hub Summary properties:
Property Description
# of Successful Message Number of requests that the Web Services Hub processed successfully.
# of Fault Responses Number of fault responses generated by web services in the Web Services Hub.
The fault responses could be due to any error.
Total Messages Total number of requests that the Web Services Hub received.
Last Server Restart Tme Date and time when the Web Services Hub was last started.
Avg. # of Service Partitions Average number of partitions allocated for all web services in the Web Services
Hub.
% of Partitions in Use Percentage of web service partitions that are in use for all web services in the
Web Services Hub.
Avg. # of Run Instances Average number of instances running for all web services in the Web Services
Hub.
Property Description
Web Service Name of the web service for which the information is displayed.
When you click the name of a web service, the Web Services Report displays the
Service Statistics window.
Avg. Service Time Average time it takes to process a service request received by the web service.
Max Service Time The largest amount of time taken by the web service to process a request.
Min Service Time The smallest amount of time taken by the web service to process a request.
Avg. DTM Time Average number of seconds it takes the PowerCenter Integration Service to
process the requests from the Web Services Hub.
Avg. Service Partitions Average number of session partitions allocated for the web service.
Avg Run Instances Average number of instances running for the web service.
The report provides the following information for each web service for the selected time interval:
Property Description
Service name Name of the web service for which the information is displayed.
Successful Requests Number of requests received by the web service that the Web Services Hub
processed successfully.
Fault Responses Number of fault responses generated by the web services in the Web Services
Hub.
Avg. Service Time Average time it takes to process a service request received by the web service.
Avg. Service Partitions Average number of session partitions allocated for the web service.
Avg. Run Instances Average number of instances of the web service running during the interval.
The report provides the following information for the selected web service:
Property Description
# of Successful Requests Number of requests received by the web service that the Web Services Hub
processed successfully.
# of Fault Responses Number of fault responses generated by the web services in the Web Services
Hub.
Total Messages Total number of requests that the Web Services Hub received.
Last Server Restart Time Date and time when the Web Services Hub was last started
Last Service Time Number of seconds it took to process the most recent service request
Average Service Time Average time it takes to process a service request received by the web service.
Avg.# of Service Partitions Average number of session partitions allocated for the web service.
Avg. # of Run Instances Average number of instances of the web service running during the interval.
The report provides the following information for each of the most active IP addresses:
Property Description
Top 10 Client IP Addresses The list of client IP addresses and the longest time taken by the web service to
process a request from the client. The client IP addresses are listed in the order of
longest to shortest service times. Use the Click here link to display the list of IP
addresses and service times.
Property Description
Web Service Name of the web service for which the information is displayed.
Fault Responses Number of requests received for the web service that could not be processed and
generated fault responses.
Avg. Service Time Average time it takes to process a service request received by the web service.
Min. Service Time The smallest amount of time taken by the web service to process a request.
Max. Service Time The largest amount of time taken by the web service to process a request.
Avg. DTM Time Average time it takes the PowerCenter Integration Service to process the requests from
the Web Services Hub.
Avg. Service Partitions Average number of session partitions allocated for the web service.
Avg. Run Instances Average number of instances running for the web service.
Before you run the Web Services Report for a Web Services Hub, verify that the Web Services Hub is enabled.
You cannot run the Web Services Report for a disabled Web Services Hub.
In a domain that contains multiple nodes, the node where you generate the SSL certificate affects how you
access the Web Services Report for a secure Web Services Hub.
Use the following rules and guidelines to run the Web Services Report for a secure Web Services Hub in a
domain with multiple nodes:
• For each secure Web Services Hub running in a domain, generate an SSL certificate and import it to a
Java certificate file.
• The Administrator tool searches for SSL certificates in the certificate file of a gateway node. The SSL
certificate for a Web Services Hub running on worker node must be generated on a gateway node and
imported into the certificate file of the same gateway node.
• To view the Web Services Report for a secure Web Services Hub, log in to the Administrator tool from the
gateway node that has the certificate file containing the SSL certificate of the Web Services Hub for which
you want to view reports.
• If a secure Web Services Hub runs on a worker node, the SSL certificate must be generated and imported
into the certificate file of the gateway node. If a secure Web Services Hub runs on a gateway and a worker
node, the SSL certificate of both nodes must be generated and imported into the certificate file of the
gateway node. To view reports for the secure Web Services Hub, log in to the Administrator tool from the
gateway node.
• If the domain has two gateway nodes and a secure Web Services Hub runs on each gateway node, access
to the Web Services Reports depends on where the SSL certificate is located.
For example, gateway node GWN01 runs Web Services Hub WSH01 and gateway node GWN02 runs Web
Services Hub WSH02. You can view the reports for the Web Services Hubs based on the location of the
SSL certificates:
- If the SSL certificate for WSH01 is in the certificate file of GWN01 but not GWN02, you can view the
reports for WSH01 if you log in to the Administrator tool through GWN01. You cannot view the reports
for WSH01 if you log in to the Administrator tool through GWN02. If GWN01 fails, you cannot view
reports for WSH01.
- If the SSL certificate for WSH01 is in the certificate files of GWN01 and GWN02, you can view the reports
for WSH01 if you log in to the Administrator tool through GWN01 or GWN02. If GWN01 fails, you can
view the reports for WSH01 if you log in to the Administrator tool through GWN02.
• To ensure successful failover when a gateway node fails, generate and import the SSL certificates of all
Web Services Hubs in the domain into the certificates files of all gateway nodes in the domain.
Node Diagnostics
This chapter includes the following topics:
You can discover comprehensive information about your technical environment and diagnose issues before
they become critical.
Generate node diagnostics from the Informatica Administrator and upload them to the Configuration Support
Manager in the Informatica MySupport Portal. Then, check the node diagnostics against business rules and
recommendations in the Configuration Support Manager.
The following image shows the operational flow to generate and upload node diagnostics:
297
2. Generate node diagnostics. The Service Manager analyzes the services of the node and generates node
diagnostics including information such as operating system details, CPU details, database details, and
patches.
3. Optionally, download node diagnostics to your local drive.
4. Upload node diagnostics to the Configuration Support Manager, a diagnostic web application outside the
firewall. The Configuration Support Manager is a part of the Informatica MySupport Portal. The Service
Manager connects to the Configuration Support Manager through the HTTPS protocol and uploads the
node diagnostics.
5. Review the node diagnostics in the Configuration Support Manager to find troubleshooting information
for your environment.
To maintain login security, you must log out of the Configuration Support Manager and the Node Diagnostics
Upload page of the Administrator tool.
• To log out of the Configuration Support Manager, click the logout link.
• To log out of the Upload page, click Close Window.
Note: If you close these windows through the web browser close button, you remain logged in to the
Configuration Support Manager. Other users can access the Configuration Support Manager without valid
credentials.
1. In the Administrator tool, click the Manage tab > Services and Nodes view.
2. In the Domain Navigator, select the domain.
3. In the contents panel, click Diagnostics.
A list of all the nodes in the domain appears.
Field Description
Email Address Email address with which you registered your customer portal account.
6. Click OK.
The XML file contains details about services, logs, environment variables, operating system parameters,
system information, and database clients. Worker node diagnostics only contain node metadata; they do not
include domain metadata.
1. In the Administrator tool, click the Manage tab > Services and Nodes view.
2. In the Domain Navigator, select the domain.
3. In the contents panel, click Diagnostics.
A list of all nodes in the domain appears.
4. Select the node.
5. Click Generate Diagnostics File.
6. Click Yes to confirm that you want to generate node diagnostics.
1. In the Administrator tool, click the Manage tab > Services and Nodes view.
2. In the Domain Navigator, select the domain.
3. In the contents panel, click Diagnostics.
A list of all nodes in the domain appears.
4. Click the diagnostics file name of the node.
The file opens in another browser window.
5. Click File > Save As. Then, specify a location to save the file.
6. Click Save.
The XML file is saved to your local drive.
When you upload node diagnostics, you can update or create a configuration in the Configuration Support
Manager. Create a configuration the first time you upload the node diagnostics. Update a configuration to
view the latest diagnostics of the configuration. To compare current and previous node configurations of an
existing configuration, upload the current node diagnostics as a new configuration.
Note: If you do not have access to the Internet, you can download the file and upload it at a later time. You
can also send the file to the Informatica Global Customer Support in an email to troubleshoot or to upload.
1. In the Administrator tool, click the Manage tab > Services and Nodes view.
2. In the Domain Navigator, select the domain.
3. In the contents panel, click Diagnostics.
A list of all nodes in the domain appears.
4. Select the node.
5. Generate node diagnostics.
6. Click Upload Diagnostics File to CSM.
Field Description
For example, when you run a Sorter session that processes a large volume of data, you notice that there is
some data loss. You generate node diagnostics and upload them to the Configuration Support Manager.
When you review the diagnostics for bug fix alerts, you see that a bug fix, EBF178626, is available for this.
You apply EBF178626, and run the session again. All data is successfully loaded.
Identify Recommendations
You can use the Configuration Support Manager to avoid issues in your environment. You can troubleshoot
issues that arise after you make changes to the node properties by comparing different node diagnostics in
the Configuration Support Manager. You can also use the Configuration Support Manager to identify
recommendations or updates that may help you improve the performance of the node.
For example, you upgrade the node memory to handle a higher volume of data. You generate node
diagnostics and upload them to the Configuration Support Manager. When you review the diagnostics for
operating system warnings, you find the recommendation to increase the total swap memory of the node to
twice that of the node memory for optimal performance. You increase swap space as suggested in the
Configuration Support Manager and avoid performance degradation.
Tip: Regularly upload node diagnostics to the Configuration Support Manager and review node diagnostics to
maintain your environment efficiently.
Understanding Globalization
This chapter includes the following topics:
Globalization Overview
Informatica can process data in different languages. Some languages require single-byte data, while other
languages require multibyte data. To process data correctly in Informatica, you must set up the following
items:
• Locale. Informatica requires that the locale settings on machines that access Informatica applications are
compatible with code pages in the domain. You may need to change the locale settings. The locale
specifies the language, territory, encoding of character set, and collation order.
• Data movement mode. The PowerCenter Integration Service can process single-byte or multibyte data and
write it to targets. Use the ASCII data movement mode to process single-byte data. Use the Unicode data
movement mode for multibyte data.
• Code pages. Code pages contain the encoding to specify characters in a set of one or more languages.
You select a code page based on the type of character data you want to process. To ensure accurate data
movement, you must ensure compatibility among code pages for Informatica and environment
components. You use code pages to distinguish between US-ASCII (7-bit ASCII), ISO 8859-1 (8-bit ASCII),
and multibyte characters.
To ensure data passes accurately through your environment, the following components must work together:
303
• Code page for each PowerCenter Integration Service process
• PowerCenter Client code page
• PowerCenter repository code page
• Source and target database code pages
• Metadata Manager repository code page
You can configure the PowerCenter Integration Service for relaxed code page validation. Relaxed validation
removes restrictions on source and target code pages.
Unicode
The Unicode Standard is the work of the Unicode Consortium, an international body that promotes the
interchange of data in all languages. The Unicode Standard is designed to support any language, no matter
how many bytes each character in that language may require. Currently, it supports all common languages
and provides limited support for other less common languages. The Unicode Consortium is continually
enhancing the Unicode Standard with new character encodings. For more information about the Unicode
Standard, see http://www.unicode.org.
The Unicode Standard includes multiple character sets. Informatica uses the following Unicode standards:
• UCS-2 (Universal Character Set, double-byte). A character set in which each character uses two bytes.
• UTF-8 (Unicode Transformation Format). An encoding format in which each character can use between
one to four bytes.
• UTF-16 (Unicode Transformation Format). An encoding format in which each character uses two or four
bytes.
• UTF-32 (Unicode Transformation Format). An encoding format in which each character uses four bytes.
• GB18030. A Unicode encoding format defined by the Chinese government in which each character can
use between one to four bytes.
Informatica is a Unicode application. The PowerCenter Client, PowerCenter Integration Service, and Data
Integration Service use UCS-2 internally. The PowerCenter Client converts user input from any language to
UCS-2 and converts it from UCS-2 before writing to the PowerCenter repository. The PowerCenter Integration
Service and Data Integration Service converts source data to UCS-2 before processing and converts it from
UCS-2 after processing. The PowerCenter repository, Model repository, PowerCenter Integration Service, and
Data Integration Service support UTF-8. You can use Informatica to process data in any language.
The domain configuration database uses the UTF-8 code page. If you need to store metadata in multiple
languages, such as Chinese, Japanese, and Arabic, you must use the UTF-8 code page for all services in that
domain.
The Service Manager synchronizes the list of users in the domain with the list of users and groups in each
application service. If a user in the domain has characters that the code page of the application services
does not recognize, characters do not convert correctly and inconsistencies occur.
Use the following guidelines when you use UTF-8 as the PowerCenter repository code page:
Locales
Every machine has a locale. A locale is a set of preferences related to the user environment, including the
input language, keyboard layout, how data is sorted, and the format for currency and dates. Informatica uses
locale settings on each machine.
• System locale. Determines the language, code pages, and associated bitmap font files that are used as
defaults for the system.
• User locale. Determines the default formats to display date, time, currency, and number formats.
• Input locale. Describes the input method, such as the keyboard, of the system language.
For more information about configuring the locale settings on Windows, consult the Windows
documentation.
System Locale
The system locale is also referred to as the system default locale. It determines which ANSI and OEM code
pages, as well as bitmap font files, are used as defaults for the system. The system locale contains the
language setting, which determines the language in which text appears in the user interface, including in
dialog boxes and error messages. A message catalog file defines the language in which messages display.
By default, the machine uses the language specified for the system locale for all processes, unless you
override the language for a specific process.
The system locale is already set on your system and you may not need to change settings to run Informatica.
If you do need to configure the system locale, you configure the locale on a Windows machine in the Regional
Options dialog box. On UNIX, you specify the locale in the LANG environment variable.
Locales 305
User Locale
The user locale displays date, time, currency, and number formats for each user. You can specify different
user locales on a single machine. Create a user locale if you are working with data on a machine that is in a
different language than the operating system. For example, you might be an English user working in Hong
Kong on a Chinese operating system. You can set English as the user locale to use English standards in your
work in Hong Kong. When you create a new user account, the machine uses a default user locale. You can
change this default setting once the account is created.
Input Locale
An input locale specifies the keyboard layout of a particular language. You can set an input locale on a
Windows machine to type characters of a specific language.
You can use the Windows Input Method Editor (IME) to enter multibyte characters from any language without
having to run the version of Windows specific for that language. For example, if you are working on an
English operating system and need to enter text in Chinese, you can use IME to set the input locale to
Chinese without having to install the Chinese version of Windows. You might want to use an input method
editor to enter multibyte characters into a PowerCenter repository that uses UTF-8.
• ASCII (American Standard Code for Information Interchange). The US-ASCII code page contains a set of 7-
bit ASCII characters and is a subset of other character sets. When the PowerCenter Integration Service
runs in ASCII data movement mode, each character requires one byte.
• Unicode. The universal character-encoding standard that supports all languages. When the PowerCenter
Integration Service runs in Unicode data movement mode, it allots up to two bytes for each character. Run
the PowerCenter Integration Service in Unicode mode when the source contains multibyte data.
Tip: You can also use ASCII or Unicode data movement mode if the source has 8-bit ASCII data. The
PowerCenter Integration Service allots an extra byte when processing data in Unicode data movement
mode. To increase performance, use the ASCII data movement mode. For example, if the source contains
characters from the ISO 8859-1 code page, use the ASCII data movement.
The data movement you choose affects the requirements for code pages. Ensure the code pages are
compatible.
When you run the PowerCenter Integration Service in Unicode mode, it enforces session code page
relationships.
The following table describes how the PowerCenter Integration Service handles session files and caches
after you change the data movement mode:
Session File or Time of Creation or Use PowerCenter Integration Service Behavior After Data
Cache Movement Mode Change
Session Log File Each session. No change in behavior. Creates a new session log for each
(*.log) session using the code page of the PowerCenter Integration
Service process.
Workflow Log Each workflow. No change in behavior. Creates a new workflow log file for
each workflow using the code page of the PowerCenter
Integration Service process.
Reject File (*.bad) Each session. No change in behavior. Appends rejected data to the
existing reject file using the code page of the PowerCenter
Integration Service process.
Output File (*.out) Sessions writing to flat file. No change in behavior. Creates a new output file for each
session using the target code page.
Indicator File (*.in) Sessions writing to flat file. No change in behavior. Creates a new indicator file for each
session.
Incremental Sessions with Incremental When files are removed or deleted, the PowerCenter
Aggregation Files Aggregation enabled. Integration Service creates new files.
(*.idx, *.dat) When files are not moved or deleted, the PowerCenter
Integration Service fails the session with the following error
message:
SM_7038 Aggregate Error: ServerMode: [server
data movement mode] and CachedMode: [data
movement mode that created the files]
mismatch.
Move or delete files created using a different code page.
Unnamed Persistent Sessions with a Lookup Rebuilds the persistent lookup cache.
Lookup Files (*.idx, transformation configured for
*.dat) an unnamed persistent
lookup cache.
Named Persistent Sessions with a Lookup When files are removed or deleted, the PowerCenter
Lookup Files (*.idx, transformation configured for Integration Service creates new files.
*.dat) a named persistent lookup When files are not moved or deleted, the PowerCenter
cache. Integration Service fails the session.
Move or delete files created using a different code page.
When you choose a code page, the program or application for which you set the code page refers to a
specific set of data that describes the characters the application recognizes. This influences the way that
application stores, receives, and sends character data.
MS Latin1 and Latin1 both support English and most Western European languages and are compatible with
each other. When you install the PowerCenter Client, PowerCenter Integration Service, or PowerCenter
You can use the IBM EBCDIC code page for the PowerCenter Integration Service process when you install it
on a mainframe system. You cannot install the PowerCenter Client or PowerCenter repository on mainframe
systems, so you cannot use the IBM EBCDIC code page for PowerCenter Client or PowerCenter repository
installations.
UNIX systems allow you to change the code page by changing the LANG, LC_CTYPE or LC_ALL environment
variable. For example, you want to change the code page an AIX machine uses. Use the following command
in the C shell to view your environment:
locale
This results in the following output, in which “C” implies “ASCII”:
LANG="C"
LC_CTYPE="C"
LC_NUMERIC="C"
LC_TIME="C"
LC_ALL="C"
To change the language to English and require the system to use the Latin1 code page, you can use the
following command:
setenv LANG en_US.iso88591
When you check the locale again, it has been changed to use Latin1 (ISO 8859-1):
LANG="en_US.iso88591"
LC_CTYPE="en_US.iso88591"
LC_NUMERIC="en_US.iso88591"
LC_TIME="en_US.iso88591"
LC_ALL="en_US.iso88591"
For more information about changing the locale or code page of a UNIX system, see the UNIX
documentation.
If you purchase Windows in the United States and use English as an input and display language, your
operating system code page is MS Latin1 (MS1252) by default. However, if you install additional display or
input languages from the Windows installation CD and use those languages, the operating system might use
a different code page.
For more information about the default code page for your Windows system, contact Microsoft.
• Single-byte. A character represented as a unique number between 0 and 255. One byte is eight bits. ASCII
characters are single-byte characters.
• Double-byte. A character two bytes or 16 bits in size represented as a unique number 256 or greater. Many
Asian languages, such as Chinese, have double-byte characters.
• Multibyte. A character two or more bytes in size is represented as a unique number 256 or greater. Many
Asian languages, such as Chinese, have multibyte characters.
A code page can be compatible with another code page, or it can be a subset or a superset of another:
• Compatible. Two code pages are compatible when the characters encoded in the two code pages are
virtually identical. For example, JapanEUC and JIPSE code pages contain identical characters and are
compatible with each other. The PowerCenter repository and PowerCenter Integration Service process
can each use one of these code pages and can pass data back and forth without data loss.
• Superset. A code page is a superset of another code page when it contains all the characters encoded in
the other code page and additional characters not encoded in the other code page. For example, MS
Latin1 is a superset of US-ASCII because it contains all characters in the US-ASCII code page.
Note: Informatica considers a code page to be a superset of itself and all other compatible code pages.
• Subset. A code page is a subset of another code page when all characters in the code page are also
encoded in the other code page. For example, US-ASCII is a subset of MS Latin1 because all characters in
the US-ASCII code page are also encoded in the MS Latin1 code page.
For accurate data movement, the target code page must be a superset of the source code page. If the target
code page is not a superset of the source code page, the PowerCenter Integration Service may not process
all characters, resulting in incorrect or missing data. For example, Latin1 is a superset of US-ASCII. If you
select Latin1 as the source code page and US-ASCII as the target code page, you might lose character data if
the source contains characters that are not included in US-ASCII.
When you install or upgrade a PowerCenter Integration Service to run in Unicode mode, you must ensure code
page compatibility among the domain configuration database, the Administrator tool, PowerCenter Clients,
PowerCenter Integration Service process nodes, the PowerCenter repository, the Metadata Manager
repository, and the machines hosting pmrep and pmcmd. In Unicode mode, the PowerCenter Integration
Service enforces code page compatibility between the PowerCenter Client and the PowerCenter repository,
and between the PowerCenter Integration Service process and the PowerCenter repository. In addition, when
you run the PowerCenter Integration Service in Unicode mode, code pages associated with sessions must
have the appropriate relationships:
• For each source in the session, the source code page must be a subset of the target code page. The
PowerCenter Integration Service does not require code page compatibility between the source and the
PowerCenter Integration Service process or between the PowerCenter Integration Service process and the
target.
• Domain configuration database. The domain configuration database must be compatible with the code
pages of the PowerCenter repository and Metadata Manager repository.
• Administrator tool. You can enter data in any language in the Administrator tool.
• PowerCenter Client. You can enter metadata in any language in the PowerCenter Client.
• PowerCenter Integration Service process. The PowerCenter Integration Service can move data in ASCII
mode and Unicode mode. The default data movement mode is ASCII, which passes 7-bit ASCII or 8-bit
ASCII character data. To pass multibyte character data from sources to targets, use the Unicode data
movement mode. When you run the PowerCenter Integration Service in Unicode mode, it uses up to three
bytes for each character to move data and performs additional checks at the session level to ensure data
integrity.
• PowerCenter repository. The PowerCenter repository can store data in any language. You can use the
UTF-8 code page for the PowerCenter repository to store multibyte data in the PowerCenter repository.
The code page for the PowerCenter repository is the same as the database code page.
• Metadata Manager repository. The Metadata Manager repository can store data in any language. You can
use the UTF-8 code page for the Metadata Manager repository to store multibyte data in the repository.
The code page for the repository is the same as the database code page.
• Sources and targets. The sources and targets store data in one or more languages. You use code pages
to specify the type of characters in the sources and targets.
• PowerCenter command line programs. You must also ensure that the code page for pmrep is a subset of
the PowerCenter repository code page and the code page for pmcmd is a subset of the PowerCenter
Integration Service process code page.
Most database servers use two code pages, a client code page to receive data from client applications and a
server code page to store the data. When the database server is running, it converts data between the two
code pages if they are different. In this type of database configuration, the PowerCenter Integration Service
process interacts with the database client code page. Thus, code pages used by the PowerCenter Integration
Service process, such as the PowerCenter repository, source, or target code pages, must be identical to the
database client code page. The database client code page is usually identical to the operating system code
page on which the PowerCenter Integration Service process runs. The database client code page is a subset
of the database server code page.
For more information about specific database client and server code pages, see your database
documentation.
The Service Manager synchronizes the list of users in the domain with the list of users and groups in each
application service. If a user name in the domain has characters that the code page of the application service
does not recognize, characters do not convert correctly and inconsistencies occur.
However, on UNIX, you can change the code page of the PowerCenter Integration Service process by
changing the LANG, LC_CTYPE or LC_ALL environment variable for the user that starts the process.
The code page of the PowerCenter Integration Service process must be:
PowerCenter Integration Services configured for Unicode mode validate code pages when you start a session
to ensure accurate data movement. It uses session code pages to convert character data. When the
PowerCenter Integration Service runs in ASCII mode, it does not validate session code pages. It reads all
character data as ASCII characters and does not perform code page conversions.
Each code page has associated sort orders. When you configure a session, you can select one of the sort
orders associated with the code page of the PowerCenter Integration Service process. When you run the
PowerCenter Integration Service in Unicode mode, it uses the selected session sort order to sort character
data. When you run the PowerCenter Integration Service in ASCII mode, it sorts all character data using a
binary sort order.
If you run the PowerCenter Integration Service in the United States on Windows, consider using MS Windows
Latin1 (ANSI) as the code page of the PowerCenter Integration Service process.
If you run the PowerCenter Integration Service in the United States on UNIX, consider using ISO 8859-1 as the
code page for the PowerCenter Integration Service process.
If you use pmcmd to communicate with the PowerCenter Integration Service, the code page of the operating
system hosting pmcmd must be identical to the code page of the PowerCenter Integration Service process.
The PowerCenter Integration Service generates the names of session log files, reject files, caches and cache
files, and performance detail files based on the code page of the PowerCenter Integration Service process.
If you copy objects from one PowerCenter repository to another PowerCenter repository, the code page for
the target PowerCenter repository must be a superset of the code page for the source PowerCenter
repository.
• Flat files and VSAM files. The code page of the data in the file. When you configure the flat file or COBOL
source definition, choose a code page that matches the code page of the data in the file.
• XML files. The PowerCenter Integration Service converts XML to Unicode when it parses an XML source.
When you create an XML source definition, the PowerCenter Designer assigns a default code page. You
cannot change the code page.
• Relational databases. The code page of the database client. When you configure the relational connection
in the PowerCenter Workflow Manager, choose a code page that is compatible with the code page of the
Note: Select IBM EBCDIC as the source database connection code page only if you access EBCDIC data, such
as data from a mainframe extract file.
• Flat files. When you configure the flat file target definition, choose a code page that matches the code
page of the data in the flat file.
• XML files. Configure the XML target code page after you create the XML target definition. The XML Wizard
assigns a default code page to the XML target. The PowerCenter Designer does not apply the code page
that appears in the XML schema.
• Relational databases. When you configure the relational connection in the PowerCenter Workflow
Manager, choose a code page that is compatible with the code page of the database client. If you set a
database environment variable to specify the language for the database, ensure the code page for the
connection is compatible with the language set for the variable. For example, if you set the NLS_LANG
environment variable for an Oracle database, ensure that the code page of the Oracle connection is
compatible with the value set in the NLS_LANG variable. If you do not use compatible code pages,
sessions may hang or you might receive a database error, such as:
ORA-00911: Invalid character specified.
The target code page must be a superset of the code page of transformations and sources that provide data
to the target. The target code page does not need to be a superset of transformations or sources that do not
provide data to the target.
The PowerCenter Integration Service creates session indicator files, session output files, and external loader
control and data files using the target flat file code page.
Note: Select IBM EBCDIC as the target database connection code page only if you access EBCDIC data, such
as data from a mainframe extract file.
The code page compatibility for pmcmd and pmrep depends on whether you configured the code page
environment variable INFA_CODEPAGENAME for pmcmd or pmrep. You can set this variable for either
command line program or for both.
If you did not set this variable for a command line program, ensure the following requirements are met:
• If you did not set the variable for pmcmd, then the code page of the machine hosting pmcmd must be a
subset of the code page for the PowerCenter Integration Service process.
• If you set INFA_CODEPAGENAME for pmcmd, the code page defined for the variable must be a subset of
the code page for the PowerCenter Integration Service process.
• If you set INFA_CODEPAGENAME for pmrep, the code page defined for the variable must be a subset of
the PowerCenter repository code page.
• If you run pmcmd and pmrep from the same machine and you set the INFA_CODEPAGENAME variable, the
code page defined for the variable must be subsets of the code pages for the PowerCenter Integration
Service process and the PowerCenter repository.
If the code pages are not compatible, the PowerCenter Integration Service process may not fetch the
workflow, session, or task from the PowerCenter repository.
To ensure compatibility, the PowerCenter Client and PowerCenter Integration Service perform the following
code page validations:
• PowerCenter restricts the use of EBCDIC-based code pages for repositories. Since you cannot install the
PowerCenter Client or PowerCenter repository on mainframe systems, you cannot select EBCDIC-based
code pages, like IBM EBCDIC, as the PowerCenter repository code page.
• PowerCenter Client can connect to the PowerCenter repository when its code page is a subset of the
PowerCenter repository code page. If the PowerCenter Client code page is not a subset of the
PowerCenter repository code page, the PowerCenter Client fails to connect to the PowerCenter repository
code page with the following error:
REP_61082 <PowerCenter Client>'s code page <PowerCenter Client code page> is not one-
way compatible to repository <PowerCenter repository name>'s code page <PowerCenter
repository code page>.
• After you set the PowerCenter repository code page, you cannot change it. After you create or upgrade a
PowerCenter repository, you cannot change the PowerCenter repository code page. This prevents data
loss and inconsistencies in the PowerCenter repository.
• The PowerCenter Integration Service process can start if its code page is a subset of the PowerCenter
repository code page. The code page of the PowerCenter Integration Service process must be a subset of
the PowerCenter repository code page to prevent data loss or inconsistencies. If it is not a subset of the
PowerCenter repository code page, the PowerCenter Integration Service writes the following message to
the log files:
REP_61082 <PowerCenter Integration Service>'s code page <PowerCenter Integration
Service code page> is not one-way compatible to repository <PowerCenter repository
name>'s code page <PowerCenter repository code page>.
• When in Unicode data movement mode, the PowerCenter Integration Service starts workflows with the
appropriate source and target code page relationships for each session. When the PowerCenter
Integration Service runs in Unicode mode, the code page for every source in a session must be a subset of
the target code page. This prevents data loss during a session.
Although relaxed code page validation removes source and target code page restrictions, it still enforces
code page compatibility between the PowerCenter Integration Service and PowerCenter repository.
Note: Relaxed code page validation does not safeguard against possible data inconsistencies when you
move data between incompatible code pages. You must verify that the characters the PowerCenter
Integration Service reads from the source are included in the target code page.
Informatica removes the following restrictions when you relax code page validation:
• Source and target code pages. You can use any code page supported by Informatica for your source and
target data.
• Session sort order. You can use any sort order supported by Informatica when you configure a session.
When you run a session with relaxed code page validation, the PowerCenter Integration Service writes the
following message to the session log:
TM_6185 WARNING! Data code page validation is disabled in this session.
When you relax code page validation, the PowerCenter Integration Service writes descriptions of the
database connection code pages to the session log.
The following text shows sample code page messages in the session log:
TM_6187 Repository code page: [MS Windows Latin 1 (ANSI), superset of Latin 1]
WRT_8222 Target file [$PMTargetFileDir\passthru.out] code page: [MS Windows Traditional
Chinese, superset of Big 5]
WRT_8221 Target database connection [Japanese Oracle] code page: [MS Windows Japanese,
superset of Shift-JIS]
TM_6189 Source database connection [Japanese Oracle] code page: [MS Windows Japanese,
superset of Shift-JIS]
CMN_1716 Lookup [LKP_sjis_lookup] uses database connection [Japanese Oracle] in code
page [MS Windows Japanese, superset of Shift-JIS]
CMN_1717 Stored procedure [J_SP_INCREMENT] uses database connection [Japanese Oracle] in
code page [MS Windows Japanese, superset of Shift-JIS]
• Disable code page validation. Disable the ValidateDataCodePages option in the PowerCenter Integration
Service properties.
• Configure the PowerCenter Integration Service for Unicode data movement mode. Select Unicode for the
Data Movement Mode option in the PowerCenter Integration Service properties.
• Configure the PowerCenter Integration Service to write to the logs using the UTF-8 character set. If you
configure sessions or workflows to write to log files, enable the LogsInUTF8 option in the PowerCenter
Integration Service properties. The PowerCenter Integration Service writes all logs in UTF-8 when you
enable the LogsInUTF8 option. The PowerCenter Integration Service writes to the Log Manager in UTF-8
by default.
When the PowerCenter Integration Service reads characters that are not included in the target code page, you
risk transformation errors, inconsistent data, or failed sessions.
Note: If you relax code page validation, it is your responsibility to ensure that data converts from the source
to target properly.
The PowerCenter Integration Service failed a session and wrote the following message to the
session log:
TM_6188 The specified sort order is incompatible with the PowerCenter Integration
Service code page.
If you want to validate code pages, select a sort order compatible with the PowerCenter Integration Service
code page. If you want to relax code page validation, configure the PowerCenter Integration Service to relax
code page validation in Unicode data movement mode.
I tried to view the session or workflow log, but it contains garbage characters.
The PowerCenter Integration Service is not configured to write session or workflow logs using the UTF-8
character set.
When you run a session, the PowerCenter Integration Service converts source, target, and lookup queries
from the PowerCenter repository code page to the source, target, or lookup code page. The PowerCenter
Integration Service also converts the name and call text of stored procedures from the PowerCenter
repository code page to the stored procedure database code page.
At run time, the PowerCenter Integration Service verifies that it can convert the following queries and
procedure text from the PowerCenter repository code page without data loss:
Example
The PowerCenter Integration Service, PowerCenter repository, and PowerCenter Client use the ISO 8859-1
Latin1 code page, and the source database contains Japanese data encoded using the Shift-JIS code page.
Each code page contains characters not encoded in the other. Using characters other than 7-bit ASCII for the
PowerCenter repository and source database metadata can cause the sessions to fail or load no rows to the
target in the following situations:
• You create a mapping that contains a string literal with characters specific to the German language range
of ISO 8859-1 in a query. The source database may reject the query or return inconsistent results.
• You use the PowerCenter Client to generate SQL queries containing characters specific to the German
language range of ISO 8859-1. The source database cannot convert the German-specific characters from
the ISO 8859-1 code page into the Shift-JIS code page.
• The source database has a table name that contains Japanese characters. The PowerCenter Designer
cannot convert the Japanese characters from the source database code page to the PowerCenter Client
code page. Instead, the PowerCenter Designer imports the Japanese characters as question marks (?),
changing the name of the table. The PowerCenter Repository Service saves the source table name in the
PowerCenter repository as question marks. If the PowerCenter Integration Service sends a query to the
source database using the changed table name, the source database cannot find the correct table, and
returns no rows or an error to the PowerCenter Integration Service, causing the session to fail.
Because the US-ASCII code page is a subset of both the ISO 8859-1 and Shift-JIS code pages, you can avoid
these data inconsistencies if you use 7-bit ASCII characters for all of your metadata.
For this case study, the ISO 8859-1 environment consists of the following elements:
1. Verify code page compatibility between the PowerCenter repository database client and the database
server.
2. Verify code page compatibility between the PowerCenter Client and the PowerCenter repository, and
between the PowerCenter Integration Service process and the PowerCenter repository.
3. Set the PowerCenter Integration Service data movement mode to ASCII.
4. Verify session code page compatibility.
5. Verify lookup and stored procedure database code page compatibility.
6. Verify External Procedure or Custom transformation procedure code page compatibility.
7. Configure session sort order.
The PowerCenter repository resides in an Oracle database. Use NLS_LANG to set the locale (language,
territory, and character set) you want the database client and server to use with your login:
NLS_LANG = LANGUAGE_TERRITORY.CHARACTERSET
By default, Oracle configures NLS_LANG for the U.S. English language, the U.S. territory, and the 7-bit ASCII
character set:
NLS_LANG = AMERICAN_AMERICA.US7ASCII
Change the default configuration to write ISO 8859-1 data to the PowerCenter repository using the Oracle
WE8ISO8859P1 code page. For example:
NLS_LANG = AMERICAN_AMERICA.WE8ISO8859P1
In this case, the PowerCenter Client on Windows systems were purchased in the United States. Thus the
system code pages for the PowerCenter Client machines are set to MS Windows Latin1 by default. To verify
system input and display languages, open the Regional Options dialog box from the Windows Control Panel.
For systems purchased in the United States, the Regional Settings and Input Locale must be configured for
English (United States).
The PowerCenter Integration Service is installed on a UNIX machine. The default code page for UNIX
operating systems is ASCII. In this environment, change the UNIX system code page to ISO 8859-1 Western
European so that it is a subset of the PowerCenter repository code page.
In this case, the environment contains source databases containing German and English data. When you
configure a source database connection in the PowerCenter Workflow Manager, the code page for the
connection must be identical to the source database code page and must be a subset of the target code
page. Since both the MS Windows Latin1 and the ISO 8859-1 Western European code pages contain German
characters, you would most likely use one of these code pages for source database connections.
Because the target code page must be a superset of the source code page, use either MS Windows Latin1,
ISO 8859-1 Western European, or UTF-8 for target database connection or flat file code pages. To ensure
data consistency, the configured target code page must match the target database or flat file system code
page.
If you configure the PowerCenter Integration Service for relaxed code page validation, the PowerCenter
Integration Service removes restrictions on source and target code page compatibility. You can select any
supported code page for source and target data. However, you must ensure that the targets only receive
character data encoded in the target code page.
For this case study, the UTF-8 environment consists of the following elements:
1. Verify code page compatibility between the PowerCenter repository database client and the database
server.
2. Verify code page compatibility between the PowerCenter Client and the PowerCenter repository, and
between the PowerCenter Integration Service and the PowerCenter repository.
3. Configure the PowerCenter Integration Service for Unicode data movement mode.
4. Verify session code page compatibility.
5. Verify lookup and stored procedure database code page compatibility.
6. Verify External Procedure or Custom transformation procedure code page compatibility.
7. Configure session sort order.
The PowerCenter repository resides in an Oracle database. With Oracle, you can use NLS_LANG to set the
locale (language, territory, and character set) you want the database client and server to use with your login:
NLS_LANG = LANGUAGE_TERRITORY.CHARACTERSET
By default, Oracle configures NLS_LANG for U.S. English language, the U.S. territory, and the 7-bit ASCII
character set:
NLS_LANG = AMERICAN_AMERICA.US7ASCII
Change the default configuration to write UTF-8 data to the PowerCenter repository using the Oracle UTF8
character set. For example:
NLS_LANG = AMERICAN_AMERICA.UTF8
For more information about verifying and changing the PowerCenter repository database code page, see your
database documentation.
In this case, the PowerCenter Client on Windows systems were purchased in Switzerland. Thus, the system
code pages for the PowerCenter Client machines are set to MS Windows Latin1 by default. To verify system
input and display languages, open the Regional Options dialog box from the Windows Control Panel.
The PowerCenter Integration Service is installed on a UNIX machine. The default code page for UNIX
operating systems is ASCII. In this environment, the UNIX system character set must be changed to UTF-8.
In this case, the environment contains a source database containing German and Japanese data. When you
configure a source database connection in the PowerCenter Workflow Manager, the code page for the
connection must be identical to the source database code page. You can use any code page for the source
database.
Because the target code page must be a superset of the source code pages, you must use UTF-8 for the
target database connections or flat files. To ensure data consistency, the configured target code page must
match the target database or flat file system code page.
In this case, the External Procedure or Custom transformations must be able to process the German and
Japanese data from the sources. However, the PowerCenter Integration Service passes data to procedures in
UCS-2. Therefore, all data processed by the External Procedure or Custom transformations must be in the
UCS-2 character set.
To sort German and Japanese data when the PowerCenter Integration Service uses UTF-8, you most likely
want to use the default binary sort order.
A distribution package is a set of distribution binaries that you install within the domain for the following
processing requirements:
You can install distribution packages if you didn't do so during the upgrade or install process or if you want to
add a distribution package. You can remove a distribution package if you want to use a different package or
if you installed a package that you don't use.
When you install or remove distribution packages, verify that you perform the operation on all service and
client machines.
326
2. Set one of the following environment variables:
Variable Description
INFA_JDK_HOME Location of the folder containing the supported Java Development Kit (JDK). Set the
INFA_JDK_HOME environment variable in the following scenarios:
- Informatica domain is on Windows or Linux
- Informatica client
INFA_JRE_HOME Location of the folder containing the supported Java Runtime Environment (JRE). If the
Informatica domain is on AIX, set the INFA_JRE_HOME environment variable.
3. Verify that the user that runs the package manager has read and write permissions on the Informatica
installation directory and execute permissions on the executable file.
4. Download the following files from the Informatica Electronic Software Download site:
• Integration Package Manager
• Distribution packages
5. Extract the Integration Package Manager ZIP files to a local drive.
6. Copy the ZIP files of distribution packages that you need to the following location: <Integration
Package Manager directory>/source
Note: The package manager fails if the ZIP files for distribution packages aren't available in the source
directory.
1. From the package manager directory, run one of the following commands:
• ./Server.sh console for Linux or UNIX
• Server.bat console for Windows
• Client.bat console for client
Note: To run the command on Windows, use the administrator command prompt.
2. Enter the installation directory of the services or client and press Enter.
3. Choose the operation type and press Enter.
• Select 1 to remove existing distribution packages.
• Select 2 to install one or more distribution packages.
The console lists the distribution packages that you can install or remove.
4. Enter the distribution packages that you want to install or remove, separating multiple packages with a
comma, and press Enter.
5. Verify the installation or removal status in the package manager log file.
You can find the log file in the following location: <Integration Package Manager directory>/
IntegrationPackageManager_<date and timestamp>.log
1. Find the developerCore.ini file in the following location: <Informatica installation directory>
\clients\DeveloperClient
Code Pages
This appendix includes the following topics:
When you assign an application service code page in the Administrator tool, you select the code page
description.
You must use UTF-8 compatible code pages for the domain, Model Repository Service, and for each Data
Integration Service process.
The following table lists the name, description, and ID for supported code pages for the PowerCenter
Repository Service, the Metadata Manager Service, and for each PowerCenter Integration Service process:
Name Description ID
330
Name Description ID
When you assign a source or target code page in the PowerCenter Client, you select the code page
description. When you assign a code page using the pmrep CreateConnection command or define a code
page in a parameter file, you enter the code page name. The following table lists the name, description, and
ID for supported code pages for sources and targets:
Name Description ID
IBM-12712 EBCDIC Hebrew (updated with euro and new sheqel, control 10161
characters)
IBM-1363 PC Korean KSC MBCS Extended (with \ <-> Won mapping) 10032
IBM-1371 EBCDIC Taiwan Extended (SBCS IBM-1159 combined with DBCS 10154
IBM-9027)
IBM424 EBCDIC Hebrew (updated with new sheqel, control characters) 10120
• Select IBM EBCDIC as your source database connection code page only if you access EBCDIC data, such
as data from a mainframe extract file.
• The following code pages are not supported for database or relational connections:
- UTF-16 encoding of Unicode (Opposite Platform Endian)
Custom Roles
This appendix includes the following topics:
The following table lists the default privilege assigned to the Analyst Service Business Glossary Consumer
custom role:
342
Metadata Manager Service Custom Roles
Metadata Manager Service custom roles include the Metadata Manager Advanced User, Metadata Manager
Basic User, and Metadata Manager Intermediate User roles.
The following table lists the default privileges assigned to the Operator custom role:
PowerCenter Developer
The following table lists the default privileges assigned to the PowerCenter Developer custom role:
Folders - Copy
- Create
- Manage Versions
Test Engineer
The following table lists the default privileges assigned to the Test Engineer custom role:
Application services and the Service Managers in a domain use TCP/IP network protocol to
communicate with other nodes and services. The clients also use TCP/IP to communicate with
application services. You can configure the host name and port number for TCP/IP communication on a
node when you install the Informatica services. You can configure the port numbers used for services on
a node during installation or in Informatica Administrator.
Native drivers
The Data Integration Service uses native drivers to communicate with databases. The PowerCenter
Integration Service and the PowerCenter Repository Service use native drivers to communicate with
databases. Native drivers are packaged with the database server and client software. Install and
configure the native database client software on the machines where the services run.
ODBC
The ODBC drivers are installed with the Informatica services and the Informatica clients. The integration
services use ODBC drivers to communicate with databases.
JDBC
The Model Repository Service uses JDBC to connect to the Model repository database. The Metadata
Manager Service uses JDBC to connect to the Metadata Manager repository and metadata source
repositories.
The gateway nodes in the Informatica domain use JDBC to connect to the domain configuration
repository.
350
Domain Connectivity
Services on a node in an Informatica domain use TCP/IP to connect to services on other nodes. Because
services can run on multiple nodes in the domain, services rely on the Service Manager to route requests. The
Service Manager on the master gateway node handles requests for services and responds with the address
of the requested service.
Nodes communicate through TCP/IP on the port you select for a node when you install Informatica Services.
When you create a node, you select a port number for the node. The Service Manager listens for incoming
TCP/IP connections on that port.
The following figure shows an overview of the connectivity for components in the platform:
The platform uses connection objects to define connectivity information for source and target databases.
The connection objects can use native or ODBC connectivity. The Data Integration Service uses connection
objects to connect to sources and targets.
The Model Repository Service uses JDBC to read or write data and metadata in the Model repository. It
uses TCP/IP to communicate with the Data Integration Service and the clients.
The Data Integration Service uses ODBC or native drivers to connect and read data from a source
database and write data to a target database. It uses TCP/IP to communicate with the Model Repository
Service, Content Management Service, and client applications.
Informatica Developer
The Developer tool uses TCP/IP to send data transformation requests to the Data Integration Service. It
uses TCP/IP to communicate with the Content Management Service to manage reference tables,
probabilistic model files, and to retrieve configuration and status information for identity population files
and address validation reference data files. When you preview mappings or data objects in the Developer
tool, it uses JDBC or ODBC drivers to connect to the source or target database to fetch the metadata
required for preview.
Informatica Analyst
The Analyst Service uses TCP/IP to send requests to the Data Integration Service. It uses TCP/IP to
communicate with the Content Management Service to manage reference tables. When an Analyst tool
user previews profiles or objects, the Analyst Service fetches the metadata that the preview requires
If you use ODBC to connect to the source or target database, install the ODBC driver on the node where
the Analyst Service runs.
The Analyst Service can also connect to an exception management audit database. The exception
management audit database is a centralized audit trail for the work that Analyst tool users perform on
Human task instances. The Analyst Service uses JDBC drivers to connect to the exception management
audit database.
The Content Management Service manages the locations and other properties for reference data. The
Content Management Service uses TCP/IP to communicate with the Data Integration Service to read and
write data in reference tables. The Content Management Service uses JDBC to communicate directly
with the reference data warehouse when it creates reference tables.
If multiple instances of a Content Management Service exist in an Informatica domain, the master
Content Management Service updates the Data Integration Service. The master Content Management
Service uses TCP/IP to communicate with the Domain Service to identify the Model Repository Service
and the Data Integration Service to use.
1. A Model repository client sends a repository connection request to the master gateway node, which is the entry
point to the domain.
2. The Service Manager sends back the host name and port number of the node running the Model Repository Service.
In the diagram, the Model Repository Service is running on node A.
3. The repository client establishes a TCP/IP connection with the Model Repository Service process on node A.
4. The Model Repository Service process communicates with the Model repository database over JDBC. The Model
Repository Service process stores objects in or retrieves objects from the Model repository database based on
requests from the Model repository client.
Note: The Model repository tables have an open architecture. Although you can view the repository tables,
never manually edit them through other utilities. Informatica is not responsible for corrupted data that is
caused by customer alteration of the repository tables or data within those tables.
PowerCenter Connectivity
PowerCenter uses the TCP/IP network protocol, native database drivers, ODBC, and JDBC for communication
between the following PowerCenter components:
• PowerCenter Repository Service. The PowerCenter Repository Service uses native database drivers to
communicate with the PowerCenter repository. The PowerCenter Repository Service uses TCP/IP to
communicate with other PowerCenter components.
• PowerCenter Integration Service. The PowerCenter Integration Service uses native database connectivity
and ODBC to connect to source and target databases. The PowerCenter Integration Service uses TCP/IP
to communicate with other PowerCenter components.
The following table describes the connectivity required to connect the Repository Service to the repository
and source and target databases:
The PowerCenter Integration Service connects to the Repository Service to retrieve metadata when it runs
workflows.
Connecting to Databases
To set up a connection from the PowerCenter Repository Service to the repository database, configure the
database properties in Informatica Administrator. You must install and configure the native database drivers
for the repository database on the machine where the PowerCenter Repository Service runs.
The following table describes the connectivity required to connect the PowerCenter Integration Service to the
platform components, source databases, and target databases:
The PowerCenter Integration Service includes ODBC libraries that you can use to connect to other ODBC
sources. The Informatica installation includes ODBC drivers.
For flat file, XML, or COBOL sources, you can either access data with network connections, such as NFS, or
transfer data to the PowerCenter Integration Service node through FTP software. For information about
connectivity software for other ODBC sources, refer to your database documentation.
Connecting to Databases
Use the Workflow Manager to create connections to databases. You can create connections using native
database drivers or ODBC. If you use native drivers, specify the database user name, password, and native
connection string for each connection. The PowerCenter Integration Service uses this information to connect
to the database when it runs the session.
Note: PowerCenter supports ODBC drivers, such as ISG Navigator, that do not need user names and
passwords to connect. To avoid using empty strings or nulls, use the reserved words PmNullUser and
PmNullPasswd for the user name and password when you configure a database connection. The
PowerCenter Integration Service treats PmNullUser and PmNullPasswd as no user and no password.
Connecting to Databases
To connect to databases from the Designer, use the Windows ODBC Data Source Administrator to create a
data source for each database you want to access. Select the data source names in the Designer when you
perform the following tasks:
• Import a table or a stored procedure definition from a database. Use the Source Analyzer or Target
Designer to import the table from a database. Use the Transformation Developer, Mapplet Designer, or
Mapping Designer to import a stored procedure or a table for a Lookup transformation.
To connect to the database, you must also provide your database user name, password, and table or
stored procedure owner name.
• Preview data. You can select the data source name when you preview data in the Source Analyzer or
Target Designer. You must also provide your database user name, password, and table owner name.
JDBC drivers are installed with the Informatica services and the Informatica clients. You can use the installed
JDBC drivers to connect to the Metadata Manager repository.
The Informatica installers do not install ODBC drivers or the JDBC-ODBC bridge for the Metadata Manager
Service.
The following table describes the syntax for the native connection string for each supported database
system:
ODBC Connectivity
Open Database Connectivity (ODBC) provides a common way to communicate with different database
systems.
To use ODBC connectivity, you must install the following components on the machine hosting the Informatica
service or client tool:
• Database client software. Install the client software for the database system. This installs the client
libraries needed to connect to the database.
Note: Some ODBC drivers contain wire protocols and do not require the database client software.
• ODBC drivers. The DataDirect closed 32-bit or 64-bit ODBC drivers are installed when you install the
Informatica services. The DataDirect closed 32-bit ODBC drivers are installed when you install the
Informatica clients. The database server can also include an ODBC driver.
After you install the necessary components you must configure an ODBC data source for each database that
you want to connect to. A data source contains information that you need to locate and access the database,
such as database name, user name, and database password. On Windows, you use the ODBC Data Source
Administrator to create a data source name. On UNIX, you add data source entries to the odbc.ini file found in
the system $ODBCHOME directory.
When you create an ODBC data source, you must also specify the driver that the ODBC driver manager sends
database calls to.
JDBC Connectivity
JDBC (Java Database Connectivity) is a Java API that provides connectivity to relational databases. Java-
based applications can use JDBC drivers to connect to databases.
JDBC drivers are installed with the Informatica services and the Informatica clients.
To use the Administrator tool, configure the following options in the browser:
• Active scripting
• Allow programmatic clipboard access
• Run ActiveX controls and plug-ins
• Script ActiveX controls marked safe for scripting
To configure the controls, click Tools > Internet options > Security > Custom level.
Trusted sites
If the Informatica domain runs on a network with Kerberos authentication, you must configure the
browser to allow access to the Informatica web applications. In Microsoft Internet Explorer, Microsoft
Edge, and Google Chrome, add the URL of the Informatica web application to the list of trusted sites. In
Safari, add the certificate of the Informatica web application to the keychain. If you are using Chrome
version 86.0.42x or later on Windows, you must also set the AuthServerWhitelist and
AuthNegotiateDelegateWhitelist policies.
360
Index
361
Cassandra connections compute role
properties 130 nodes 93
catalina.out Configuration Support Manager
troubleshooting 267 using to analyze node diagnostics 301
category using to review node diagnostics 297
domain log events 276 Confluent Kafka connection
changing Confluent Kafka broker properties 132
password for user account 29 create using infacmd 133
character sizes general properties 132
double byte 310 connect string
multibyte 310 examples 358
single byte 310 syntax 358
COBOL connecting
connectivity 355 SQL data service 116
code page relaxation connecting to databases
compatible code pages, selecting 319 JDBC 357
configuring the Integration Service 319 Connection
data inconsistencies 318 details 179
overview 318 properties 179
troubleshooting 319 connection pooling
code page validation properties 118
overview 317 connection properties
relaxed validation 318 Databricks 133
code pages blockchain 128
Administrator tool 312 connection strings
application sources 313 native connectivity 358
application targets 314 connections
choosing 310 properties 142
compatibility diagram 315 adding pass-through security 117
compatibility overview 310 creating database connections 114
conversion 320 database identifier properties 212
Custom transformation 315 deleting 116
Data Integration Service process 330 editing 115
descriptions 332 Google PubSub 141
domain configuration database 311 overview 113
External Procedure transformation 315 pass-through security 116
flat file sources 313 refreshing 114
flat file targets 314 Salesforce Marketing Cloud 196
ID 332 testing 115
lookup database 315 web services properties 211
Metadata Manager Service 313 connectivity
names 332 COBOL 355
overview 308 connect string examples 358
pmcmd 312 Content Management Service 351
PowerCenter Client 312 Data Integration Service 351
PowerCenter Integration Service process 312, 330 diagram of 350
relational sources 313 Informatica Analyst 351
relational targets 314 Informatica Developer 351
relationships 317 Integration Service 355
relaxed validation for sources and targets 318 Metadata Manager 357
repository 313, 330 Model Repository Service 351
sort order overview 312 overview 350
sources 313, 332 PowerCenter Client 356
stored procedure database 315 PowerCenter Repository Service 355
supported code pages 330, 332 Content Management Service
targets 314, 332 application service 40
UNIX 309 connectivity 351
validation 317 Cosmos DB connection
Windows 309 creating 180
command line programs CPU detail
resilience, configuring 110 License Management Report 286
compatibility CPU profile
between code pages 310 node property 96
between source and target code pages 319 CPU summary
compatible License Management Report 285
defined for code page compatibility 310 CPUs
complete history statistics exceeding the limit 285
Web Services Report 294
362 Index
creating domain
Cosmos DB connection 180 log event categories 276
custom filters reports 284
date and time 263 user activity, monitoring 284
elapsed time 263 user security 77
multi-select 263 domain configuration
custom properties description 81
domain 90 log events 276
custom roles migrating 82
Analyst Service 342 domain configuration database
Metadata Manager Service 343 backing up 81
Operator 344 code page 311
PowerCenter Repository Service 345 connection for gateway node 84
description 81
migrating 82
D restoring 82
secure database 88
Data Analyzer updating 84
ODBC (Open Database Connectivity) 350 domain properties
Data Integration Service Informatica domain 86
application service 40 domain reports
connectivity 351 License Management Report 284
log events 277 running 284
recovery 106 Web Services Report 291
Data Integration Service process domains
supported code pages 330 multiple 72
viewing status 95
Data Integration Services
monitoring 240
data movement mode
E
ASCII 307 editing
changing 307 connections 115
description 306 schedules 219
effect on session files and caches 307 environment variables
overview 306 LANG_C 309
Unicode 307 LC_ALL 309
data object caching LC_CTYPE 309
with pass-through security 117 NLS_LANG 321, 324
database troubleshooting 79
domain configuration 81
database connections
identifier properties 212
updating for domain configuration 84
F
database drivers failover
Integration Service 350 application service 106
Repository Service 350 domain 105
database properties flat files
Informatica domain 87 connectivity 355
Databricks connection properties 133 exporting logs 274
DataDirect ODBC drivers source code page 313
platform-specific drivers required 358 target code page 314
deleting folders
connections 116 Administrator tool 75
schedules 220 creating 75
delimited identifiers managing 75
database connections 213 objects, moving 76
dependencies overview 40
application services 66 removing 76
grids 66 FTP
nodes 66 achieving high availability 111
viewing for services and nodes 66
deployed mapping jobs
monitoring 246
detailed statistics
G
monitoring 241 gateway
disable mode managing 80
PowerCenter Integration Services and Service Processes 78 gateway node
configuring 80
Index 363
gateway node (continued) IBM DB2 connections
description 92 properties 157
log directory 80 IBM DB2 for i5/OS connections
logging 267 properties 160
GB18030 IBM DB2 for z/OS connections
description 304 properties 163
general properties identifiers
Informatica domain 86 delimited 213
license 231 regular 213
globalization IME (Windows Input Method Editor)
overview 303 input locales 306
Google Analytics connections IMS connections
properties 137 properties 166
Google BigQuery connection incremental keys
properties 137 licenses 227
Google BigQuery connections Informatica Administrator
connection modes 139 logging in 32
Google Cloud Spanner connection Logs tab 56
properties 139 Manage tab 33, 38
Google Cloud Storage connections Monitor tab 48, 49
properties 140 Navigator 57
Google PubSub overview 31, 72
connection properties 141 Reports tab 57
graphics display server searching 57
requirement 284 Security page 57
Greenplum connections service process, enabling and disabling 78
properties 135 Services and Nodes view 39
grids services, enabling and disabling 78
dependencies 66 tabs, viewing 31
Informatica Administrator tabs 46 Informatica Analyst
searching 39 connectivity 351
groups Informatica Data Explorer
overview 58 connectivity 351
Guaranteed Message Delivery files Informatica Data Quality
Log Manager 266 connectivity 351
Informatica Data Services
connectivity 351
H Informatica Developer
connectivity 351
hardware configuration Informatica domain
License Management Report 288 alerts 73
HBase connections database properties 87
MapR-DB properties 151 description 17
properties 148 domain configuration database 88
HDFS connections domain properties 86
properties 149 general properties 86
high availability log and gateway configuration 88
description 25, 101 multiple domains 72
failover 105 permissions 77
recovery 106 privileges 77
restart 105 restarting 85
TCP KeepAlive timeout 111 shutting down 85
high availability persistence tables state of operations 106
PowerCenter Integration Service 109 user security 77
historical statistics Informatica Network
monitoring 241 logging in 298
Hive connections Information and Content Exchange (ICE)
properties 151 log files 274
Hive pushdown input locales
connection properties 142 configuring 306
HTTP connections IME (Windows Input Method Editor) 306
properties 155 Integration Service
connectivity 355
ODBC (Open Database Connectivity) 350
I
IBM DB2
connect string syntax 358
364 Index
J License Management Report (continued)
licensing 285
JD Edwards EnterpriseOne connection multibyte characters 290
properties 173 node configuration 288
JDBC (Java Database Connectivity) repository summary 287
overview 359 running 284, 289
JDBC connections Unicode font 290
properties 168 user detail 287
JDBC drivers user summary 287
Data Analyzer 350 license usage
installed drivers 357 log events 276
Metadata Manager 350 licensed options
Metadata Manager connection to databases 357 License Management Report 289
PowerCenter domain 350 licensing
Reference Table Manager 350 License Management Report 285
JDBC V2 connection log events 279
properties 171 managing 226
job scheduling licensing logs
overview 216 log events 226
job status linked domain
domain failover 262 multiple domains 72
jobs Listener Service
monitoring 241 log events 278
locales
overview 305
K localhost_.txt
troubleshooting 267
Kafka connection Log Agent
create using infacmd 177 description 265
general properties 175 log events 276
Kafka broker properties 175 log and gateway configuration
Kerberos authentication Informatica domain 88
troubleshooting 33 log directory
Kudu connection for gateway node 80
properties 177 location, configuring 268
log errors
Administrator tool 275
Index 365
log events (continued)
timestamps 276 M
user activity 280 Manage tab
user management 276 Connections view 46
viewing 270 Informatica Administrator 33, 38
Web Services Hub 280 Navigator 33, 38
workflow 260 Schedules view 47
log files Services and Nodes view 38
Mapping tasks 283 managing
Log Manager accounts 28
architecture 266 user accounts 28
catalina.out 267 Mapping task
configuring 270 log files 283
directory location, configuring 268 master gateway node
domain log events 276 description 92
log event components 276 Maximum CPU Run Queue Length
log events 276 node property 96
log events, purging 269 Maximum Memory Percent
log events, saving 273 node property 96
logs, viewing 270 Maximum Processes
message 276 node property 96
message code 276 Maximum Restart Attempts (property)
node 276 Informatica domain 79
node.log 267 message code
PowerCenter Integration Service log events 279 Log Manager 276
PowerCenter Repository Service log events 279 Messaging connection
ProcessID 276 Confluent Kafka connection 131
purge properties 269 Kafka connection 174
recovery 267 metadata
SAP NetWeaver BI log events 279 adding to repository 320
security audit trail 279 choosing characters 320
service name 276 Metadata Access Service
severity levels 276 application service 40
thread 276 Metadata Manager
time zone 270 connectivity 357
timestamp 276 ODBC (Open Database Connectivity) 350
troubleshooting 267 Metadata Manager Service
user activity log events 280 application service 40
using 265 code page 313
Logger Service custom roles 343
log events 278 log events 278
logical CPUs Microsoft Azure Data Lake Storage Gen1 connection
calculation 285 properties 181
logical data objects Microsoft Azure Data Lake Storage Gen2 connection
monitoring 249 properties 182
login Microsoft Azure SQL Data Warehouse connection
troubleshooting 33 properties 183
logs Microsoft SQL Server
components 276 connect string syntax 358
configuring 268 migrate
domain 276 domain configuration 82
location 268 Model Repository Service
PowerCenter Integration Service 279 application service 40
PowerCenter Repository Service 279 connectivity 351
purging 269 log events 278
SAP BW Service 279 Monitor tab
saving 273 Informatica Administrator 48, 49
user activity 280 monitoring
viewing 270 applications 245
workflow 260 configuring 236
Logs tab Data Integration Services 240
Informatica Administrator 56 deployed mapping jobs 246
lookup databases description 234
code pages 315 detailed statistics 241
exporting summary statistics 239
historical statistics 241
jobs 241
logical data objects 249
366 Index
monitoring (continued) nodes (continued)
preferences, configuring 237 searching 39
reports 53 shutting down 98
setup 235 starting 98
SQL data services 250 TCP/IP network protocol 350
statistics 52 types 92
summary statistics 238, 241 worker 92
viewing summary statistics 239
web services 253
workflows 254
monitoring Model Repository Service
O
application service 40 OData connections
MS SQL Server connections properties 190
properties 185 ODBC (Open Database Connectivity)
multibyte data DataDirect driver issues 358
entering in PowerCenter Client 306 establishing connectivity 358
Integration Service 350
Metadata Manager 350
Index 367
PowerCenter Client (continued) Reports tab
multibyte characters, entering 306 Informatica Administrator 57
ODBC (Open Database Connectivity) 350 repositories
resilience 102 backup directory 96
TCP/IP network protocol 350 code pages 313
PowerCenter domains supported code pages 330
connectivity 353 Unicode 304
TCP/IP network protocol 350 UTF-8 304
PowerCenter Integration Service repository metadata
application service 40 choosing characters 320
enabling and disabling 78 repository summary
failover configuration 109 License Management Report 287
high availability persistence tables 109 resilience
log events 279 application service 103
recovery 106 application service configuration 108
recovery configuration 109 command line program configuration 110
resilience 103 in exclusive mode 110
state of operations 106 PowerCenter Client 102
PowerCenter Integration Service process PowerCenter Integration Service 103
code page 312 PowerCenter Repository Service 103
enabling and disabling 78 TCP KeepAlive timeout 111
restart, configuring 79 Resource Manager Service
supported code pages 330 log events 279
viewing status 95 system services 44
PowerCenter Repository Service resource provision thresholds
application service 40 setting for nodes 96
connectivity requirements 355 restart
custom roles 345 application service 106
log events 279 configuring for PowerCenter Integration Service processes 79
recovery 106 restoring
resilience 103 domain configuration database 82
state of operations 106 roles
PowerCenter security nodes 93
managing 57 overview 59
PowerExchange Listener Service run-time statistics
application service 40 Web Services Report 293
PowerExchange Logger Service
application service 40
preferences
monitoring 237
S
process identification number Salesforce Marketing Cloud
Log Manager 276 connection properties 196
ProcessID SAP BW Service
Log Manager 276 application service 40
message code 276 log events 279
purge properties SAP connections
Log Manager 269 properties 197
Scheduler Service
log events 280
R schedules
creating schedules 217
recovery deleting 220
Data Integration Service 106 editing 219
high availability 106 overview 216
Integration Service 106 Search section
PowerCenter Repository Service 106 Informatica Administrator 57
regular identifiers security
database connections 213 audit trail, viewing 279
Reporting and Dashboards Service permissions 77
application service 40 privileges 77
Reporting Service Security page
application service 40 Informatica Administrator 57
reports Navigator 57
Administrator tool 284 Sequential connections
domain 284 properties 200
License 284 Service Manager
monitoring 53 authorization 18
Web Services 284 description 18
368 Index
Service Manager (continued) system services
log events 276 Resource Manager Service 44
service name
log events 276
service role
nodes 93
T
services Tableau V3 connection
searching 39 properties 206
services and nodes target databases
viewing dependencies 66 code page 314
Services and Nodes view targets
Informatica Administrator 39 code pages 314, 332
sessions tasks
sort order 312 states 257
severity TCP KeepAlive timeout
log events 276 high availability 111
Show Custom Properties (property) TCP/IP network protocol
user preference 29 nodes 350
shutting down PowerCenter Client 350
Informatica domain 85 PowerCenter domains 350
SMTP configuration requirement for Integration Service 356
alerts 73 Teradata
Snowflake connection connect string syntax 358
properties 202 Teradata Parallel Transporter connections
sort order properties 203
code page 312 testing
source databases database connections 115
code page 313 thread identification
sources Logs tab 276
code pages 313, 332 threads
Spark deploy mode Log Manager 276
Hadoop connection properties 142 time zone
Spark engine Log Manager 270
connection properties 142 timestamps
Spark Event Log directory Log Manager 276
Hadoop connection properties 142 troubleshooting
Spark execution parameters catalina.out 267
Hadoop connection properties 142 code page relaxation 319
Spark HDFS staging directory environment variables 79
Hadoop connection properties 142 Kerberos authentication 33
SQL data services localhost_.txt 267
monitoring 250 logging in 33
stack traces node.log 267
viewing 271 Twitter Streaming connections
state of operations properties 207
domain 106
PowerCenter Integration Service 106
PowerCenter Repository Service 106
statistics
U
for monitoring 52 UCS-2
Web Services Hub 291 description 304
stopping Unicode
Informatica domain 85 GB18030 304
stored procedures repositories 304
code pages 315 UCS-2 304
Subscribe for Alerts UTF-16 304
user preference 29 UTF-32 304
subset UTF-8 304
defined for code page compatibility 310 Unicode mode
summary statistics overview 306
monitoring 238 UNIX
superset code pages 309
defined for code page compatibility 310 UNIX environment variables
Sybase ASE LANG_C 309
connect string syntax 358 LC_ALL 309
system locales LC_CTYPE 309
description 305 user accounts
changing the password 29
Index 369
user accounts (continued) web services
managing 28 monitoring 253
user activity Web Services Hub
log event categories 280 application service 25, 40
user detail log events 280
License Management Report 287 statistics 291
user locales Web Services Report
description 306 activity data 292
user management Average Service Time (property) 292
log events 276 Avg DTM Time (property) 292
user preferences Avg. No. of Run Instances (property) 292
description 29 Avg. No. of Service Partitions (property) 292
user summary complete history statistics 294
License Management Report 287 contents 292
users Percent Partitions in Use (property) 292
license activity, monitoring 284 run-time statistics 293
overview 58 Within Restart Period (property)
UTF-16 Informatica domain 79
description 304 worker node
UTF-32 configuring as gateway 80
description 304 description 92
UTF-8 workflow recovery
description 304 overview 259
repository 313 running 260
workflows
aborting 259
V canceling 259
logs 260
validating monitoring 254
code pages 317 recovering 260
licenses 226 states 256
viewing
dependencies for services and nodes 66
VSAM connections
properties 208
X
X Virtual Frame Buffer
for License Report 284
W XML
for Web Services Report 284
370 Index