What is the “Integration” check box for in the workflow Start task?
Assuming you are referring to the Start task of an asynchronous workflow, when this property is selected, the workflow is used to migrate data from staging tables in IBM TRIRIGA records. This type of workflow is used extensively in IBM TRIRIGA DataConnect.
Check out this IBM Knowledge Center topic about DataConnect that describes the “Integration” check box: Workflow task settings.
[Admin: The same question is also posted in the main Application Platform forum. To see other related posts, use the Staging tag or DataConnect tag.]
I’m seeing issues within report results, where if the user profile language is in US English, the results are of one type, but if the user profile language is in German, it is showing some other data. The record while checked is the same for both cases. While opening and checking the record, the ID and name of the record varies from what it showed as a result in the report. While investigating, what we observed is that these fields are marked as “Localizable” in Data Modeler. What is the use and impact of the “Localizable” field property? Any suggestions?
Here is a PDF link to the 3.5.3 Globalization (Localization) user guide. Also, perhaps this technote on localized database storage will help provide some insight.
[Admin: The same question is also posted in the main Application Platform forum. To see other related posts, use the Localization tag or Language tag.]
We have some reservable spaces with system geography and system organization settings. A non-Admin user also has the same geography settings. There are security groups for reservations, and organizations and geography security groups are assigned to him. The geo and org security groups have the same geo and org as the space and profile. But the non-Admin user still isn’t able to see spaces.
He is only able to see them when the first level of the org hierarchy is provided in the group (i.e. \Organization). But as soon as the second level is given in the group, he isn’t able to see them. Can anyone help me on this? I think there is some issue in the org, but I don’t know exactly where it is.
[Admin: To see other related posts, use the Geography tag or Organizations tag.]
Object Labels and Revisions
Starting in TRIRIGA 3.5.3, consider converting the TRIRIGA as-shipped objects that you renamed and modified back to the TRIRIGA objects from which they originated.
The purpose of the new conversion process is to enable the objects to use the new TRIRIGA object revisioning capabilities while preserving your object modifications. When you make future modifications to the objects, the modifications are saved in object revisions. Ultimately, the object revisioning and object labeling capabilities and tools aid in the future upgrade of your TRIRIGA applications.
For more information, see the documents, (1) “Converting your modified TRIRIGA objects to use the TRIRIGA revisioning features and application upgrade best practices” and (2) “Best practices for upgrading your TRIRIGA applications“.
[Admin: This post is related to the 09.24.15 post about the best practices for configuring and upgrading TRIRIGA applications. To see other related posts, use the Best Practices tag or Object Label tag.]
Is there a way to query which label class is being used on which fields? I want to tweak the existing ones slightly, but I’m not sure what is going to be updated when I make the change.
[Admin: To see other related posts, use the Labels tag or Query tag.]
You can configure TRIRIGA to use Tivoli Directory Integrator as its ETL runtime engine to run ETLJobItems from within TRIRIGA.
Before you begin
Install Tivoli Directory Integrator, if not already installed, on all the TRIRIGA systems that could run a TDI ETL Job Item. During the TDI install:
- Make note of the installation directory you enter on the Destination panel. You will enter this value later in TRIRIGAWEB.properties.
- Select either installation type. TRIRIGA requires only the TDI Server component.
- When prompted for the location of the Solution Directory, you can select any option. TRIRIGA specifies its own solution directory at runtime. However selecting the option “Use Install Directory” may simplify troubleshooting.
- Make note of the value you enter in the Server Port field on the Server Port Values Panel. You will enter this value later in TRIRIGAWEB.properties.
- Clear the “Start the Configuration Editor” check box on the Install Complete panel.
- Note: This step is very important for TDI/TRIRIGA integration to work. After you have installed Tivoli Directory Integrator, update it with the recommended fix packs (per TRIRIGA support matrix). TDI must be at least at FP04 (220.127.116.11) or it will not automatically start the TririgiaETLDispatch.xml assembly line which will result in ETL job items failing to run successfully.
- Edit TRIRIGAWEB.properties file to enable TRIRIGA to manage TDI server. Set the following properties…
- Install a JDBC driver library so that Tivoli Directory Integrator can use it to access TRIRIGA database…
- Edit TDI global.properties file to allow TRIRIGA to check and stop the TDI server from localhost without requiring authentication and authorization certificates. Set the api.remote.ssl.on property to false to tell TDI to trust requests from localhost…
- Start Tivoli Directory Integrator Agent from TRIRIGA Admin Console and verify that it starts successfully…
[Admin: This post is related to the 08.03.16 post about installing, upgrading, or uninstalling TRIRIGA TDI, and the 05.01.16 post about documentation on developing TDI with TRIRIGA. To see other related posts, use the TDI tag.]
Starting in version 3.5.3, IBM TRIRIGA Application Platform includes a Performance Analyzer. The Performance Analyzer is a system tool that helps you to diagnose system performance issues.
Performance Analyzer overview
- The Performance Analyzer provides a more streamlined approach to troubleshooting performance issues than the traditional IBM TRIRIGA performance log analysis.
- The Performance Analyzer helps you to better isolate and analyze the causes of performance issues by generating a log that is more targeted at the problem area. When you contact IBM to report a TRIRIGA performance issue, you might be asked to run the Performance Analyzer and provide the resulting data.
- When you select the performance timing categories that you want and start a performance run, the analyzer rolls the performance log and captures data while you do the process or actions that cause the slowness.
- When you stop the performance run, the analyzer automatically loads the performance data and does a high-level analysis that shows the longest running item in each category in the performance run. A result summary of the performance run shows the top events that took the longest time to run. You can further analyze the data by reviewing the result details where all the selected category data is displayed.
- You can also upload an existing performance log to the analyzer and the analyzer does the same high-level analysis of the data.
The Performance Analyzer has the same base requirements and restrictions as the traditional TRIRIGA performance logging, which includes nightly rolling, space consumption, and logs for all users. Details include…
[Admin: This post is related to the 02.10.17 post about finding the performance best practices. To see other related posts, use the Performance tag.]