I installed IBM TRIRIGA Application Platform 3.5.3 in Linux CentOS 7. While running the TRIRIGA installer, I selected an embedded server, that is IBM WebSphere Application Server (WAS) Liberty Profile 126.96.36.199. This server is successfully installed and up and running.
From the log, I found that the server host where IBM TRIRIGA is accessed has an URL something like this: http://some.static.ab-xyz.com:8001. This URL value was taken by default. I did not provide this value. My question is: How do I change this host name URL? Where is the setting for this?
I’m not sure what log you’re looking at and what specifically you’re seeing. What is set as your FRONT_END_SERVER value in your TRIRIGAWEB.properties file?
[Admin: To see other related posts, use the Hostname tag.]
During TRIRIGA install, we encountered “Build Failed” and the Oracle WebLogic log contained the following. What happened?
####<Jan 23, 2018 4:47:14 PM ART> <Error> <Deployer>
<S-Tririga> <[ACTIVE] ExecuteThread: '3' for queue: 'weblogic.kernel.Default (self-tuning)'> <<WLS Kernel>> <> <7bf5687e-a5eb-47f9-9610-fd0f96db07ae-0000042e> <1516736834996> <BEA-149265> <Failure occurred in the execution of deployment request with ID "7285840482438638" for task "2". Error is:
[Admin: To see other related posts, use the WebLogic or Installation tag.]
Our customer has seen an issue when installing TRIRIGA 3.5.3 (Linux, Server build number: 276955) on an existing database (on 3.4.2 / 10.4.2). Everything goes well until starting up the server. Generally, TRIRIGA will run a database upgrade on the first startup when a build number difference is detected.
In the OM log, we notice that TRIRIGA tried to import the upgrade OM package… The import process started with the triPlatformObjectLabelManager package, but it failed to import a navigation item, which is newly created for Object Label Manager. I haven’t found any log which can explain this failure. I’ve checked the NAV_ITEM table. This navigation item wasn’t there before the upgrade process. Then all of the other packages are stuck on a pending status. Nothing happens after “Creating package from Zip file”. This behavior causes a lot of SQL update failures.
Meanwhile, on our Dev environment (Windows, Server build number: 279835), the upgrade went very well. You can find the difference in the logs. The OM log was set on “Debug” level on both servers. Note that the build number is slightly different between these two enviroments. Have you seen this kind of issue? Where can I find more details about the navigation item import failure?
[Admin: This post is related to the 02.17.17 post and 05.19.16 post about inconsistent OM validation results. To see other related posts, use the Object Migration tag or Upgrade tag.]
I am attempting a clean install of TRIRIGA 3.5.3 and according to the logs, it appears that the installation was successful within the was-ant.log. But I receive this error in the ant.log… When I access the application, I was unable to pull up the login page. Checking the server logs, I noticed these errors… The odd thing is that when I did a TRIRIGA 3.5.3 install on WebSphere Liberty, it worked perfectly…
The errors seem to indicate that you selected an upgrade, not a new install. Are you sure you selected a new install? Some of the environment properties that are needed to start up do not seem to be available. Did the database get created or are you pointing at an existing database? What happens if you do a select * from environment_properties on the database you are using?
The DB2 exception you are receiving seems to indicate you are running out of statement handles, but it’s hard to see this happen with a successful install.
[Admin: To see other related posts, use the WebSphere tag or “install error” search phrase.]
I made a big import of data in TRIRIGA, but the resources were not enough to proceed. So I have to stop TRIRIGA and truncate all events to stop the import. But now, the database log never stops growing and crashing the database. Is there a way to clean up and make TRIRIGA stable?
If this is Microsoft SQL Server, this may be related to the following SQL Server defect: SQL Server crashes when the log file of tempdb database is full in SQL Server 2012 or SQL Server 2014.
[Admin: To see other related posts, use the SQL tag or “sql server” search phrase.]
Is there a way to clear server caches without logging into the Admin Console?
Beginning in IBM TRIRIGA Platform 3.5.1, TRIRIGA delivered an enhancement for this to be done via workflow. The pertinent release notes can be found from this wiki page. Here is an excerpt from the release notes on the topic:
A custom task class has been added for workflow which triggers a global cache clear across all servers.
You can create a custom task and specify the following in the class field: com.tririga.platform.admin.cache.web.CacheProcessingCustomTask $RefreshAllCache
The custom task will perform a global cache clear on the server where the workflow runs as if it were triggered from that server’s Administrator Console. (Tri-211723)
[Admin: To see other related posts, use the Admin Console tag or Cache tag.]
We have quite a lot of rows in the extended formula queue that started to back up since Aug 29. At one point, we had over 900K in the queue and the process server is processing it pretty slowly. As of Sep 8, the system is still processing the queue. We are adding daily on average over 125K rows to the EF_QUEUE table.
We tried starting and stopping the Extended Formula Agent and Formula Recalc Agent and tried rebooting the process servers. We swapped the agents to run on different servers. We also ran the segment shrink on the EF_QUEUE table and gathered statistics. It is still extremely slow. We still have over 500K in the queue.
In parallel, we did open the platform logging on the extended formula activity and cache building, and we are working to identify the processes that can be fixed to reduce the multiple calls to the Extended Formula Agent queue. Do you have any recommendations to speed up the process to clear the extended formula queue?
Review the following two APARs. They may need to be applied to your environment to address this issue: (1) Platform: APAR IV87030. (2) Application: APAR IV87104.
[Admin: This post is related to the 07.28.16 post and 05.04.16 post about the Extended Formula (EF) Agent not processing and backing up the queue. To see other related posts, use the Extended Formula tag.]
After a successful load of the TRIRIGA CAD Integrator/Publisher (CI), the menu is not displayed. After enabling the log to C:\ProgramData\IBM\TRIRIGA\CAD Integrator\log\ci.out, the file says: “Could not load resource: Could not find a part of the path…” Why?
In fact, the main issue was that the CAD menu title was not globally activated. When the menu bar is set, the TRIRIGA CAD Integrator menu is now displayed.
[Admin: To see other related posts, use the Integrator tag or AutoCAD tag.]
I know that TRIRIGA SaaS comes with predefined ETL jobs, and one of them is energy log fact. I have some energy logs on some buildings. However, all of my attempts to run the ETL job for energy logs have failed. Is there any missing parameter for me to add to make this work? I did try adding triEnvEnergyItem as a BO name, but no luck.
The ETLs each require different inputs depending on the ETL. Since, in general, ETL processing is a background process and runs through the scheduler, you would need to look in the server.log to see what the ETL needs if you are running it directly from the ETL Job form. You would also need to make sure that you have a license that allows you to run the ETLs, which would be a license that grants privileges to “Technology Metrics”.
Maybe this link will help: ETL and Metric Query Troubleshooting.
[Admin: To see other related posts, use the ETL tag.]
Starting in version 3.5.3, IBM TRIRIGA Application Platform includes a Performance Analyzer. The Performance Analyzer is a system tool that helps you to diagnose system performance issues.
Performance Analyzer overview
- The Performance Analyzer provides a more streamlined approach to troubleshooting performance issues than the traditional IBM TRIRIGA performance log analysis.
- The Performance Analyzer helps you to better isolate and analyze the causes of performance issues by generating a log that is more targeted at the problem area. When you contact IBM to report a TRIRIGA performance issue, you might be asked to run the Performance Analyzer and provide the resulting data.
- When you select the performance timing categories that you want and start a performance run, the analyzer rolls the performance log and captures data while you do the process or actions that cause the slowness.
- When you stop the performance run, the analyzer automatically loads the performance data and does a high-level analysis that shows the longest running item in each category in the performance run. A result summary of the performance run shows the top events that took the longest time to run. You can further analyze the data by reviewing the result details where all the selected category data is displayed.
- You can also upload an existing performance log to the analyzer and the analyzer does the same high-level analysis of the data.
The Performance Analyzer has the same base requirements and restrictions as the traditional TRIRIGA performance logging, which includes nightly rolling, space consumption, and logs for all users. Details include…
[Admin: This post is related to the 02.10.17 post about finding the performance best practices. To see other related posts, use the Performance tag.]