Pyramid Usage Model and Analysis
Using Pyramid to analyze itself.
The following document describes how to build a data model using Pyramid’s database repository as a data source to be able to analyze content usage and performance in Pyramid itself.
The resulting model will give you insights into:
- WHO: Query transactions executed by users and tenants
- WHEN: Query transactions by time - down to the hour and minute level
- WHAT: Query transactions by data source server and type, database and model, with further details related to content items, the main content folders, specific hierarchies and measures used in each query (from the “elements” hierarchies)
- WHERE: Query transactions by processing server
- WHICH: Query transactions by which types of content
And of course, any combination of the above.
Setup
- Flow:
- Start a new advanced data model in Pyramid. Add a relevant data source (PostgreSQL, MS SQL Server, Oracle), depending on which type of repository you have elected to deploy
- From the database listing in the property cards, select the Pyramid database repository
- From the table selection, choose three views
- contentview*
- modelview**
- transactionview
*Available in 2020.24 and later versions
**Use the server_log_transaction_columns table instead for versions older than 2020.24
- Click Add Selected Nodes, to add these views to the flow
- Add a new Convert Columns node after the transactionview select node in the flow from the Column Operations side menu
- Change the date column to type “date”, with format yyyy-mm-dd
- Add a new In-memory target database node and provide a name for the database (“Pyramid Usage”)
- Start a new advanced data model in Pyramid. Add a relevant data source (PostgreSQL, MS SQL Server, Oracle), depending on which type of repository you have elected to deploy
- Model:
- Change the default measure to “Total Time” on the configuration tab
- On the Tables tab:
- Rename the modelview to “Elements”
- Rename the transactionview table to “Events”
- Make sure the Elements table is inner joined to the Events table using the “transaction_id” column
- Make sure the contentview table is inner joined to the Events table using the “item_id” column
- Change the following measures in the Events table to “Average” aggregations: column_count, row_count, cell_count, connection_time, post_query_time, pre_query_time, other_time, query_time, total_time
- Set client as a non-measure and select it
- Right-click on transaction_id column and add a count measure
- Optional: Right-click on user_name and add a distinct count measure (to count unique users)
- Under contentview, set item_type and sub_type as non-measures and select them
- On the Hierarchy tab:
- Add a time hierarchy based on Events with year, quarter-year, month-year, date, hour ,minute
- Add a content tree Parent-child hierarchy with Orphans Handling set to Roots and set the following Attribute Selections:
Child Key – item_id
Parent Key – folder_id
Caption - name
- Add a time hierarchy based on Events with year, quarter-year, month-year, date, hour ,minute
- Save and then Process your model.
The resulting model will take a snapshot of your transactional activity in Pyramid for further analysis.
Set up a schedule on this model to reprocess the database once a day, typically off-peak.
Metrics
Response Size Metrics
- Column Count: The number of RAW columns in the result set
- Row Count: The number of RAW rows in the result set
- Cell count = columns x rows
The RAW result set is not always indicative of the drawn result in the visual.
For instance, matrix grids can be much larger drawn than their underlying RAW results.
Time Metrics
- Pre-query: Pyramid engine time before the query is submitted.
- Post-query: Pyramid engine time after the query is submitted.
- Other: any time spent on scripting logic and dynamic functions as part of the query
- Server: pre-query + post-query + other
- Connection: the time taken to connect to the data source
- Query: the time data source takes to respond to the query
- Query Engine: Connection + Queries and Calculations
- Total: server + query engine + network + client
The time aggregates may be slightly off because:
- the measurements are in millisecond integers - so rounding is ignored.
- not every time element is always fully additive - because some items may run in parallel to each other (like 'connection' and 'other' times)
Using these metrics, you should be able to spot performance issues in your query designs vs data source performance vs Pyramid's processing performance.
12 replies
-
Hi Itamar,
This is very useful feature. But I have one question regarding this. The view is returning only data for last 4 months. Is there any way we can extend this period?
Thanks,
Uroš
-
Hi Uroš,
The Usage Model is based on the transactional data that is stored in the Pyramid database repository.
This data is stored in the db for a limited time which Administrators can change if needed.To configure Pyramid to keep the data for a longer period of time, you can change the "Keep Transactions Logs For These Days" setting in the Admin Console > Settings > Logging.
Before making this change, please make sure the database is capable of storing that much information.
-
For explanations on the model metrics see this posting.
https://community.pyramidanalytics.com/t/184nyf/pyramid-usage-model
-
Hi Itamar,
Is there a way to find exactly how many users are currently logged into Pyramid Application?
Thanks,
Shyam Pendyala. -
Oh, this seems to be be very useful...
Does it work for version 2020.27 as well?In our ORACLE repository DB I found two databases which sound reasonable - PYRAMID_APPLICATION (no views, only tables) and PYRAMID_ADMIN (empty). But I could not find especially the mentioned views...
THANKS anyway for the inspiration
Michael -
Hi Michael,
You should be able to see the views if you have selected the repository database. Please open a ticket for support to assist you with this
-
Now that it works on our server, I like to say it again: Very useful showcase!
But I still have issues with the TIME fragments. In the SQL code of the views I find the following:... trans.pre_query_time, trans.post_query_time, trans.r_script_time + trans.dynamic_function_time as other_time, trans.connection_time, trans.query_time, trans.engine_time, trans.post_query_time+trans.engine_time+trans.r_script_time + trans.dynamic_function_time as total_time, ...
Together with Itamar's initial explanations regarding the "Time Metrics" I think the first three columns sum up to something we could call "server_time". The sum "server_time" is not part of the view but it could easily be added, right?
Then, I understand that "engine_time" is more than the sum of "connection_time" and "query_time", due to the calculations which most probably cannot be extracted from the repository database.
But, why is "pre_query_time" not part of "total_time"? I understand that the time consumption for network and client is hard to retrieve, that is okay - but as Itamar explained, besides "engine_time", the entire "server_time" (= the possibly additional column) should belong to it!
And, finally, when I look at the concrete times on my server, only "post_query_time", "other_time" and "engine_time" add up to "total_time". And for some rows the sum of "connection_time" and "query_time" is much greater than "engine_time"! How can that be?
-
Hi,
We are using the usage model and it is very useful.
We started using the mobile devices for Pyramid in addition to the desktops, how is it possible to see the usage driven by mobile vs the desktop ?
Tnx,
Michael
Content aside
- 12 Likes
- 1 yr agoLast active
- 12Replies
- 1544Views
- 11 Following