-
Global information
- Generated on Mon Jan 5 09:59:47 2026
- Log file: /home/postgres/pg_data/data/pg_log/postgresql-2026-01-05_110000.log
- Parsed 1,808,180 log entries in 45s
- Log start from 2026-01-05 11:00:00 to 2026-01-05 11:59:44
-
Overview
Global Stats
- 330 Number of unique normalized queries
- 225,674 Number of queries
- 1h35m45s Total query duration
- 2026-01-05 11:00:00 First query
- 2026-01-05 11:59:44 Last query
- 4,813 queries/s at 2026-01-05 11:30:04 Query peak
- 1h35m45s Total query duration
- 5s399ms Prepare/parse total duration
- 47s30ms Bind total duration
- 1h34m53s Execute total duration
- 2 Number of events
- 2 Number of unique normalized events
- 1 Max number of times the same event was reported
- 0 Number of cancellation
- 41 Total number of automatic vacuums
- 57 Total number of automatic analyzes
- 843 Number temporary file
- 166.09 MiB Max size of temporary file
- 5.83 MiB Average size of temporary file
- 2,711 Total number of sessions
- 13 sessions at 2026-01-05 11:53:48 Session peak
- 2d1h35m50s Total duration of sessions
- 1m5s Average duration of sessions
- 83 Average queries per session
- 2s119ms Average queries duration per session
- 1m3s Average idle time per session
- 2,714 Total number of connections
- 29 connections/s at 2026-01-05 11:40:02 Connection peak
- 3 Total number of databases
SQL Traffic
Key values
- 4,813 queries/s Query Peak
- 2026-01-05 11:30:04 Date
SELECT Traffic
Key values
- 2,367 queries/s Query Peak
- 2026-01-05 11:30:04 Date
INSERT/UPDATE/DELETE Traffic
Key values
- 218 queries/s Query Peak
- 2026-01-05 11:00:55 Date
Queries duration
Key values
- 1h35m45s Total query duration
Prepared queries ratio
Key values
- 0.00 Ratio of bind vs prepare
- 0.00 % Ratio between prepared and "usual" statements
General Activity
↑ Back to the top of the General Activity tableDay Hour Count Min duration Max duration Avg duration Latency Percentile(90) Latency Percentile(95) Latency Percentile(99) Jan 05 11 225,674 0ms 15s73ms 25ms 3m12s 3m30s 4m11s Day Hour SELECT COPY TO Average Duration Latency Percentile(90) Latency Percentile(95) Latency Percentile(99) Jan 05 11 66,056 26 0ms 0ms 0ms 0ms Day Hour INSERT UPDATE DELETE COPY FROM Average Duration Latency Percentile(90) Latency Percentile(95) Latency Percentile(99) Jan 05 11 32,766 3,320 16 96 0ms 0ms 0ms 0ms Day Hour Prepare Bind Bind/Prepare Percentage of prepare Jan 05 11 17,599 73,698 4.19 16.15% Day Hour Count Average / Second Jan 05 11 2,714 0.75/s Day Hour Count Average Duration Average idle time Jan 05 11 2,711 1m5s 1m3s -
Connections
Established Connections
Key values
- 29 connections Connection Peak
- 2026-01-05 11:40:02 Date
Connections per database
Key values
- acaweb_fx Main Database
- 2,714 connections Total
Connections per user
Key values
- postgres Main User
- 2,714 connections Total
Connections per host
Key values
- 192.168.4.142 Main host with 1204 connections
- 2,714 Total connections
Host Count 127.0.0.1 113 192.168.0.114 6 192.168.0.216 102 192.168.0.236 1 192.168.0.74 93 192.168.1.127 38 192.168.1.145 58 192.168.1.15 74 192.168.1.20 69 192.168.1.23 2 192.168.1.239 3 192.168.1.90 57 192.168.2.126 67 192.168.2.182 12 192.168.2.82 48 192.168.3.199 36 192.168.4.104 1 192.168.4.142 1,204 192.168.4.150 10 192.168.4.238 12 192.168.4.240 4 192.168.4.26 7 192.168.4.33 92 192.168.4.98 330 [local] 275 -
Sessions
Simultaneous sessions
Key values
- 13 sessions Session Peak
- 2026-01-05 11:53:48 Date
Histogram of session times
Key values
- 2,079 0-500ms duration
Sessions per database
Key values
- acaweb_fx Main Database
- 2,711 sessions Total
Sessions per user
Key values
- postgres Main User
- 2,711 sessions Total
Sessions per host
Key values
- 192.168.4.142 Main Host
- 2,711 sessions Total
Host Count Total Duration Average Duration 127.0.0.1 113 10s843ms 95ms 192.168.0.114 5 29m32s 5m54s 192.168.0.216 102 1m12s 709ms 192.168.0.74 93 5h40m35s 3m39s 192.168.1.127 38 1m3s 1s670ms 192.168.1.145 58 6h5m26s 6m18s 192.168.1.15 74 3h30m36s 2m50s 192.168.1.20 69 13h24m21s 11m39s 192.168.1.23 1 5ms 5ms 192.168.1.239 3 18ms 6ms 192.168.1.90 57 39s695ms 696ms 192.168.2.126 67 10s613ms 158ms 192.168.2.182 12 850ms 70ms 192.168.2.82 48 22s256ms 463ms 192.168.3.199 36 1s339ms 37ms 192.168.4.104 1 268ms 268ms 192.168.4.142 1,204 10m21s 516ms 192.168.4.150 10 20h4m10s 2h25s 192.168.4.238 12 15s846ms 1s320ms 192.168.4.240 4 37ms 9ms 192.168.4.26 7 37s432ms 5s347ms 192.168.4.33 92 3m8s 2s48ms 192.168.4.98 330 14s763ms 44ms [local] 275 2m47s 608ms -
Checkpoints / Restartpoints
Checkpoints Buffers
Key values
- 11,027 buffers Checkpoint Peak
- 2026-01-05 11:08:56 Date
- 209.988 seconds Highest write time
- 0.009 seconds Sync time
Checkpoints Wal files
Key values
- 6 files Wal files usage Peak
- 2026-01-05 11:08:56 Date
Checkpoints distance
Key values
- 202.85 Mo Distance Peak
- 2026-01-05 11:08:56 Date
Checkpoints Activity
↑ Back to the top of the Checkpoint Activity tableDay Hour Written buffers Write time Sync time Total time Jan 05 11 43,968 2,192.043s 0.048s 2,192.409s Day Hour Added Removed Recycled Synced files Longest sync Average sync Jan 05 11 0 0 27 2,365 0.003s 0s Day Hour Count Avg time (sec) Jan 05 11 0 0s Day Hour Mean distance Mean estimate Jan 05 11 36,763.50 kB 77,860.58 kB -
Temporary Files
Size of temporary files
Key values
- 184.00 MiB Temp Files size Peak
- 2026-01-05 11:10:07 Date
Number of temporary files
Key values
- 62 per second Temp Files Peak
- 2026-01-05 11:32:09 Date
Temporary Files Activity
↑ Back to the top of the Temporary Files Activity tableDay Hour Count Total size Average size Jan 05 11 843 4.80 GiB 5.83 MiB Queries generating the most temporary files (N)
Rank Count Total size Min size Max size Avg size Query 1 29 1.65 GiB 3.55 MiB 166.09 MiB 58.30 MiB with rankedmt4 as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors ), last_feed_entry as ( select * from rankedmt4 where r = ? ), ok_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where status = ? ), earliest_entry_after_ok as ( select m.datafeedname, min(m.eventtimestamp) as eventtimestamp from mt4datafeederrors m left outer join ( select datafeedname, eventtimestamp from ok_entries where r = ?) oo on m.datafeedname = oo.datafeedname where m.eventtimestamp > coalesce(oo.eventtimestamp, ?::timestamp without time zone) group by m.datafeedname ), notified_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where notified is not null and notified <> ? ), broker as ( select *, row_number() over (partition by feedname order by brokerid) r from ( select distinct b.brokerid, b.name as brokername, dss.classname as feedname from downloadersymbolsettings dss inner join brokersymbollist bsl on dss.symbolid = bsl.symbolid inner join broker b on bsl.brokerid = b.brokerid where dss.enabled = ?) a ) select last.id, last.datafeedname, last.eventtimestamp, last.status, last.errordescription, last.serveraddress, last.username, note.notified, note.eventtimestamp, broker.brokername from last_feed_entry last inner join earliest_entry_after_ok after_ok on last.datafeedname = after_ok.datafeedname inner join broker on last.datafeedname = broker.feedname left outer join ok_entries ok on ok.datafeedname = last.datafeedname left outer join notified_entries note on note.datafeedname = last.datafeedname and note.r = ? where (ok.r is null or ok.r = ?) and last.datafeedname not in ( select distinct datafeedname from last_feed_entry where status = ?) and extract(epoch from (last.eventtimestamp - after_ok.eventtimestamp)) > ? * ? and last.eventtimestamp > current_timestamp - interval ? and (note.eventtimestamp is null or note.eventtimestamp < current_timestamp - interval ?) and last.eventtimestamp > current_timestamp - interval ? and broker.r = ?;-
with rankedmt4 as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors ), last_feed_entry as ( select * from rankedmt4 where r = 1 ), ok_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where status = 'OK' ), earliest_entry_after_ok as ( select m.datafeedname, min(m.eventtimestamp) as eventtimestamp from mt4datafeederrors m left outer join ( select datafeedname, eventtimestamp from ok_entries where r = 1) oo on m.datafeedname = oo.datafeedname where m.eventtimestamp > coalesce(oo.eventtimestamp, '1900-01-01'::timestamp without time zone) group by m.datafeedname ), notified_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where notified is not null and notified <> '' ), broker as ( select *, row_number() over (partition by feedname order by brokerid) r from ( select distinct b.brokerid, b.name as brokername, dss.classname as feedname from downloadersymbolsettings dss inner join brokersymbollist bsl on dss.symbolid = bsl.symbolid inner join broker b on bsl.brokerid = b.brokerid where dss.enabled = 1) a ) select last.id, last.datafeedname, last.eventtimestamp, last.status, last.errordescription, last.serveraddress, last.username, note.notified, note.eventtimestamp, broker.brokername from last_feed_entry last inner join earliest_entry_after_ok after_ok on last.datafeedname = after_ok.datafeedname inner join broker on last.datafeedname = broker.feedname left outer join ok_entries ok on ok.datafeedname = last.datafeedname left outer join notified_entries note on note.datafeedname = last.datafeedname and note.r = 1 where (ok.r is null or ok.r = 1) and last.datafeedname not in ( select distinct datafeedname from last_feed_entry where status = 'OK') and extract(epoch from (last.eventtimestamp - after_ok.eventtimestamp)) > 60 * 60 and last.eventtimestamp > current_timestamp - interval '1 day' and (note.eventtimestamp is null or note.eventtimestamp < current_timestamp - interval '10 hours') and last.eventtimestamp > current_timestamp - interval '1 hour' and broker.r = 1;
Date: 2026-01-05 11:00:06 Duration: 0ms
2 16 616.25 MiB 38.52 MiB 38.52 MiB 38.52 MiB update solr_relevance_old set new_hod_correct = sub.hod_correct, new_hod_percent = sub.hod_percent, new_hod_total = sub.hod_total, new_pattern_correct = sub.pattern_correct, new_pattern_percent = sub.pattern_percent, new_pattern_total = sub.pattern_total, new_percent = sub.percent, new_symbol_correct = sub.symbol_correct, new_symbol_percent = sub.symbol_percent, new_symbol_total = sub.symbol_total from ( select distinct resultuid, hod_correct, hod_percent, hod_total, hod, pattern_correct, pattern_percent, pattern_total, percent, symbol_correct, symbol_percent, symbol_total from whatshot_probability where type = ?) sub where result_uid = sub.resultuid;-
UPDATE solr_relevance_old SET new_hod_correct = sub.hod_correct, new_hod_percent = sub.hod_percent, new_hod_total = sub.hod_total, new_pattern_correct = sub.pattern_correct, new_pattern_percent = sub.pattern_percent, new_pattern_total = sub.pattern_total, new_percent = sub.percent, new_symbol_correct = sub.symbol_correct, new_symbol_percent = sub.symbol_percent, new_symbol_total = sub.symbol_total FROM ( select distinct resultuid, hod_correct, hod_percent, hod_total, hod, pattern_correct, pattern_percent, pattern_total, percent, symbol_correct, symbol_percent, symbol_total FROM whatshot_probability WHERE type = 'cp') sub WHERE result_uid = sub.resultuid;
Date: 2026-01-05 11:01:13 Duration: 0ms
3 16 1.11 GiB 70.77 MiB 70.77 MiB 70.77 MiB with max_ra as ( select resultuid from relevance_keylevels_results order by resultuid desc limit ?) update solr_relevance_old set newrelevant = sub.relevant, newage = sub.age from ( select so.uuid, case when ra.relevant is not null then ra.relevant when so.result_uid < max_ra.resultuid then ? else ? end as relevant, case when ra.age is not null then ra.age when so.result_uid < max_ra.resultuid then ? else ? end as age, so.result_uid from max_ra, solr_relevance_old so inner join keylevels_results k on so.result_uid = k.resultuid and so.uuid ilike ? inner join downloadersymbolsettings dss on k.symbolid = dss.symbolid left outer join relevance_keylevels_results ra on so.result_uid = ra.resultuid and so.uuid ilike ?) sub where solr_relevance_old.result_uid = sub.result_uid and solr_relevance_old.uuid ilike ?; update solr_relevance_old set newrelevant = ? where result_uid in ( select result_uid from solr_relevance_old s left outer join keylevels_results a on a.resultuid = s.result_uid where s.uuid ilike ? and a.resultuid is null); update solr_relevance_old set new_hod_correct = sub.hod_correct, new_hod_percent = sub.hod_percent, new_hod_total = sub.hod_total, new_pattern_correct = sub.pattern_correct, new_pattern_percent = sub.pattern_percent, new_pattern_total = sub.pattern_total, new_percent = sub.percent, new_symbol_correct = sub.symbol_correct, new_symbol_percent = sub.symbol_percent, new_symbol_total = sub.symbol_total from ( select distinct resultuid, hod_correct, hod_percent, hod_total, hod, pattern_correct, pattern_percent, pattern_total, percent, symbol_correct, symbol_percent, symbol_total from whatshot_probability where type in (...)) sub where result_uid = sub.resultuid;-
with max_ra as ( select resultuid from relevance_keylevels_results order by resultuid desc limit 1) update solr_relevance_old set newrelevant = sub.relevant, newage = sub.age from ( select so.uuid, case when ra.relevant is not null then ra.relevant when so.result_uid < max_ra.resultuid then 0 else 1 end as relevant, case when ra.age is not null then ra.age when so.result_uid < max_ra.resultuid then 11 else 0 end as age, so.result_uid from max_ra, solr_relevance_old so inner join keylevels_results k on so.result_uid = k.resultuid and so.uuid ilike 'kl_%' inner join downloadersymbolsettings dss on k.symbolid = dss.symbolid left outer join relevance_keylevels_results ra on so.result_uid = ra.resultuid and so.uuid ilike 'kl_%') sub where solr_relevance_old.result_uid = sub.result_uid and solr_relevance_old.uuid ilike 'kl_%'; update solr_relevance_old set newrelevant = 0 where result_uid in ( select result_uid from solr_relevance_old s left outer join keylevels_results a on a.resultuid = s.result_uid where s.uuid ilike 'kl_%' and a.resultuid is null); UPDATE solr_relevance_old SET new_hod_correct = sub.hod_correct, new_hod_percent = sub.hod_percent, new_hod_total = sub.hod_total, new_pattern_correct = sub.pattern_correct, new_pattern_percent = sub.pattern_percent, new_pattern_total = sub.pattern_total, new_percent = sub.percent, new_symbol_correct = sub.symbol_correct, new_symbol_percent = sub.symbol_percent, new_symbol_total = sub.symbol_total FROM ( select distinct resultuid, hod_correct, hod_percent, hod_total, hod, pattern_correct, pattern_percent, pattern_total, percent, symbol_correct, symbol_percent, symbol_total FROM whatshot_probability WHERE type in ('kl', 'ekl')) sub WHERE result_uid = sub.resultuid;
Date: 2026-01-05 11:01:17 Duration: 0ms
4 8 1010.35 MiB 126.27 MiB 126.31 MiB 126.29 MiB select updateresultsmaterializedview ();-
select updateresultsmaterializedview ();
Date: 2026-01-05 11:02:14 Duration: 0ms
5 4 367.30 MiB 91.78 MiB 91.89 MiB 91.82 MiB select updateageforrelevantresults ();-
select updateageforrelevantresults ();
Date: 2026-01-05 11:02:04 Duration: 0ms
Queries generating the largest temporary files
Rank Size Query 1 166.09 MiB with rankedmt4 as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors ), last_feed_entry as ( select * from rankedmt4 where r = 1 ), ok_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where status = 'OK' ), earliest_entry_after_ok as ( select m.datafeedname, min(m.eventtimestamp) as eventtimestamp from mt4datafeederrors m left outer join ( select datafeedname, eventtimestamp from ok_entries where r = 1) oo on m.datafeedname = oo.datafeedname where m.eventtimestamp > coalesce(oo.eventtimestamp, '1900-01-01'::timestamp without time zone) group by m.datafeedname ), notified_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where notified is not null and notified <> '' ), broker as ( select *, row_number() over (partition by feedname order by brokerid) r from ( select distinct b.brokerid, b.name as brokername, dss.classname as feedname from downloadersymbolsettings dss inner join brokersymbollist bsl on dss.symbolid = bsl.symbolid inner join broker b on bsl.brokerid = b.brokerid where dss.enabled = 1) a ) select last.id, last.datafeedname, last.eventtimestamp, last.status, last.errordescription, last.serveraddress, last.username, note.notified, note.eventtimestamp, broker.brokername from last_feed_entry last inner join earliest_entry_after_ok after_ok on last.datafeedname = after_ok.datafeedname inner join broker on last.datafeedname = broker.feedname left outer join ok_entries ok on ok.datafeedname = last.datafeedname left outer join notified_entries note on note.datafeedname = last.datafeedname and note.r = 1 where (ok.r is null or ok.r = 1) and last.datafeedname not in ( select distinct datafeedname from last_feed_entry where status = 'OK') and extract(epoch from (last.eventtimestamp - after_ok.eventtimestamp)) > 60 * 60 and last.eventtimestamp > current_timestamp - interval '1 day' and (note.eventtimestamp is null or note.eventtimestamp < current_timestamp - interval '10 hours') and last.eventtimestamp > current_timestamp - interval '1 hour' and broker.r = 1;[ Date: 2026-01-05 11:30:05 ]
2 163.68 MiB with rankedmt4 as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors ), last_feed_entry as ( select * from rankedmt4 where r = 1 ), ok_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where status = 'OK' ), earliest_entry_after_ok as ( select m.datafeedname, min(m.eventtimestamp) as eventtimestamp from mt4datafeederrors m left outer join ( select datafeedname, eventtimestamp from ok_entries where r = 1) oo on m.datafeedname = oo.datafeedname where m.eventtimestamp > coalesce(oo.eventtimestamp, '1900-01-01'::timestamp without time zone) group by m.datafeedname ), notified_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where notified is not null and notified <> '' ), broker as ( select *, row_number() over (partition by feedname order by brokerid) r from ( select distinct b.brokerid, b.name as brokername, dss.classname as feedname from downloadersymbolsettings dss inner join brokersymbollist bsl on dss.symbolid = bsl.symbolid inner join broker b on bsl.brokerid = b.brokerid where dss.enabled = 1) a ) select last.id, last.datafeedname, last.eventtimestamp, last.status, last.errordescription, last.serveraddress, last.username, note.notified, note.eventtimestamp, broker.brokername from last_feed_entry last inner join earliest_entry_after_ok after_ok on last.datafeedname = after_ok.datafeedname inner join broker on last.datafeedname = broker.feedname left outer join ok_entries ok on ok.datafeedname = last.datafeedname left outer join notified_entries note on note.datafeedname = last.datafeedname and note.r = 1 where (ok.r is null or ok.r = 1) and last.datafeedname not in ( select distinct datafeedname from last_feed_entry where status = 'OK') and extract(epoch from (last.eventtimestamp - after_ok.eventtimestamp)) > 60 * 60 and last.eventtimestamp > current_timestamp - interval '1 day' and (note.eventtimestamp is null or note.eventtimestamp < current_timestamp - interval '10 hours') and last.eventtimestamp > current_timestamp - interval '1 hour' and broker.r = 1;[ Date: 2026-01-05 11:50:05 ]
3 144.76 MiB with rankedmt4 as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors ), last_feed_entry as ( select * from rankedmt4 where r = 1 ), ok_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where status = 'OK' ), earliest_entry_after_ok as ( select m.datafeedname, min(m.eventtimestamp) as eventtimestamp from mt4datafeederrors m left outer join ( select datafeedname, eventtimestamp from ok_entries where r = 1) oo on m.datafeedname = oo.datafeedname where m.eventtimestamp > coalesce(oo.eventtimestamp, '1900-01-01'::timestamp without time zone) group by m.datafeedname ), notified_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where notified is not null and notified <> '' ), broker as ( select *, row_number() over (partition by feedname order by brokerid) r from ( select distinct b.brokerid, b.name as brokername, dss.classname as feedname from downloadersymbolsettings dss inner join brokersymbollist bsl on dss.symbolid = bsl.symbolid inner join broker b on bsl.brokerid = b.brokerid where dss.enabled = 1) a ) select last.id, last.datafeedname, last.eventtimestamp, last.status, last.errordescription, last.serveraddress, last.username, note.notified, note.eventtimestamp, broker.brokername from last_feed_entry last inner join earliest_entry_after_ok after_ok on last.datafeedname = after_ok.datafeedname inner join broker on last.datafeedname = broker.feedname left outer join ok_entries ok on ok.datafeedname = last.datafeedname left outer join notified_entries note on note.datafeedname = last.datafeedname and note.r = 1 where (ok.r is null or ok.r = 1) and last.datafeedname not in ( select distinct datafeedname from last_feed_entry where status = 'OK') and extract(epoch from (last.eventtimestamp - after_ok.eventtimestamp)) > 60 * 60 and last.eventtimestamp > current_timestamp - interval '1 day' and (note.eventtimestamp is null or note.eventtimestamp < current_timestamp - interval '10 hours') and last.eventtimestamp > current_timestamp - interval '1 hour' and broker.r = 1;[ Date: 2026-01-05 11:20:03 ]
4 126.31 MiB select updateresultsmaterializedview ();[ Date: 2026-01-05 11:17:14 ]
5 126.31 MiB select updateresultsmaterializedview ();[ Date: 2026-01-05 11:47:14 ]
6 126.30 MiB select updateresultsmaterializedview ();[ Date: 2026-01-05 11:32:14 ]
7 126.30 MiB select updateresultsmaterializedview ();[ Date: 2026-01-05 11:50:33 ]
8 126.29 MiB select updateresultsmaterializedview ();[ Date: 2026-01-05 11:20:33 ]
9 126.28 MiB select updateresultsmaterializedview ();[ Date: 2026-01-05 11:35:32 ]
10 126.28 MiB select updateresultsmaterializedview ();[ Date: 2026-01-05 11:02:14 ]
11 126.27 MiB select updateresultsmaterializedview ();[ Date: 2026-01-05 11:05:32 ]
12 109.73 MiB with rankedmt4 as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors ), last_feed_entry as ( select * from rankedmt4 where r = 1 ), ok_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where status = 'OK' ), earliest_entry_after_ok as ( select m.datafeedname, min(m.eventtimestamp) as eventtimestamp from mt4datafeederrors m left outer join ( select datafeedname, eventtimestamp from ok_entries where r = 1) oo on m.datafeedname = oo.datafeedname where m.eventtimestamp > coalesce(oo.eventtimestamp, '1900-01-01'::timestamp without time zone) group by m.datafeedname ), notified_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where notified is not null and notified <> '' ), broker as ( select *, row_number() over (partition by feedname order by brokerid) r from ( select distinct b.brokerid, b.name as brokername, dss.classname as feedname from downloadersymbolsettings dss inner join brokersymbollist bsl on dss.symbolid = bsl.symbolid inner join broker b on bsl.brokerid = b.brokerid where dss.enabled = 1) a ) select last.id, last.datafeedname, last.eventtimestamp, last.status, last.errordescription, last.serveraddress, last.username, note.notified, note.eventtimestamp, broker.brokername from last_feed_entry last inner join earliest_entry_after_ok after_ok on last.datafeedname = after_ok.datafeedname inner join broker on last.datafeedname = broker.feedname left outer join ok_entries ok on ok.datafeedname = last.datafeedname left outer join notified_entries note on note.datafeedname = last.datafeedname and note.r = 1 where (ok.r is null or ok.r = 1) and last.datafeedname not in ( select distinct datafeedname from last_feed_entry where status = 'OK') and extract(epoch from (last.eventtimestamp - after_ok.eventtimestamp)) > 60 * 60 and last.eventtimestamp > current_timestamp - interval '1 day' and (note.eventtimestamp is null or note.eventtimestamp < current_timestamp - interval '10 hours') and last.eventtimestamp > current_timestamp - interval '1 hour' and broker.r = 1;[ Date: 2026-01-05 11:00:05 ]
13 101.62 MiB with rankedmt4 as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors ), last_feed_entry as ( select * from rankedmt4 where r = 1 ), ok_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where status = 'OK' ), earliest_entry_after_ok as ( select m.datafeedname, min(m.eventtimestamp) as eventtimestamp from mt4datafeederrors m left outer join ( select datafeedname, eventtimestamp from ok_entries where r = 1) oo on m.datafeedname = oo.datafeedname where m.eventtimestamp > coalesce(oo.eventtimestamp, '1900-01-01'::timestamp without time zone) group by m.datafeedname ), notified_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where notified is not null and notified <> '' ), broker as ( select *, row_number() over (partition by feedname order by brokerid) r from ( select distinct b.brokerid, b.name as brokername, dss.classname as feedname from downloadersymbolsettings dss inner join brokersymbollist bsl on dss.symbolid = bsl.symbolid inner join broker b on bsl.brokerid = b.brokerid where dss.enabled = 1) a ) select last.id, last.datafeedname, last.eventtimestamp, last.status, last.errordescription, last.serveraddress, last.username, note.notified, note.eventtimestamp, broker.brokername from last_feed_entry last inner join earliest_entry_after_ok after_ok on last.datafeedname = after_ok.datafeedname inner join broker on last.datafeedname = broker.feedname left outer join ok_entries ok on ok.datafeedname = last.datafeedname left outer join notified_entries note on note.datafeedname = last.datafeedname and note.r = 1 where (ok.r is null or ok.r = 1) and last.datafeedname not in ( select distinct datafeedname from last_feed_entry where status = 'OK') and extract(epoch from (last.eventtimestamp - after_ok.eventtimestamp)) > 60 * 60 and last.eventtimestamp > current_timestamp - interval '1 day' and (note.eventtimestamp is null or note.eventtimestamp < current_timestamp - interval '10 hours') and last.eventtimestamp > current_timestamp - interval '1 hour' and broker.r = 1;[ Date: 2026-01-05 11:40:05 ]
14 100.36 MiB with rankedmt4 as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors ), last_feed_entry as ( select * from rankedmt4 where r = 1 ), ok_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where status = 'OK' ), earliest_entry_after_ok as ( select m.datafeedname, min(m.eventtimestamp) as eventtimestamp from mt4datafeederrors m left outer join ( select datafeedname, eventtimestamp from ok_entries where r = 1) oo on m.datafeedname = oo.datafeedname where m.eventtimestamp > coalesce(oo.eventtimestamp, '1900-01-01'::timestamp without time zone) group by m.datafeedname ), notified_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where notified is not null and notified <> '' ), broker as ( select *, row_number() over (partition by feedname order by brokerid) r from ( select distinct b.brokerid, b.name as brokername, dss.classname as feedname from downloadersymbolsettings dss inner join brokersymbollist bsl on dss.symbolid = bsl.symbolid inner join broker b on bsl.brokerid = b.brokerid where dss.enabled = 1) a ) select last.id, last.datafeedname, last.eventtimestamp, last.status, last.errordescription, last.serveraddress, last.username, note.notified, note.eventtimestamp, broker.brokername from last_feed_entry last inner join earliest_entry_after_ok after_ok on last.datafeedname = after_ok.datafeedname inner join broker on last.datafeedname = broker.feedname left outer join ok_entries ok on ok.datafeedname = last.datafeedname left outer join notified_entries note on note.datafeedname = last.datafeedname and note.r = 1 where (ok.r is null or ok.r = 1) and last.datafeedname not in ( select distinct datafeedname from last_feed_entry where status = 'OK') and extract(epoch from (last.eventtimestamp - after_ok.eventtimestamp)) > 60 * 60 and last.eventtimestamp > current_timestamp - interval '1 day' and (note.eventtimestamp is null or note.eventtimestamp < current_timestamp - interval '10 hours') and last.eventtimestamp > current_timestamp - interval '1 hour' and broker.r = 1;[ Date: 2026-01-05 11:10:04 ]
15 100.01 MiB with rankedmt4 as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors ), last_feed_entry as ( select * from rankedmt4 where r = 1 ), ok_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where status = 'OK' ), earliest_entry_after_ok as ( select m.datafeedname, min(m.eventtimestamp) as eventtimestamp from mt4datafeederrors m left outer join ( select datafeedname, eventtimestamp from ok_entries where r = 1) oo on m.datafeedname = oo.datafeedname where m.eventtimestamp > coalesce(oo.eventtimestamp, '1900-01-01'::timestamp without time zone) group by m.datafeedname ), notified_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where notified is not null and notified <> '' ), broker as ( select *, row_number() over (partition by feedname order by brokerid) r from ( select distinct b.brokerid, b.name as brokername, dss.classname as feedname from downloadersymbolsettings dss inner join brokersymbollist bsl on dss.symbolid = bsl.symbolid inner join broker b on bsl.brokerid = b.brokerid where dss.enabled = 1) a ) select last.id, last.datafeedname, last.eventtimestamp, last.status, last.errordescription, last.serveraddress, last.username, note.notified, note.eventtimestamp, broker.brokername from last_feed_entry last inner join earliest_entry_after_ok after_ok on last.datafeedname = after_ok.datafeedname inner join broker on last.datafeedname = broker.feedname left outer join ok_entries ok on ok.datafeedname = last.datafeedname left outer join notified_entries note on note.datafeedname = last.datafeedname and note.r = 1 where (ok.r is null or ok.r = 1) and last.datafeedname not in ( select distinct datafeedname from last_feed_entry where status = 'OK') and extract(epoch from (last.eventtimestamp - after_ok.eventtimestamp)) > 60 * 60 and last.eventtimestamp > current_timestamp - interval '1 day' and (note.eventtimestamp is null or note.eventtimestamp < current_timestamp - interval '10 hours') and last.eventtimestamp > current_timestamp - interval '1 hour' and broker.r = 1;[ Date: 2026-01-05 11:40:05 ]
16 95.33 MiB with rankedmt4 as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors ), last_feed_entry as ( select * from rankedmt4 where r = 1 ), ok_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where status = 'OK' ), earliest_entry_after_ok as ( select m.datafeedname, min(m.eventtimestamp) as eventtimestamp from mt4datafeederrors m left outer join ( select datafeedname, eventtimestamp from ok_entries where r = 1) oo on m.datafeedname = oo.datafeedname where m.eventtimestamp > coalesce(oo.eventtimestamp, '1900-01-01'::timestamp without time zone) group by m.datafeedname ), notified_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where notified is not null and notified <> '' ), broker as ( select *, row_number() over (partition by feedname order by brokerid) r from ( select distinct b.brokerid, b.name as brokername, dss.classname as feedname from downloadersymbolsettings dss inner join brokersymbollist bsl on dss.symbolid = bsl.symbolid inner join broker b on bsl.brokerid = b.brokerid where dss.enabled = 1) a ) select last.id, last.datafeedname, last.eventtimestamp, last.status, last.errordescription, last.serveraddress, last.username, note.notified, note.eventtimestamp, broker.brokername from last_feed_entry last inner join earliest_entry_after_ok after_ok on last.datafeedname = after_ok.datafeedname inner join broker on last.datafeedname = broker.feedname left outer join ok_entries ok on ok.datafeedname = last.datafeedname left outer join notified_entries note on note.datafeedname = last.datafeedname and note.r = 1 where (ok.r is null or ok.r = 1) and last.datafeedname not in ( select distinct datafeedname from last_feed_entry where status = 'OK') and extract(epoch from (last.eventtimestamp - after_ok.eventtimestamp)) > 60 * 60 and last.eventtimestamp > current_timestamp - interval '1 day' and (note.eventtimestamp is null or note.eventtimestamp < current_timestamp - interval '10 hours') and last.eventtimestamp > current_timestamp - interval '1 hour' and broker.r = 1;[ Date: 2026-01-05 11:10:05 ]
17 91.89 MiB select updateageforrelevantresults ();[ Date: 2026-01-05 11:02:04 ]
18 91.84 MiB select updateageforrelevantresults ();[ Date: 2026-01-05 11:32:05 ]
19 91.79 MiB select updateageforrelevantresults ();[ Date: 2026-01-05 11:17:05 ]
20 91.78 MiB select updateageforrelevantresults ();[ Date: 2026-01-05 11:47:05 ]
-
Vacuums
Vacuums / Analyzes Distribution
Key values
- 0 sec Highest CPU-cost vacuum
Table
Database - Date
- 0 sec Highest CPU-cost analyze
Table
Database - Date
Analyzes per table
Key values
- public.solr_relevance_old (16) Main table analyzed (database acaweb_fx)
- 57 analyzes Total
Table Number of analyzes acaweb_fx.public.solr_relevance_old 16 acaweb_fx.pg_catalog.pg_attribute 5 acaweb_fx.public.datafeeds_latestrun 4 acaweb_fx.public.relevance_keylevels_results 4 acaweb_fx.pg_catalog.pg_class 4 acaweb_fx.public.relevance_fibonacci_results 4 acaweb_fx.public.relevance_autochartist_results 4 acaweb_fx.pg_catalog.pg_type 3 acaweb_fx.pg_catalog.pg_index 2 acaweb_fx.public.autochartist_symbolupdates 2 acaweb_fx.public.latest_t15_candle_view 2 acaweb_fx.pg_catalog.pg_depend 2 socialmedia.public.processes 1 acaweb_fx.public.latest_candle_datetime_per_receng 1 acaweb_fx.public.patternresultsage 1 acaweb_fx.public.relevance_consecutivecandles_results 1 acaweb_fx.public.symbollatestupdatetime 1 Total 57 Vacuums per table
Key values
- public.solr_relevance_old (16) Main table vacuumed on database acaweb_fx
- 41 vacuums Total
Index Buffer usage Skipped WAL usage Table Vacuums scans hits misses dirtied pins frozen records full page bytes acaweb_fx.public.solr_relevance_old 16 16 13,116 0 54 0 0 8,405 1,071 5,455,318 acaweb_fx.public.datafeeds_latestrun 4 0 480 0 12 0 0 60 12 65,388 acaweb_fx.public.relevance_fibonacci_results 3 3 3,830 0 154 4 143 693 204 532,527 acaweb_fx.public.relevance_autochartist_results 3 3 10,553 0 367 2 717 2,087 383 704,058 acaweb_fx.pg_catalog.pg_attribute 2 2 1,595 0 358 0 134 734 274 1,635,771 acaweb_fx.public.relevance_keylevels_results 2 2 8,503 0 535 4 96 2,529 540 1,061,920 acaweb_fx.pg_catalog.pg_class 2 2 926 0 118 0 0 290 115 572,604 acaweb_fx.pg_catalog.pg_index 1 1 88 0 11 0 0 27 10 75,230 acaweb_fx.pg_toast.pg_toast_2619 1 1 164 0 46 0 0 122 38 152,967 acaweb_fx.public.autochartist_symbolupdates 1 1 24,689 0 3,423 2 37,429 8,261 3,430 1,595,088 acaweb_fx.pg_catalog.pg_depend 1 1 377 0 72 0 59 163 63 330,702 acaweb_fx.public.latest_t15_candle_view 1 1 66 0 1 0 0 6 1 9,007 acaweb_fx.pg_catalog.pg_type 1 1 138 0 19 0 0 45 14 91,704 acaweb_fx.pg_catalog.pg_statistic 1 1 994 0 140 0 582 466 124 515,142 acaweb_fx.public.relevance_consecutivecandles_results 1 1 69 0 8 1 0 22 7 38,108 acaweb_fx.public.symbollatestupdatetime 1 0 1,458 0 157 0 666 1,131 150 451,813 Total 41 36 67,046 51,309 5,475 13 39,826 25,041 6,436 13,287,347 Tuples removed per table
Key values
- public.solr_relevance_old (24458) Main table with removed tuples on database acaweb_fx
- 57759 tuples Total removed
Index Tuples Pages Table Vacuums scans removed remain not yet removable removed remain acaweb_fx.public.solr_relevance_old 16 16 24,458 91,633 0 0 3,189 acaweb_fx.public.symbollatestupdatetime 1 0 19,128 89,230 0 0 1,714 acaweb_fx.public.autochartist_symbolupdates 1 1 5,673 53,782 1 0 40,691 acaweb_fx.pg_catalog.pg_attribute 2 2 2,908 21,845 0 40 480 acaweb_fx.public.relevance_keylevels_results 2 2 2,089 23,561 0 0 558 acaweb_fx.public.relevance_autochartist_results 3 3 1,149 24,446 781 0 1,140 acaweb_fx.pg_catalog.pg_depend 1 1 578 14,647 0 0 135 acaweb_fx.pg_catalog.pg_statistic 1 1 549 3,743 0 0 1,194 acaweb_fx.public.relevance_fibonacci_results 3 3 429 4,156 0 0 306 acaweb_fx.pg_catalog.pg_class 2 2 294 3,298 0 0 300 acaweb_fx.public.datafeeds_latestrun 4 0 247 56 0 0 64 acaweb_fx.pg_toast.pg_toast_2619 1 1 67 169 0 0 53 acaweb_fx.public.relevance_consecutivecandles_results 1 1 63 326 0 0 7 acaweb_fx.pg_catalog.pg_type 1 1 59 1,482 36 0 38 acaweb_fx.public.latest_t15_candle_view 1 1 53 14 0 0 1 acaweb_fx.pg_catalog.pg_index 1 1 15 813 0 0 22 Total 41 36 57,759 333,201 818 40 49,892 Pages removed per table
Key values
- pg_catalog.pg_attribute (40) Main table with removed pages on database acaweb_fx
- 40 pages Total removed
Table Number of vacuums Index scans Tuples removed Pages removed acaweb_fx.pg_catalog.pg_attribute 2 2 2908 40 acaweb_fx.pg_catalog.pg_index 1 1 15 0 acaweb_fx.pg_toast.pg_toast_2619 1 1 67 0 acaweb_fx.public.autochartist_symbolupdates 1 1 5673 0 acaweb_fx.public.datafeeds_latestrun 4 0 247 0 acaweb_fx.pg_catalog.pg_depend 1 1 578 0 acaweb_fx.public.latest_t15_candle_view 1 1 53 0 acaweb_fx.public.relevance_keylevels_results 2 2 2089 0 acaweb_fx.pg_catalog.pg_class 2 2 294 0 acaweb_fx.public.relevance_fibonacci_results 3 3 429 0 acaweb_fx.pg_catalog.pg_type 1 1 59 0 acaweb_fx.pg_catalog.pg_statistic 1 1 549 0 acaweb_fx.public.relevance_consecutivecandles_results 1 1 63 0 acaweb_fx.public.symbollatestupdatetime 1 0 19128 0 acaweb_fx.public.solr_relevance_old 16 16 24458 0 acaweb_fx.public.relevance_autochartist_results 3 3 1149 0 Total 41 36 57,759 40 Autovacuum Activity
↑ Back to the top of the Autovacuum Activity tableDay Hour VACUUMs ANALYZEs Jan 05 11 41 57 - 0 sec Highest CPU-cost vacuum
-
Locks
Locks by types
Key values
- ShareLock Main Lock Type
- 1 locks Total
Most frequent waiting queries (N)
Rank Count Total time Min time Max time Avg duration Query 1 1 1s424ms 1s424ms 1s424ms 1s424ms insert into t60 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) values (?, ?, ?, ?, ?, ?, ?, ?, ?, ?) on conflict (pricedatetime, symbolid) do update set open = ?, high = ?, low = ?, close = ?, volume = ?, bsf = ?, sastdatetimewritten = ?, sastdatetimereceived = ?;-
INSERT INTO T60 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;
Date: 2026-01-05 11:06:16
Queries that waited the most
Rank Wait time Query 1 1s424ms INSERT INTO T60 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;[ Date: 2026-01-05 11:06:16 ]
-
Queries
Queries by type
Key values
- 66,056 Total read queries
- 42,906 Total write queries
Queries by database
Key values
- unknown Main database
- 224,619 Requests
- 1h34m53s (unknown)
- Main time consuming database
Database Request type Count Duration acaweb_fx Total 921 0ms copy from 80 0ms copy to 26 0ms cte 104 0ms ddl 16 0ms delete 16 0ms others 205 0ms select 103 0ms tcl 332 0ms update 39 0ms socialmedia Total 134 0ms others 38 0ms select 92 0ms tcl 4 0ms unknown Total 224,619 1h34m53s copy from 16 0ms cte 5,814 0ms insert 32,766 0ms others 3,631 0ms select 65,861 0ms tcl 412 0ms update 3,281 0ms Queries by user
Key values
- unknown Main user
- 224,619 Requests
User Request type Count Duration postgres Total 1,055 0ms copy from 80 0ms copy to 26 0ms cte 104 0ms ddl 16 0ms delete 16 0ms others 243 0ms select 195 0ms tcl 336 0ms update 39 0ms unknown Total 224,619 1h34m53s copy from 16 0ms cte 5,814 0ms insert 32,766 0ms others 3,631 0ms select 65,861 0ms tcl 412 0ms update 3,281 0ms Duration by user
Key values
- 1h34m53s (unknown) Main time consuming user
User Request type Count Duration postgres Total 1,055 0ms copy from 80 0ms copy to 26 0ms cte 104 0ms ddl 16 0ms delete 16 0ms others 243 0ms select 195 0ms tcl 336 0ms update 39 0ms unknown Total 224,619 1h34m53s copy from 16 0ms cte 5,814 0ms insert 32,766 0ms others 3,631 0ms select 65,861 0ms tcl 412 0ms update 3,281 0ms Queries by host
Key values
- unknown Main host
- 225,674 Requests
- 1h34m53s (unknown)
- Main time consuming host
Queries by application
Key values
- unknown Main application
- 225,286 Requests
- 1h34m53s (unknown)
- Main time consuming application
Number of cancelled queries
Key values
- 0 per second Cancelled query Peak
- 2026-01-05 11:11:06 Date
Number of cancelled queries (5 minutes period)
NO DATASET
-
Top Queries
Histogram of query times
Key values
- 69,566 0-1ms duration
Slowest individual queries
Rank Duration Query NO DATASET
Time consuming queries
Rank Total duration Times executed Min duration Max duration Avg duration Query 1 0ms 1 0ms 0ms 0ms insert into t30 (symbolid, pricedatetime, open, high, low, close, volume, bsf, sastdatetimereceived) values (?, ?::timestamp without time zone, ?.?, ?.?, ?.?, ?, ?, ?, ?::timestamp without time zone) on conflict (symbolid, pricedatetime) do nothing;Times Reported Time consuming queries #1
Day Hour Count Duration Avg duration Jan 05 11 1 0ms 0ms 2 0ms 40 0ms 0ms 0ms select key, value from datasources ds inner join datasourceparams dsp on ds.id = dsp.datasourceid where ds.name = ?;Times Reported Time consuming queries #2
Day Hour Count Duration Avg duration Jan 05 11 40 0ms 0ms 3 0ms 1 0ms 0ms 0ms select "public"."processparameters"."id" AS "id", "public"."processparameters"."processid" AS "processid", "public"."processparameters"."key" AS "key", "public"."processparameters"."value" AS "value" from "public"."processparameters" where "public"."processparameters"."id" = ? and "public"."processparameters"."id" = ? limit ? offset ?;Times Reported Time consuming queries #3
Day Hour Count Duration Avg duration Jan 05 11 1 0ms 0ms 4 0ms 1 0ms 0ms 0ms select distinct "public"."processes"."live" AS "live" from "public"."processes" left outer join "public"."brokers" "LT?" on "LT?"."id" = "public"."processes"."brokerid" left outer join "public"."contenttypes" "LT?" on "LT?"."id" = "public"."processes"."contenttypeid" where "public"."processes"."id" = ? and "public"."processes"."id" = ? order by ? asc;Times Reported Time consuming queries #4
Day Hour Count Duration Avg duration Jan 05 11 1 0ms 0ms 5 0ms 10 0ms 0ms 0ms with rar_max as ( select resultuid from relevance_bigmovement_results order by resultuid desc limit ? ) select bmr.symbolid, patternstarttime, patternendtime, timegranularity, ? as direction, case when bmr.old_resultuid = ? then bmr.old_resultuid else bmr.resultuid end as uid, s.exchange, s.symbol, s.longname, s.shortname, dtt.timezone, bmr.patternmovement, bmr.statisticalmovement, bmr.fromprice, bmr.toprice, bmr.percentile, bmr.patternlengthbars, case when rbr.age is not null then rbr.age when bmr.resultuid <= rm.resultuid then ? else ? end as age, case when rbr.relevant is not null then rbr.relevant when bmr.resultuid <= rm.resultuid then ? else ? end as relevant, cps.pip from bigmovement_results bmr inner join downloadersymbolsettings dss on bmr.symbolid = dss.symbolid inner join datafeedstimetable dtt on dss.classname = dtt.classname inner join symbols s on bmr.symbolid = s.symbolid inner join rar_max rm on ? = ? left outer join relevance_bigmovement_results rbr on rbr.resultuid = bmr.resultuid left join currencypips cps on cps.symbol = s.symbol where (bmr.old_resultuid = ? or bmr.resultuid = ?) and dtt.dayofweek = ?;Times Reported Time consuming queries #5
Day Hour Count Duration Avg duration Jan 05 11 10 0ms 0ms 6 0ms 2,387 0ms 0ms 0ms insert into t60 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) values (?, ?, ?, ?, ?, ?, ?, ?, ?, ?) on conflict (pricedatetime, symbolid) do update set open = ?, high = ?, low = ?, close = ?, volume = ?, bsf = ?, sastdatetimewritten = ?, sastdatetimereceived = ?;Times Reported Time consuming queries #6
Day Hour Count Duration Avg duration Jan 05 11 2,387 0ms 0ms -
INSERT INTO T60 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;
Date: 2026-01-05 11:06:16 Duration: 0ms
7 0ms 48 0ms 0ms 0ms select count(*) from datafeeds_latestrun where feedname ilike ? and ((latestrxtime > current_timestamp - interval ? and latestdbwritetime > current_timestamp - interval ?) or (latestdbwritetime > current_timestamp - interval ? and lateststartuptime > current_timestamp - interval ?));Times Reported Time consuming queries #7
Day Hour Count Duration Avg duration Jan 05 11 48 0ms 0ms 8 0ms 4 0ms 0ms 0ms select updaterelevantforrelevantresults ();Times Reported Time consuming queries #8
Day Hour Count Duration Avg duration Jan 05 11 4 0ms 0ms 9 0ms 38 0ms 0ms 0ms set datestyle = iso;Times Reported Time consuming queries #9
Day Hour Count Duration Avg duration Jan 05 11 38 0ms 0ms 10 0ms 38 0ms 0ms 0ms set client_encoding to ?;Times Reported Time consuming queries #10
Day Hour Count Duration Avg duration Jan 05 11 38 0ms 0ms 11 0ms 18 0ms 0ms 0ms select cast(count(*) / cast(setting as numeric) * ? as int) from pg_stat_activity, pg_settings where name = ? group by setting;Times Reported Time consuming queries #11
Day Hour Count Duration Avg duration Jan 05 11 18 0ms 0ms 12 0ms 1 0ms 0ms 0ms select distinct "public"."processes"."enabled" AS "enabled" from "public"."processes" left outer join "public"."brokers" "LT?" on "LT?"."id" = "public"."processes"."brokerid" left outer join "public"."contenttypes" "LT?" on "LT?"."id" = "public"."processes"."contenttypeid" where "public"."processes"."id" = ? and "public"."processes"."id" = ? order by ? asc;Times Reported Time consuming queries #12
Day Hour Count Duration Avg duration Jan 05 11 1 0ms 0ms 13 0ms 2 0ms 0ms 0ms select count(*) from ( select max(lastupdated) from sa_hist_bigmove where symbolid in ( select symbolid from t60 group by symbolid order by max(sastdatetimereceived) desc limit ?) group by symbolid having max(lastupdated) > current_timestamp - interval ? order by max(lastupdated) desc limit ?) as k;Times Reported Time consuming queries #13
Day Hour Count Duration Avg duration Jan 05 11 2 0ms 0ms 14 0ms 374 0ms 0ms 0ms commit;Times Reported Time consuming queries #14
Day Hour Count Duration Avg duration Jan 05 11 374 0ms 0ms 15 0ms 397 0ms 0ms 0ms with rar_max as ( select resultuid from relevance_keylevels_results order by resultuid desc limit ? ), kr as ( select a.*, rr.age, rr.relevant from keylevels_results a left outer join relevance_keylevels_results rr on a.resultuid = rr.resultuid where case when false = ? then true else a.resultuid > ( select min(resultuid) from relevance_keylevels_results) end ), all_results as ( select kr.resultuid as resultuid, kr.direction as direction, s.exchange as exchange, s.symbolid as symbolid, coalesce(bim.code, s.symbol) as symbol_code, s.longname as symbol_name, s.timegranularity as interval, p.patternname as pattern_name, kr.breakout as breakout, kr.atbaridentified as identified, dtt.timezone as timezone, kr.patternlengthbars as length, g.basegroupname, newlevels.filtered, case when kr.age is not null then kr.age when kr.resultuid <= rm.resultuid then ? else ? end as age, case when kr.relevant is not null then kr.relevant when kr.resultuid <= rm.resultuid then ? else ? end as relevant, cps.pip from kr inner join brokersymbollist bsl on bsl.brokerid = ? and bsl.symbolid = kr.symbolid inner join symbols s on bsl.symbolid = s.symbolid and s.nonliquid = ? inner join symbolgroup sg on s.symbolid = sg.symbolid inner join groups g on sg.groupid = g.groupid inner join brokergroups bg on g.groupid = bg.groupid and bsl.brokerid = bg.brokerid inner join hrspatterns p on kr.patternid = p.patternid inner join downloadersymbolsettings dss on s.symbolid = dss.symbolid inner join datafeedstimetable dtt on dss.classname = dtt.classname and dtt.dayofweek = ? inner join rar_max rm on ? = ? left outer join autochartist_symbolupdates au on dss.symbolid = au.symbolid left outer join relevance_keylevels_results rar on rar.resultuid = kr.resultuid left join lateral calc_kl_signal_filter (kr.resultuid) newlevels on true left join currencypips cps on cps.symbol = s.symbol left outer join brokerinstrumentmap bim on dss.datafeedinstrumentid = bim.datafeedinstrumentid and bim.brokerid = bsl.brokerid and bim.type = ? where kr.gmttimefound > now() - interval ? and dss.enabled = ? and s.deleted = ? and (kr.simulation = ? or kr.simulation is null) and (? = ? or s.timegranularity in (...)) and (? = ? or s.exchange in (...)) and (? = ? or coalesce(bim.code, s.symbol) in (...)) and (? = ? or p.patternname in (...)) and (? = ? or kr.patternclassid in (...)) and (? = ? or kr.patternlengthbars <= ?) and kr.patternstarttime::timestamp without time zone >= coalesce(au.earliestpricedatetime, ?::timestamp without time zone) -- to make sure patternstarttime is in our t-tables ), results as ( select distinct on (symbolid) * from all_results where (false = ? or relevant = ?) and (? = ? or age <= ?) order by symbolid, resultuid ) select * from results order by identified desc, length desc limit ?;Times Reported Time consuming queries #15
Day Hour Count Duration Avg duration Jan 05 11 397 0ms 0ms 16 0ms 1 0ms 0ms 0ms select "public"."processes"."id" AS "id", "public"."processes"."locale" AS "locale", "public"."processes"."region" AS "region", "public"."processes"."schedule" AS "schedule", "public"."processes"."enabled" AS "enabled", "public"."processes"."live" AS "live", "public"."processes"."lastmodified" AS "lastmodified", "public"."processes"."lastrun" AS "lastrun", "public"."processes"."contenttypeid" AS "contenttypeid", "public"."processes"."brokerid" AS "brokerid", "public"."processes"."uuid" AS "uuid", "LT?"."name" AS "LA?", "LT?"."name" AS "LA?" from "public"."processes" left outer join "public"."brokers" "LT?" on "LT?"."id" = "public"."processes"."brokerid" left outer join "public"."contenttypes" "LT?" on "LT?"."id" = "public"."processes"."contenttypeid" where "public"."processes"."id" = ? and "public"."processes"."id" = ? and (brokerid = ?) order by "public"."processes"."id" asc limit ? offset ?;Times Reported Time consuming queries #16
Day Hour Count Duration Avg duration Jan 05 11 1 0ms 0ms 17 0ms 239 0ms 0ms 0ms select count(*), sum(size), extract(epoch from now() - min(modification)) from pg_ls_waldir ();Times Reported Time consuming queries #17
Day Hour Count Duration Avg duration Jan 05 11 239 0ms 0ms 18 0ms 239 0ms 0ms 0ms select system_identifier from pg_control_system ();Times Reported Time consuming queries #18
Day Hour Count Duration Avg duration Jan 05 11 239 0ms 0ms 19 0ms 2 0ms 0ms 0ms select groupid, exchange, groupname, symbol, longname from prfsymboltree where brokerid = ? order by groupname, symbol;Times Reported Time consuming queries #19
Day Hour Count Duration Avg duration Jan 05 11 2 0ms 0ms 20 0ms 1 0ms 0ms 0ms insert into t15 (symbolid, pricedatetime, open, high, low, close, volume, bsf, sastdatetimereceived) values (?, ?::timestamp without time zone, ?, ?.?, ?.?, ?.?, ?, ?, ?::timestamp without time zone) on conflict (symbolid, pricedatetime) do nothing;Times Reported Time consuming queries #20
Day Hour Count Duration Avg duration Jan 05 11 1 0ms 0ms Most frequent queries (N)
Rank Times executed Total duration Min duration Max duration Avg duration Query 1 21,979 0ms 0ms 0ms 0ms select ?;Times Reported Time consuming queries #1
Day Hour Count Duration Avg duration Jan 05 11 21,979 0ms 0ms 2 12,988 0ms 0ms 0ms 0ms select distinct on (coalesce(bim.code, s.symbol) , s.exchange, s.timegranularity, df.timezone) s.symbolid as id, coalesce(bim.code, s.symbol) as name, s.symbol as symbol, dss.downloadersymbol as ticker, s.exchange as exchange, s.timegranularity as interval, df.timezone as timezone from symbols s inner join downloadersymbolsettings dss on dss.symbolid = s.symbolid inner join datafeedstimetable df on df.classname ilike dss.classname left join brokersymbollist bsl on bsl.brokerid = ? and bsl.symbolid = s.symbolid left outer join brokerinstrumentmap bim on dss.datafeedinstrumentid = bim.datafeedinstrumentid and bim.brokerid = ? and bim.type = ? where s.symbolid = ?;Times Reported Time consuming queries #2
Day Hour Count Duration Avg duration Jan 05 11 12,988 0ms 0ms 3 9,877 0ms 0ms 0ms 0ms select s.symbolid as id, s.symbol as name, s.exchange as exchange, s.timegranularity as interval, dtt.timezone as timezone from symbols s inner join downloadersymbolsettings dss on dss.symbolid = s.symbolid inner join datafeedstimetable dtt on dss.classname = dtt.classname and dtt.dayofweek = ? inner join brokersymbollist bsl on bsl.symbolid = s.symbolid where bsl.brokerid = ? and (? = ? or s.timegranularity = ?) and (s.symbol = ? or dss.downloadersymbol = ?) and dss.enabled = ?;Times Reported Time consuming queries #3
Day Hour Count Duration Avg duration Jan 05 11 9,877 0ms 0ms 4 8,202 0ms 0ms 0ms 0ms insert into executionlogs (executionid, status, message, details, detailtype) values (null, ?, ?, null, null);Times Reported Time consuming queries #4
Day Hour Count Duration Avg duration Jan 05 11 8,202 0ms 0ms 5 6,676 0ms 0ms 0ms 0ms insert into t15 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) values (?, ?, ?, ?, ?, ?, ?, ?, ?, ?) on conflict (pricedatetime, symbolid) do update set open = ?, high = ?, low = ?, close = ?, volume = ?, bsf = ?, sastdatetimewritten = ?, sastdatetimereceived = ?;Times Reported Time consuming queries #5
Day Hour Count Duration Avg duration Jan 05 11 6,676 0ms 0ms 6 5,472 0ms 0ms 0ms 0ms select * from status_perbroker;Times Reported Time consuming queries #6
Day Hour Count Duration Avg duration Jan 05 11 5,472 0ms 0ms 7 5,322 0ms 0ms 0ms 0ms insert into autochartist_results (resultid, symbolid, bandwidth, pattern, qtytp, gmttimefound, direction, initialtrend, breakout, volumeincrease, noise, symmetry, predictionpricefrom, predictionpriceto, predictiontimefrom, predictiontimeto, patternstarttime, patternendtime, patternstartprice, patternendprice, resx0, resx1, supportx0, supportx1, resy0, resy1, supporty0, supporty1, supportgradient, resgradient, riskreward, patternquality, trendchange, maxmovementafterbreakout, latestbaratbreakouttime, latestbaratbreakoutprice, patternlengthbars, temporarypattern, relevancestartdistance, simulation, writtendatetime) values (?, ?, ?.?, ?, ?, ?::timestamp without time zone, ?, ?.?, ?.?, ?.?, ?.?, ?.?, ?.?, ?.?, ?::timestamp without time zone, ?::timestamp without time zone, ?::timestamp without time zone, ?::timestamp without time zone, ?.?, ?.?, ?::timestamp without time zone, ?::timestamp without time zone, ?::timestamp without time zone, ?::timestamp without time zone, ?.?, ?.?, ?.?, ?.?, ?.?, ?.?, ?.?, ?.?, ?, ?.?, ?::timestamp without time zone, ?.?, ?, ?, ?.?, ?, current_timestamp::timestamp without time zone) on conflict do nothing;Times Reported Time consuming queries #7
Day Hour Count Duration Avg duration Jan 05 11 5,322 0ms 0ms 8 3,737 0ms 0ms 0ms 0ms insert into t30 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) values (?, ?, ?, ?, ?, ?, ?, ?, ?, ?) on conflict (pricedatetime, symbolid) do update set open = ?, high = ?, low = ?, close = ?, volume = ?, bsf = ?, sastdatetimewritten = ?, sastdatetimereceived = ?;Times Reported Time consuming queries #8
Day Hour Count Duration Avg duration Jan 05 11 3,737 0ms 0ms 9 2,722 0ms 0ms 0ms 0ms update patternresultsrelevance set relevant = ?, saxo_relevant = ?, notrelevantpricedatetime = ?, reason = ? where uniqueindex = ? and relevant = ?;Times Reported Time consuming queries #9
Day Hour Count Duration Avg duration Jan 05 11 2,722 0ms 0ms 10 2,647 0ms 0ms 0ms 0ms insert into fibonacci_results (bandwidth, pattern, gmttimefound, direction, patternstarttime, patternendtime, patternstartprice, patternendprice, qtytp, pricex, timex, pricea, timea, priceb, timeb, pricec, timec, priced, timed, averagequality, timequality, errormargin, patternlengthbars, target10, target06, target16, target07, target12, target05, target03, symbolid, noise, ratiosfound, temporarypattern, uniqueindex, completed, simulation, writtendatetime) values (?.?, ?, ?::timestamp without time zone, ?, ?::timestamp without time zone, ?::timestamp without time zone, ?.?, ?.?, ?, ?.?, ?::timestamp without time zone, ?.?, ?::timestamp without time zone, ?.?, ?::timestamp without time zone, ?.?, ?::timestamp without time zone, ?.?, ?::timestamp without time zone, ?.?, ?.?, ?.?, ?, ?.?, ?.?, ?.?, ?.?, ?.?, ?.?, ?.?, ?, ?.?, ?, ?, ?, ?, ?, current_timestamp::timestamp without time zone) on conflict do nothing;Times Reported Time consuming queries #10
Day Hour Count Duration Avg duration Jan 05 11 2,647 0ms 0ms 11 2,390 0ms 0ms 0ms 0ms insert into keylevels_results (bandwidth, breakout, patternid, gmttimefound, approachingtimestamp, approachingregion, qtytp, patternlengthbars, patternprice, x0, x1, x2, x3, x4, x5, x6, x7, x8, x9, breakoutbars, breakoutprice, patternendtime, atbaridentified, atpriceidentified, errormargin, direction, symbolid, predictionpricefrom, predictionpriceto, predictiontimefrom, predictiontimebars, uniquepointsvalue, furthestprice, relevancestartdistance, patternclassid, patternstarttime, stoplosslevel, simulation, writtendatetime) values (?.?, ?, ?, ?::timestamp without time zone, ?, ?.?, ?, ?, ?.?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?.?, ?::timestamp without time zone, ?, ?.?, ?.?, ?, ?, ?.?, ?.?, ?::timestamp without time zone, ?, ?, ?.?, ?.?, ?, ?, ?.?, ?, current_timestamp::timestamp without time zone) on conflict do nothing;Times Reported Time consuming queries #11
Day Hour Count Duration Avg duration Jan 05 11 2,390 0ms 0ms 12 2,387 0ms 0ms 0ms 0ms insert into t60 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) values (?, ?, ?, ?, ?, ?, ?, ?, ?, ?) on conflict (pricedatetime, symbolid) do update set open = ?, high = ?, low = ?, close = ?, volume = ?, bsf = ?, sastdatetimewritten = ?, sastdatetimereceived = ?;Times Reported Time consuming queries #12
Day Hour Count Duration Avg duration Jan 05 11 2,387 0ms 0ms -
INSERT INTO T60 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;
Date: 2026-01-05 11:06:16 Duration: 0ms
13 1,561 0ms 0ms 0ms 0ms set extra_float_digits = ?;Times Reported Time consuming queries #13
Day Hour Count Duration Avg duration Jan 05 11 1,561 0ms 0ms 14 1,540 0ms 0ms 0ms 0ms select ew.processid, "Errors", "Warnings" from quantity_errors_warnings_perprocess ew;Times Reported Time consuming queries #14
Day Hour Count Duration Avg duration Jan 05 11 1,540 0ms 0ms 15 1,534 0ms 0ms 0ms 0ms set application_name = ?;Times Reported Time consuming queries #15
Day Hour Count Duration Avg duration Jan 05 11 1,534 0ms 0ms 16 1,495 0ms 0ms 0ms 0ms with rar_max as ( select resultuid from relevance_keylevels_results order by resultuid desc limit ? ) select case when a.old_resultuid = ? then a.old_resultuid else a.resultuid end as ruid, s.symbolid as sid, s.symbol as sym, longname, shortname, exchange as e, timegranularity as tg, a.patternid as pid, a.direction as d, a.patternprice as pp, atbaridentified as pet, case when (x9 != ?) then x9 when (x8 != ?) then x8 when (x7 != ?) then x7 when (x6 != ?) then x6 when (x5 != ?) then x5 when (x4 != ?) then x4 when (x3 != ?) then x3 when (x2 != ?) then x2 end as pst, patternprice as patp, x0, x1, x2, case when (x3 != ?) then x3 else ? end as x3, case when (x4 != ?) then x4 else ? end as x4, case when (x5 != ?) then x5 else ? end as x5, case when (x6 != ?) then x6 else ? end as x6, case when (x7 != ?) then x7 else ? end as x7, case when (x8 != ?) then x8 else ? end as x8, errormargin as erm, breakoutprice as pe, breakoutbars as be, breakout, atbaridentified as atbar, atpriceidentified as atprice, patternlengthbars as l, bandwidth as bw, qtytp as qtp, p.patternname as patternname, dtt.absolutetimezoneoffset as tzos, dtt.timezone as timezone, approachingtimestamp as apt, approachingregion as apr, predictionpricefrom as ppf, predictionpriceto as ppt, predictiontimefrom as ptf, predictiontimebars as ptb, furthestprice as fp, newlevels.filtered, a.uniquepointsvalue as upv, case when rar.age is not null then rar.age when a.resultuid <= rm.resultuid then ? else ? end as age, case when rar.relevant is not null then rar.relevant when a.resultuid <= rm.resultuid then ? else ? end as relevant, cps.pip from keylevels_results a inner join downloadersymbolsettings dss on a.symbolid = dss.symbolid inner join datafeedstimetable dtt on dss.classname = dtt.classname inner join symbols s on a.symbolid = s.symbolid inner join hrspatterns p on a.patternid = p.patternid inner join rar_max rm on ? = ? left outer join relevance_keylevels_results rar on a.resultuid = rar.resultuid left join lateral calc_kl_signal_filter (a.resultuid) newlevels on true left join currencypips cps on cps.symbol = s.symbol where (a.old_resultuid = ? or a.resultuid = ?) and dtt.dayofweek = ?;Times Reported Time consuming queries #16
Day Hour Count Duration Avg duration Jan 05 11 1,495 0ms 0ms 17 1,435 0ms 0ms 0ms 0ms with rar_max as ( select resultuid from relevance_autochartist_results order by resultuid desc limit ? ) select a.symbolid, pattern, patternid, resy0, resy1, resx0, resx1, supporty0, supporty1, supportx0, supportx1, predictiontimeto, patternstarttime, timegranularity, patternendtime, direction, trendchange, patternlengthbars, patternquality, case when a.old_resultuid = ? then a.old_resultuid else a.resultuid end as uid, breakout, initialtrend, volumeincrease, symmetry as uniformity, predictionpricefrom, predictionpriceto, noise, s.exchange, s.symbol, s.longname, s.shortname, breakout, dtt.timezone, patternstartprice, patternendprice, qtytp, newlevels.profit, newlevels.stop, newlevels.filtered, case when rar.age is not null then rar.age when a.resultuid <= rm.resultuid then ? else ? end as age, case when rar.relevant is not null then rar.relevant when a.resultuid <= rm.resultuid then ? else ? end as relevant, cps.pip from autochartist_results a inner join downloadersymbolsettings dss on a.symbolid = dss.symbolid inner join datafeedstimetable dtt on dss.classname = dtt.classname inner join symbols s on a.symbolid = s.symbolid inner join patterns p on p.patternname = a.pattern inner join rar_max rm on ? = ? left outer join relevance_autochartist_results rar on rar.resultuid = a.resultuid left join lateral calc_cp_signal (a.resultuid) newlevels on true left join currencypips cps on cps.symbol = s.symbol where (a.old_resultuid = ? or a.resultuid = ?) and dtt.dayofweek = ?;Times Reported Time consuming queries #17
Day Hour Count Duration Avg duration Jan 05 11 1,435 0ms 0ms 18 1,048 0ms 0ms 0ms 0ms select symbolid, pricedatetime, classname, downloadfrequency, downloadersymbol, open, high, low, close, volume, bsf, sastdatetimereceived from ( select pricedatetime, dss.classname, dss.downloadfrequency, dss.symbolid, dss.downloadersymbol, t.open, t.high, t.low, t.close, t.volume, t.bsf, t.sastdatetimereceived, row_number() over (partition by t.symbolid order by t.pricedatetime desc) as rn from t15 t, downloadersymbolsettings dss, symbols s where dss.classname = ? and dss.downloadfrequency = ? and dss.symbolid = t.symbolid and s.symbolid = dss.symbolid and dss.enabled = ? and s.deleted = ? and dss.downloadersymbol in (...) and t.pricedatetime > now() - interval ?) as ranked_candles_table where rn = ?;Times Reported Time consuming queries #18
Day Hour Count Duration Avg duration Jan 05 11 1,048 0ms 0ms 19 538 0ms 0ms 0ms 0ms select downloadersymbol, spike_threshold from price_datafeed_spike_threshold where classname = ?;Times Reported Time consuming queries #19
Day Hour Count Duration Avg duration Jan 05 11 538 0ms 0ms 20 538 0ms 0ms 0ms 0ms select s.symbolid, dss.downloadfrequency, dss.downloadersymbol from downloadersymbolsettings dss inner join symbols s on dss.symbolid = s.symbolid where dss.classname = ? and s.deleted = ? and dss.enabled = ?;Times Reported Time consuming queries #20
Day Hour Count Duration Avg duration Jan 05 11 538 0ms 0ms Normalized slowest queries (N)
Rank Min duration Max duration Avg duration Times executed Total duration Query 1 0ms 0ms 0ms 1 0ms insert into t30 (symbolid, pricedatetime, open, high, low, close, volume, bsf, sastdatetimereceived) values (?, ?::timestamp without time zone, ?.?, ?.?, ?.?, ?, ?, ?, ?::timestamp without time zone) on conflict (symbolid, pricedatetime) do nothing;Times Reported Time consuming queries #1
Day Hour Count Duration Avg duration Jan 05 11 1 0ms 0ms 2 0ms 0ms 0ms 40 0ms select key, value from datasources ds inner join datasourceparams dsp on ds.id = dsp.datasourceid where ds.name = ?;Times Reported Time consuming queries #2
Day Hour Count Duration Avg duration Jan 05 11 40 0ms 0ms 3 0ms 0ms 0ms 1 0ms select "public"."processparameters"."id" AS "id", "public"."processparameters"."processid" AS "processid", "public"."processparameters"."key" AS "key", "public"."processparameters"."value" AS "value" from "public"."processparameters" where "public"."processparameters"."id" = ? and "public"."processparameters"."id" = ? limit ? offset ?;Times Reported Time consuming queries #3
Day Hour Count Duration Avg duration Jan 05 11 1 0ms 0ms 4 0ms 0ms 0ms 1 0ms select distinct "public"."processes"."live" AS "live" from "public"."processes" left outer join "public"."brokers" "LT?" on "LT?"."id" = "public"."processes"."brokerid" left outer join "public"."contenttypes" "LT?" on "LT?"."id" = "public"."processes"."contenttypeid" where "public"."processes"."id" = ? and "public"."processes"."id" = ? order by ? asc;Times Reported Time consuming queries #4
Day Hour Count Duration Avg duration Jan 05 11 1 0ms 0ms 5 0ms 0ms 0ms 10 0ms with rar_max as ( select resultuid from relevance_bigmovement_results order by resultuid desc limit ? ) select bmr.symbolid, patternstarttime, patternendtime, timegranularity, ? as direction, case when bmr.old_resultuid = ? then bmr.old_resultuid else bmr.resultuid end as uid, s.exchange, s.symbol, s.longname, s.shortname, dtt.timezone, bmr.patternmovement, bmr.statisticalmovement, bmr.fromprice, bmr.toprice, bmr.percentile, bmr.patternlengthbars, case when rbr.age is not null then rbr.age when bmr.resultuid <= rm.resultuid then ? else ? end as age, case when rbr.relevant is not null then rbr.relevant when bmr.resultuid <= rm.resultuid then ? else ? end as relevant, cps.pip from bigmovement_results bmr inner join downloadersymbolsettings dss on bmr.symbolid = dss.symbolid inner join datafeedstimetable dtt on dss.classname = dtt.classname inner join symbols s on bmr.symbolid = s.symbolid inner join rar_max rm on ? = ? left outer join relevance_bigmovement_results rbr on rbr.resultuid = bmr.resultuid left join currencypips cps on cps.symbol = s.symbol where (bmr.old_resultuid = ? or bmr.resultuid = ?) and dtt.dayofweek = ?;Times Reported Time consuming queries #5
Day Hour Count Duration Avg duration Jan 05 11 10 0ms 0ms 6 0ms 0ms 0ms 2,387 0ms insert into t60 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) values (?, ?, ?, ?, ?, ?, ?, ?, ?, ?) on conflict (pricedatetime, symbolid) do update set open = ?, high = ?, low = ?, close = ?, volume = ?, bsf = ?, sastdatetimewritten = ?, sastdatetimereceived = ?;Times Reported Time consuming queries #6
Day Hour Count Duration Avg duration Jan 05 11 2,387 0ms 0ms -
INSERT INTO T60 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;
Date: 2026-01-05 11:06:16 Duration: 0ms
7 0ms 0ms 0ms 48 0ms select count(*) from datafeeds_latestrun where feedname ilike ? and ((latestrxtime > current_timestamp - interval ? and latestdbwritetime > current_timestamp - interval ?) or (latestdbwritetime > current_timestamp - interval ? and lateststartuptime > current_timestamp - interval ?));Times Reported Time consuming queries #7
Day Hour Count Duration Avg duration Jan 05 11 48 0ms 0ms 8 0ms 0ms 0ms 4 0ms select updaterelevantforrelevantresults ();Times Reported Time consuming queries #8
Day Hour Count Duration Avg duration Jan 05 11 4 0ms 0ms 9 0ms 0ms 0ms 38 0ms set datestyle = iso;Times Reported Time consuming queries #9
Day Hour Count Duration Avg duration Jan 05 11 38 0ms 0ms 10 0ms 0ms 0ms 38 0ms set client_encoding to ?;Times Reported Time consuming queries #10
Day Hour Count Duration Avg duration Jan 05 11 38 0ms 0ms 11 0ms 0ms 0ms 18 0ms select cast(count(*) / cast(setting as numeric) * ? as int) from pg_stat_activity, pg_settings where name = ? group by setting;Times Reported Time consuming queries #11
Day Hour Count Duration Avg duration Jan 05 11 18 0ms 0ms 12 0ms 0ms 0ms 1 0ms select distinct "public"."processes"."enabled" AS "enabled" from "public"."processes" left outer join "public"."brokers" "LT?" on "LT?"."id" = "public"."processes"."brokerid" left outer join "public"."contenttypes" "LT?" on "LT?"."id" = "public"."processes"."contenttypeid" where "public"."processes"."id" = ? and "public"."processes"."id" = ? order by ? asc;Times Reported Time consuming queries #12
Day Hour Count Duration Avg duration Jan 05 11 1 0ms 0ms 13 0ms 0ms 0ms 2 0ms select count(*) from ( select max(lastupdated) from sa_hist_bigmove where symbolid in ( select symbolid from t60 group by symbolid order by max(sastdatetimereceived) desc limit ?) group by symbolid having max(lastupdated) > current_timestamp - interval ? order by max(lastupdated) desc limit ?) as k;Times Reported Time consuming queries #13
Day Hour Count Duration Avg duration Jan 05 11 2 0ms 0ms 14 0ms 0ms 0ms 374 0ms commit;Times Reported Time consuming queries #14
Day Hour Count Duration Avg duration Jan 05 11 374 0ms 0ms 15 0ms 0ms 0ms 397 0ms with rar_max as ( select resultuid from relevance_keylevels_results order by resultuid desc limit ? ), kr as ( select a.*, rr.age, rr.relevant from keylevels_results a left outer join relevance_keylevels_results rr on a.resultuid = rr.resultuid where case when false = ? then true else a.resultuid > ( select min(resultuid) from relevance_keylevels_results) end ), all_results as ( select kr.resultuid as resultuid, kr.direction as direction, s.exchange as exchange, s.symbolid as symbolid, coalesce(bim.code, s.symbol) as symbol_code, s.longname as symbol_name, s.timegranularity as interval, p.patternname as pattern_name, kr.breakout as breakout, kr.atbaridentified as identified, dtt.timezone as timezone, kr.patternlengthbars as length, g.basegroupname, newlevels.filtered, case when kr.age is not null then kr.age when kr.resultuid <= rm.resultuid then ? else ? end as age, case when kr.relevant is not null then kr.relevant when kr.resultuid <= rm.resultuid then ? else ? end as relevant, cps.pip from kr inner join brokersymbollist bsl on bsl.brokerid = ? and bsl.symbolid = kr.symbolid inner join symbols s on bsl.symbolid = s.symbolid and s.nonliquid = ? inner join symbolgroup sg on s.symbolid = sg.symbolid inner join groups g on sg.groupid = g.groupid inner join brokergroups bg on g.groupid = bg.groupid and bsl.brokerid = bg.brokerid inner join hrspatterns p on kr.patternid = p.patternid inner join downloadersymbolsettings dss on s.symbolid = dss.symbolid inner join datafeedstimetable dtt on dss.classname = dtt.classname and dtt.dayofweek = ? inner join rar_max rm on ? = ? left outer join autochartist_symbolupdates au on dss.symbolid = au.symbolid left outer join relevance_keylevels_results rar on rar.resultuid = kr.resultuid left join lateral calc_kl_signal_filter (kr.resultuid) newlevels on true left join currencypips cps on cps.symbol = s.symbol left outer join brokerinstrumentmap bim on dss.datafeedinstrumentid = bim.datafeedinstrumentid and bim.brokerid = bsl.brokerid and bim.type = ? where kr.gmttimefound > now() - interval ? and dss.enabled = ? and s.deleted = ? and (kr.simulation = ? or kr.simulation is null) and (? = ? or s.timegranularity in (...)) and (? = ? or s.exchange in (...)) and (? = ? or coalesce(bim.code, s.symbol) in (...)) and (? = ? or p.patternname in (...)) and (? = ? or kr.patternclassid in (...)) and (? = ? or kr.patternlengthbars <= ?) and kr.patternstarttime::timestamp without time zone >= coalesce(au.earliestpricedatetime, ?::timestamp without time zone) -- to make sure patternstarttime is in our t-tables ), results as ( select distinct on (symbolid) * from all_results where (false = ? or relevant = ?) and (? = ? or age <= ?) order by symbolid, resultuid ) select * from results order by identified desc, length desc limit ?;Times Reported Time consuming queries #15
Day Hour Count Duration Avg duration Jan 05 11 397 0ms 0ms 16 0ms 0ms 0ms 1 0ms select "public"."processes"."id" AS "id", "public"."processes"."locale" AS "locale", "public"."processes"."region" AS "region", "public"."processes"."schedule" AS "schedule", "public"."processes"."enabled" AS "enabled", "public"."processes"."live" AS "live", "public"."processes"."lastmodified" AS "lastmodified", "public"."processes"."lastrun" AS "lastrun", "public"."processes"."contenttypeid" AS "contenttypeid", "public"."processes"."brokerid" AS "brokerid", "public"."processes"."uuid" AS "uuid", "LT?"."name" AS "LA?", "LT?"."name" AS "LA?" from "public"."processes" left outer join "public"."brokers" "LT?" on "LT?"."id" = "public"."processes"."brokerid" left outer join "public"."contenttypes" "LT?" on "LT?"."id" = "public"."processes"."contenttypeid" where "public"."processes"."id" = ? and "public"."processes"."id" = ? and (brokerid = ?) order by "public"."processes"."id" asc limit ? offset ?;Times Reported Time consuming queries #16
Day Hour Count Duration Avg duration Jan 05 11 1 0ms 0ms 17 0ms 0ms 0ms 239 0ms select count(*), sum(size), extract(epoch from now() - min(modification)) from pg_ls_waldir ();Times Reported Time consuming queries #17
Day Hour Count Duration Avg duration Jan 05 11 239 0ms 0ms 18 0ms 0ms 0ms 239 0ms select system_identifier from pg_control_system ();Times Reported Time consuming queries #18
Day Hour Count Duration Avg duration Jan 05 11 239 0ms 0ms 19 0ms 0ms 0ms 2 0ms select groupid, exchange, groupname, symbol, longname from prfsymboltree where brokerid = ? order by groupname, symbol;Times Reported Time consuming queries #19
Day Hour Count Duration Avg duration Jan 05 11 2 0ms 0ms 20 0ms 0ms 0ms 1 0ms insert into t15 (symbolid, pricedatetime, open, high, low, close, volume, bsf, sastdatetimereceived) values (?, ?::timestamp without time zone, ?, ?.?, ?.?, ?.?, ?, ?, ?::timestamp without time zone) on conflict (symbolid, pricedatetime) do nothing;Times Reported Time consuming queries #20
Day Hour Count Duration Avg duration Jan 05 11 1 0ms 0ms Time consuming prepare
Rank Total duration Times executed Min duration Max duration Avg duration Query 1 1s542ms 2,136 0ms 10ms 0ms WITH rar_max as ( ;Times Reported Time consuming prepare #1
Day Hour Count Duration Avg duration Jan 05 11 2,136 1s542ms 0ms -
WITH rar_max as ( ;
Date: 2026-01-05 11:13:01 Duration: 10ms Database: postgres
-
WITH rar_max as ( ;
Date: 2026-01-05 11:32:19 Duration: 7ms Database: postgres
-
WITH rar_max as ( ;
Date: 2026-01-05 11:17:15 Duration: 6ms Database: postgres
2 1s402ms 1,172 0ms 2ms 1ms SELECT symbolid, ;Times Reported Time consuming prepare #2
Day Hour Count Duration Avg duration 11 1,172 1s402ms 1ms -
SELECT symbolid, ;
Date: 2026-01-05 11:47:14 Duration: 2ms Database: postgres
-
SELECT symbolid, ;
Date: 2026-01-05 11:30:52 Duration: 2ms Database: postgres
-
SELECT symbolid, ;
Date: 2026-01-05 11:00:55 Duration: 2ms Database: postgres
3 618ms 2,522 0ms 3ms 0ms SELECT ;Times Reported Time consuming prepare #3
Day Hour Count Duration Avg duration 11 2,522 618ms 0ms -
SELECT ;
Date: 2026-01-05 11:35:25 Duration: 3ms Database: postgres
-
SELECT ;
Date: 2026-01-05 11:32:19 Duration: 3ms Database: postgres
-
SELECT ;
Date: 2026-01-05 11:32:19 Duration: 3ms Database: postgres
4 533ms 538 0ms 1ms 0ms SELECT s.symbolid, dss.downloadfrequency, dss.downloadersymbol;Times Reported Time consuming prepare #4
Day Hour Count Duration Avg duration 11 538 533ms 0ms -
SELECT s.symbolid, dss.downloadfrequency, dss.downloadersymbol;
Date: 2026-01-05 11:00:42 Duration: 1ms Database: postgres
-
SELECT s.symbolid, dss.downloadfrequency, dss.downloadersymbol;
Date: 2026-01-05 11:45:35 Duration: 1ms Database: postgres
-
SELECT s.symbolid, dss.downloadfrequency, dss.downloadersymbol;
Date: 2026-01-05 11:45:45 Duration: 1ms Database: postgres
5 272ms 3,179 0ms 0ms 0ms INSERT INTO T30 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;Times Reported Time consuming prepare #5
Day Hour Count Duration Avg duration 11 3,179 272ms 0ms -
INSERT INTO T30 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;
Date: 2026-01-05 11:11:55 Duration: 0ms Database: postgres
-
INSERT INTO T30 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;
Date: 2026-01-05 11:11:39 Duration: 0ms Database: postgres
-
INSERT INTO T30 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;
Date: 2026-01-05 11:41:40 Duration: 0ms Database: postgres
6 212ms 2,062 0ms 0ms 0ms INSERT INTO T60 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;Times Reported Time consuming prepare #6
Day Hour Count Duration Avg duration 11 2,062 212ms 0ms -
INSERT INTO T60 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;
Date: 2026-01-05 11:11:33 Duration: 0ms Database: postgres
-
INSERT INTO T60 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;
Date: 2026-01-05 11:10:55 Duration: 0ms Database: postgres
-
INSERT INTO T60 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;
Date: 2026-01-05 11:11:27 Duration: 0ms Database: postgres
7 196ms 1,561 0ms 4ms 0ms SET extra_float_digits = 3;Times Reported Time consuming prepare #7
Day Hour Count Duration Avg duration 11 1,561 196ms 0ms -
SET extra_float_digits = 3;
Date: 2026-01-05 11:01:32 Duration: 4ms Database: postgres
-
SET extra_float_digits = 3;
Date: 2026-01-05 11:32:19 Duration: 0ms Database: postgres
-
SET extra_float_digits = 3;
Date: 2026-01-05 11:17:15 Duration: 0ms Database: postgres
8 161ms 1,070 0ms 0ms 0ms INSERT INTO T15 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;Times Reported Time consuming prepare #8
Day Hour Count Duration Avg duration 11 1,070 161ms 0ms -
INSERT INTO T15 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;
Date: 2026-01-05 11:26:40 Duration: 0ms Database: postgres
-
INSERT INTO T15 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;
Date: 2026-01-05 11:17:58 Duration: 0ms Database: postgres
-
INSERT INTO T15 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;
Date: 2026-01-05 11:56:40 Duration: 0ms Database: postgres
9 69ms 474 0ms 0ms 0ms SELECT NULL AS TABLE_CAT, n.nspname AS TABLE_SCHEM, c.relname AS TABLE_NAME, CASE n.nspname ~ '^pg_' OR n.nspname = 'information_schema' WHEN true THEN CASE WHEN n.nspname = 'pg_catalog' OR n.nspname = 'information_schema' THEN CASE c.relkind WHEN 'r' THEN 'SYSTEM TABLE' WHEN 'v' THEN 'SYSTEM VIEW' WHEN 'i' THEN 'SYSTEM INDEX' ELSE NULL END WHEN n.nspname = 'pg_toast' THEN CASE c.relkind WHEN 'r' THEN 'SYSTEM TOAST TABLE' WHEN 'i' THEN 'SYSTEM TOAST INDEX' ELSE NULL END ELSE CASE c.relkind WHEN 'r' THEN 'TEMPORARY TABLE' WHEN 'p' THEN 'TEMPORARY TABLE' WHEN 'i' THEN 'TEMPORARY INDEX' WHEN 'S' THEN 'TEMPORARY SEQUENCE' WHEN 'v' THEN 'TEMPORARY VIEW' ELSE NULL END END WHEN false THEN CASE c.relkind WHEN 'r' THEN 'TABLE' WHEN 'p' THEN 'PARTITIONED TABLE' WHEN 'i' THEN 'INDEX' WHEN 'S' THEN 'SEQUENCE' WHEN 'v' THEN 'VIEW' WHEN 'c' THEN 'TYPE' WHEN 'f' THEN 'FOREIGN TABLE' WHEN 'm' THEN 'MATERIALIZED VIEW' ELSE NULL END ELSE NULL END AS TABLE_TYPE, d.description AS REMARKS, '' as TYPE_CAT, '' as TYPE_SCHEM, '' as TYPE_NAME, '' AS SELF_REFERENCING_COL_NAME, '' AS REF_GENERATION FROM pg_catalog.pg_namespace n, pg_catalog.pg_class c LEFT JOIN pg_catalog.pg_description d ON (c.oid = d.objoid AND d.objsubid = 0) LEFT JOIN pg_catalog.pg_class dc ON (d.classoid = dc.oid AND dc.relname = 'pg_class') LEFT JOIN pg_catalog.pg_namespace dn ON (dn.oid = dc.relnamespace AND dn.nspname = 'pg_catalog') WHERE c.relnamespace = n.oid AND c.relname LIKE 'PROBABLYNOT' AND (false OR (c.relkind = 'r' AND n.nspname !~ '^pg_' AND n.nspname <> 'information_schema')) ORDER BY TABLE_TYPE, TABLE_SCHEM, TABLE_NAME;Times Reported Time consuming prepare #9
Day Hour Count Duration Avg duration 11 474 69ms 0ms -
SELECT NULL AS TABLE_CAT, n.nspname AS TABLE_SCHEM, c.relname AS TABLE_NAME, CASE n.nspname ~ '^pg_' OR n.nspname = 'information_schema' WHEN true THEN CASE WHEN n.nspname = 'pg_catalog' OR n.nspname = 'information_schema' THEN CASE c.relkind WHEN 'r' THEN 'SYSTEM TABLE' WHEN 'v' THEN 'SYSTEM VIEW' WHEN 'i' THEN 'SYSTEM INDEX' ELSE NULL END WHEN n.nspname = 'pg_toast' THEN CASE c.relkind WHEN 'r' THEN 'SYSTEM TOAST TABLE' WHEN 'i' THEN 'SYSTEM TOAST INDEX' ELSE NULL END ELSE CASE c.relkind WHEN 'r' THEN 'TEMPORARY TABLE' WHEN 'p' THEN 'TEMPORARY TABLE' WHEN 'i' THEN 'TEMPORARY INDEX' WHEN 'S' THEN 'TEMPORARY SEQUENCE' WHEN 'v' THEN 'TEMPORARY VIEW' ELSE NULL END END WHEN false THEN CASE c.relkind WHEN 'r' THEN 'TABLE' WHEN 'p' THEN 'PARTITIONED TABLE' WHEN 'i' THEN 'INDEX' WHEN 'S' THEN 'SEQUENCE' WHEN 'v' THEN 'VIEW' WHEN 'c' THEN 'TYPE' WHEN 'f' THEN 'FOREIGN TABLE' WHEN 'm' THEN 'MATERIALIZED VIEW' ELSE NULL END ELSE NULL END AS TABLE_TYPE, d.description AS REMARKS, '' as TYPE_CAT, '' as TYPE_SCHEM, '' as TYPE_NAME, '' AS SELF_REFERENCING_COL_NAME, '' AS REF_GENERATION FROM pg_catalog.pg_namespace n, pg_catalog.pg_class c LEFT JOIN pg_catalog.pg_description d ON (c.oid = d.objoid AND d.objsubid = 0) LEFT JOIN pg_catalog.pg_class dc ON (d.classoid = dc.oid AND dc.relname = 'pg_class') LEFT JOIN pg_catalog.pg_namespace dn ON (dn.oid = dc.relnamespace AND dn.nspname = 'pg_catalog') WHERE c.relnamespace = n.oid AND c.relname LIKE 'PROBABLYNOT' AND (false OR (c.relkind = 'r' AND n.nspname !~ '^pg_' AND n.nspname <> 'information_schema')) ORDER BY TABLE_TYPE, TABLE_SCHEM, TABLE_NAME;
Date: 2026-01-05 11:14:22 Duration: 0ms Database: postgres
-
SELECT NULL AS TABLE_CAT, n.nspname AS TABLE_SCHEM, c.relname AS TABLE_NAME, CASE n.nspname ~ '^pg_' OR n.nspname = 'information_schema' WHEN true THEN CASE WHEN n.nspname = 'pg_catalog' OR n.nspname = 'information_schema' THEN CASE c.relkind WHEN 'r' THEN 'SYSTEM TABLE' WHEN 'v' THEN 'SYSTEM VIEW' WHEN 'i' THEN 'SYSTEM INDEX' ELSE NULL END WHEN n.nspname = 'pg_toast' THEN CASE c.relkind WHEN 'r' THEN 'SYSTEM TOAST TABLE' WHEN 'i' THEN 'SYSTEM TOAST INDEX' ELSE NULL END ELSE CASE c.relkind WHEN 'r' THEN 'TEMPORARY TABLE' WHEN 'p' THEN 'TEMPORARY TABLE' WHEN 'i' THEN 'TEMPORARY INDEX' WHEN 'S' THEN 'TEMPORARY SEQUENCE' WHEN 'v' THEN 'TEMPORARY VIEW' ELSE NULL END END WHEN false THEN CASE c.relkind WHEN 'r' THEN 'TABLE' WHEN 'p' THEN 'PARTITIONED TABLE' WHEN 'i' THEN 'INDEX' WHEN 'S' THEN 'SEQUENCE' WHEN 'v' THEN 'VIEW' WHEN 'c' THEN 'TYPE' WHEN 'f' THEN 'FOREIGN TABLE' WHEN 'm' THEN 'MATERIALIZED VIEW' ELSE NULL END ELSE NULL END AS TABLE_TYPE, d.description AS REMARKS, '' as TYPE_CAT, '' as TYPE_SCHEM, '' as TYPE_NAME, '' AS SELF_REFERENCING_COL_NAME, '' AS REF_GENERATION FROM pg_catalog.pg_namespace n, pg_catalog.pg_class c LEFT JOIN pg_catalog.pg_description d ON (c.oid = d.objoid AND d.objsubid = 0) LEFT JOIN pg_catalog.pg_class dc ON (d.classoid = dc.oid AND dc.relname = 'pg_class') LEFT JOIN pg_catalog.pg_namespace dn ON (dn.oid = dc.relnamespace AND dn.nspname = 'pg_catalog') WHERE c.relnamespace = n.oid AND c.relname LIKE 'PROBABLYNOT' AND (false OR (c.relkind = 'r' AND n.nspname !~ '^pg_' AND n.nspname <> 'information_schema')) ORDER BY TABLE_TYPE, TABLE_SCHEM, TABLE_NAME;
Date: 2026-01-05 11:14:23 Duration: 0ms Database: postgres
-
SELECT NULL AS TABLE_CAT, n.nspname AS TABLE_SCHEM, c.relname AS TABLE_NAME, CASE n.nspname ~ '^pg_' OR n.nspname = 'information_schema' WHEN true THEN CASE WHEN n.nspname = 'pg_catalog' OR n.nspname = 'information_schema' THEN CASE c.relkind WHEN 'r' THEN 'SYSTEM TABLE' WHEN 'v' THEN 'SYSTEM VIEW' WHEN 'i' THEN 'SYSTEM INDEX' ELSE NULL END WHEN n.nspname = 'pg_toast' THEN CASE c.relkind WHEN 'r' THEN 'SYSTEM TOAST TABLE' WHEN 'i' THEN 'SYSTEM TOAST INDEX' ELSE NULL END ELSE CASE c.relkind WHEN 'r' THEN 'TEMPORARY TABLE' WHEN 'p' THEN 'TEMPORARY TABLE' WHEN 'i' THEN 'TEMPORARY INDEX' WHEN 'S' THEN 'TEMPORARY SEQUENCE' WHEN 'v' THEN 'TEMPORARY VIEW' ELSE NULL END END WHEN false THEN CASE c.relkind WHEN 'r' THEN 'TABLE' WHEN 'p' THEN 'PARTITIONED TABLE' WHEN 'i' THEN 'INDEX' WHEN 'S' THEN 'SEQUENCE' WHEN 'v' THEN 'VIEW' WHEN 'c' THEN 'TYPE' WHEN 'f' THEN 'FOREIGN TABLE' WHEN 'm' THEN 'MATERIALIZED VIEW' ELSE NULL END ELSE NULL END AS TABLE_TYPE, d.description AS REMARKS, '' as TYPE_CAT, '' as TYPE_SCHEM, '' as TYPE_NAME, '' AS SELF_REFERENCING_COL_NAME, '' AS REF_GENERATION FROM pg_catalog.pg_namespace n, pg_catalog.pg_class c LEFT JOIN pg_catalog.pg_description d ON (c.oid = d.objoid AND d.objsubid = 0) LEFT JOIN pg_catalog.pg_class dc ON (d.classoid = dc.oid AND dc.relname = 'pg_class') LEFT JOIN pg_catalog.pg_namespace dn ON (dn.oid = dc.relnamespace AND dn.nspname = 'pg_catalog') WHERE c.relnamespace = n.oid AND c.relname LIKE 'PROBABLYNOT' AND (false OR (c.relkind = 'r' AND n.nspname !~ '^pg_' AND n.nspname <> 'information_schema')) ORDER BY TABLE_TYPE, TABLE_SCHEM, TABLE_NAME;
Date: 2026-01-05 11:14:23 Duration: 0ms Database: postgres
10 64ms 12 4ms 7ms 5ms with sym_info as ( ;Times Reported Time consuming prepare #10
Day Hour Count Duration Avg duration 11 12 64ms 5ms -
with sym_info as ( ;
Date: 2026-01-05 11:36:44 Duration: 7ms Database: postgres
-
with sym_info as ( ;
Date: 2026-01-05 11:51:44 Duration: 7ms Database: postgres
-
with sym_info as ( ;
Date: 2026-01-05 11:06:46 Duration: 7ms Database: postgres
11 45ms 18 1ms 4ms 2ms select cast(count(*) / cast(setting as numeric) * 100 as int) from pg_stat_activity, pg_settings WHERE name = 'max_connections' group by setting;Times Reported Time consuming prepare #11
Day Hour Count Duration Avg duration 11 18 45ms 2ms -
select cast(count(*) / cast(setting as numeric) * 100 as int) from pg_stat_activity, pg_settings WHERE name = 'max_connections' group by setting;
Date: 2026-01-05 11:01:15 Duration: 4ms Database: postgres
-
select cast(count(*) / cast(setting as numeric) * 100 as int) from pg_stat_activity, pg_settings WHERE name = 'max_connections' group by setting;
Date: 2026-01-05 11:01:15 Duration: 2ms Database: postgres
-
select cast(count(*) / cast(setting as numeric) * 100 as int) from pg_stat_activity, pg_settings WHERE name = 'max_connections' group by setting;
Date: 2026-01-05 11:41:03 Duration: 2ms Database: postgres
12 36ms 31 0ms 3ms 1ms WITH last_candle AS ( ;Times Reported Time consuming prepare #12
Day Hour Count Duration Avg duration 11 31 36ms 1ms -
WITH last_candle AS ( ;
Date: 2026-01-05 11:04:00 Duration: 3ms Database: postgres
-
WITH last_candle AS ( ;
Date: 2026-01-05 11:32:00 Duration: 3ms Database: postgres
-
WITH last_candle AS ( ;
Date: 2026-01-05 11:32:00 Duration: 3ms Database: postgres
13 32ms 825 0ms 0ms 0ms select 1;Times Reported Time consuming prepare #13
Day Hour Count Duration Avg duration 11 825 32ms 0ms -
select 1;
Date: 2026-01-05 11:20:36 Duration: 0ms Database: postgres
-
select 1;
Date: 2026-01-05 11:00:04 Duration: 0ms Database: postgres
-
select 1;
Date: 2026-01-05 11:35:54 Duration: 0ms Database: postgres
14 17ms 1,534 0ms 0ms 0ms SET application_name = 'PostgreSQL JDBC Driver';Times Reported Time consuming prepare #14
Day Hour Count Duration Avg duration 11 1,534 17ms 0ms -
SET application_name = 'PostgreSQL JDBC Driver';
Date: 2026-01-05 11:17:15 Duration: 0ms Database: postgres
-
SET application_name = 'PostgreSQL JDBC Driver';
Date: 2026-01-05 11:51:30 Duration: 0ms Database: postgres
-
SET application_name = 'PostgreSQL JDBC Driver';
Date: 2026-01-05 11:01:47 Duration: 0ms Database: postgres
15 15ms 141 0ms 0ms 0ms INSERT INTO T240 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;Times Reported Time consuming prepare #15
Day Hour Count Duration Avg duration 11 141 15ms 0ms -
INSERT INTO T240 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;
Date: 2026-01-05 11:45:53 Duration: 0ms Database: postgres
-
INSERT INTO T240 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;
Date: 2026-01-05 11:46:09 Duration: 0ms Database: postgres
-
INSERT INTO T240 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;
Date: 2026-01-05 11:46:25 Duration: 0ms Database: postgres
16 14ms 24 0ms 0ms 0ms select count(*) from datafeed_restarter_events where is_current_entry = 1;Times Reported Time consuming prepare #16
Day Hour Count Duration Avg duration 11 24 14ms 0ms -
select count(*) from datafeed_restarter_events where is_current_entry = 1;
Date: 2026-01-05 11:00:02 Duration: 0ms Database: postgres
-
select count(*) from datafeed_restarter_events where is_current_entry = 1;
Date: 2026-01-05 11:00:02 Duration: 0ms Database: postgres
-
select count(*) from datafeed_restarter_events where is_current_entry = 1;
Date: 2026-01-05 11:20:02 Duration: 0ms Database: postgres
17 14ms 6 2ms 3ms 2ms select client_addr, count(1) from pg_stat_activity, pg_settings WHERE name = 'max_connections' group by client_addr, setting having (client_addr is not null OR (client_addr is null and count(1) > (cast(setting as numeric) / 3 * 2))) order by count desc;Times Reported Time consuming prepare #17
Day Hour Count Duration Avg duration 11 6 14ms 2ms -
select client_addr, count(1) from pg_stat_activity, pg_settings WHERE name = 'max_connections' group by client_addr, setting having (client_addr is not null OR (client_addr is null and count(1) > (cast(setting as numeric) / 3 * 2))) order by count desc;
Date: 2026-01-05 11:00:04 Duration: 3ms Database: postgres
-
select client_addr, count(1) from pg_stat_activity, pg_settings WHERE name = 'max_connections' group by client_addr, setting having (client_addr is not null OR (client_addr is null and count(1) > (cast(setting as numeric) / 3 * 2))) order by count desc;
Date: 2026-01-05 11:20:05 Duration: 2ms Database: postgres
-
select client_addr, count(1) from pg_stat_activity, pg_settings WHERE name = 'max_connections' group by client_addr, setting having (client_addr is not null OR (client_addr is null and count(1) > (cast(setting as numeric) / 3 * 2))) order by count desc;
Date: 2026-01-05 11:40:05 Duration: 2ms Database: postgres
18 13ms 6 2ms 2ms 2ms with rankedmt4 as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors ), last_feed_entry as ( select * from rankedmt4 where r = 1 ), ok_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where status = 'OK' ), earliest_entry_after_ok as ( select m.datafeedname, min(m.eventtimestamp) as eventtimestamp from mt4datafeederrors m left outer join ( select datafeedname, eventtimestamp from ok_entries where r = 1) oo on m.datafeedname = oo.datafeedname where m.eventtimestamp > coalesce(oo.eventtimestamp, '1900-01-01'::timestamp without time zone) group by m.datafeedname ), notified_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where notified is not null and notified <> '' ), broker as ( select *, row_number() over (partition by feedname order by brokerid) r from ( select distinct b.brokerid, b.name as brokername, dss.classname as feedname from downloadersymbolsettings dss inner join brokersymbollist bsl on dss.symbolid = bsl.symbolid inner join broker b on bsl.brokerid = b.brokerid where dss.enabled = 1) a ) select last.id, last.datafeedname, last.eventtimestamp, last.status, last.errordescription, last.serveraddress, last.username, note.notified, note.eventtimestamp, broker.brokername from last_feed_entry last inner join earliest_entry_after_ok after_ok on last.datafeedname = after_ok.datafeedname inner join broker on last.datafeedname = broker.feedname left outer join ok_entries ok on ok.datafeedname = last.datafeedname left outer join notified_entries note on note.datafeedname = last.datafeedname and note.r = 1 where (ok.r is null or ok.r = 1) and last.datafeedname not in ( select distinct datafeedname from last_feed_entry where status = 'OK') and extract(epoch from (last.eventtimestamp - after_ok.eventtimestamp)) > 60 * 60 and last.eventtimestamp > current_timestamp - interval '1 day' and (note.eventtimestamp is null or note.eventtimestamp < current_timestamp - interval '10 hours') and last.eventtimestamp > current_timestamp - interval '1 hour' and broker.r = 1;Times Reported Time consuming prepare #18
Day Hour Count Duration Avg duration 11 6 13ms 2ms -
with rankedmt4 as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors ), last_feed_entry as ( select * from rankedmt4 where r = 1 ), ok_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where status = 'OK' ), earliest_entry_after_ok as ( select m.datafeedname, min(m.eventtimestamp) as eventtimestamp from mt4datafeederrors m left outer join ( select datafeedname, eventtimestamp from ok_entries where r = 1) oo on m.datafeedname = oo.datafeedname where m.eventtimestamp > coalesce(oo.eventtimestamp, '1900-01-01'::timestamp without time zone) group by m.datafeedname ), notified_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where notified is not null and notified <> '' ), broker as ( select *, row_number() over (partition by feedname order by brokerid) r from ( select distinct b.brokerid, b.name as brokername, dss.classname as feedname from downloadersymbolsettings dss inner join brokersymbollist bsl on dss.symbolid = bsl.symbolid inner join broker b on bsl.brokerid = b.brokerid where dss.enabled = 1) a ) select last.id, last.datafeedname, last.eventtimestamp, last.status, last.errordescription, last.serveraddress, last.username, note.notified, note.eventtimestamp, broker.brokername from last_feed_entry last inner join earliest_entry_after_ok after_ok on last.datafeedname = after_ok.datafeedname inner join broker on last.datafeedname = broker.feedname left outer join ok_entries ok on ok.datafeedname = last.datafeedname left outer join notified_entries note on note.datafeedname = last.datafeedname and note.r = 1 where (ok.r is null or ok.r = 1) and last.datafeedname not in ( select distinct datafeedname from last_feed_entry where status = 'OK') and extract(epoch from (last.eventtimestamp - after_ok.eventtimestamp)) > 60 * 60 and last.eventtimestamp > current_timestamp - interval '1 day' and (note.eventtimestamp is null or note.eventtimestamp < current_timestamp - interval '10 hours') and last.eventtimestamp > current_timestamp - interval '1 hour' and broker.r = 1;
Date: 2026-01-05 11:10:02 Duration: 2ms Database: postgres
-
with rankedmt4 as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors ), last_feed_entry as ( select * from rankedmt4 where r = 1 ), ok_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where status = 'OK' ), earliest_entry_after_ok as ( select m.datafeedname, min(m.eventtimestamp) as eventtimestamp from mt4datafeederrors m left outer join ( select datafeedname, eventtimestamp from ok_entries where r = 1) oo on m.datafeedname = oo.datafeedname where m.eventtimestamp > coalesce(oo.eventtimestamp, '1900-01-01'::timestamp without time zone) group by m.datafeedname ), notified_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where notified is not null and notified <> '' ), broker as ( select *, row_number() over (partition by feedname order by brokerid) r from ( select distinct b.brokerid, b.name as brokername, dss.classname as feedname from downloadersymbolsettings dss inner join brokersymbollist bsl on dss.symbolid = bsl.symbolid inner join broker b on bsl.brokerid = b.brokerid where dss.enabled = 1) a ) select last.id, last.datafeedname, last.eventtimestamp, last.status, last.errordescription, last.serveraddress, last.username, note.notified, note.eventtimestamp, broker.brokername from last_feed_entry last inner join earliest_entry_after_ok after_ok on last.datafeedname = after_ok.datafeedname inner join broker on last.datafeedname = broker.feedname left outer join ok_entries ok on ok.datafeedname = last.datafeedname left outer join notified_entries note on note.datafeedname = last.datafeedname and note.r = 1 where (ok.r is null or ok.r = 1) and last.datafeedname not in ( select distinct datafeedname from last_feed_entry where status = 'OK') and extract(epoch from (last.eventtimestamp - after_ok.eventtimestamp)) > 60 * 60 and last.eventtimestamp > current_timestamp - interval '1 day' and (note.eventtimestamp is null or note.eventtimestamp < current_timestamp - interval '10 hours') and last.eventtimestamp > current_timestamp - interval '1 hour' and broker.r = 1;
Date: 2026-01-05 11:50:03 Duration: 2ms Database: postgres
-
with rankedmt4 as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors ), last_feed_entry as ( select * from rankedmt4 where r = 1 ), ok_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where status = 'OK' ), earliest_entry_after_ok as ( select m.datafeedname, min(m.eventtimestamp) as eventtimestamp from mt4datafeederrors m left outer join ( select datafeedname, eventtimestamp from ok_entries where r = 1) oo on m.datafeedname = oo.datafeedname where m.eventtimestamp > coalesce(oo.eventtimestamp, '1900-01-01'::timestamp without time zone) group by m.datafeedname ), notified_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where notified is not null and notified <> '' ), broker as ( select *, row_number() over (partition by feedname order by brokerid) r from ( select distinct b.brokerid, b.name as brokername, dss.classname as feedname from downloadersymbolsettings dss inner join brokersymbollist bsl on dss.symbolid = bsl.symbolid inner join broker b on bsl.brokerid = b.brokerid where dss.enabled = 1) a ) select last.id, last.datafeedname, last.eventtimestamp, last.status, last.errordescription, last.serveraddress, last.username, note.notified, note.eventtimestamp, broker.brokername from last_feed_entry last inner join earliest_entry_after_ok after_ok on last.datafeedname = after_ok.datafeedname inner join broker on last.datafeedname = broker.feedname left outer join ok_entries ok on ok.datafeedname = last.datafeedname left outer join notified_entries note on note.datafeedname = last.datafeedname and note.r = 1 where (ok.r is null or ok.r = 1) and last.datafeedname not in ( select distinct datafeedname from last_feed_entry where status = 'OK') and extract(epoch from (last.eventtimestamp - after_ok.eventtimestamp)) > 60 * 60 and last.eventtimestamp > current_timestamp - interval '1 day' and (note.eventtimestamp is null or note.eventtimestamp < current_timestamp - interval '10 hours') and last.eventtimestamp > current_timestamp - interval '1 hour' and broker.r = 1;
Date: 2026-01-05 11:30:02 Duration: 2ms Database: postgres
19 11ms 8 0ms 3ms 1ms WITH pre_symbols AS ( /* find relevant symbols */ ;Times Reported Time consuming prepare #19
Day Hour Count Duration Avg duration 11 8 11ms 1ms -
WITH pre_symbols AS ( /* find relevant symbols */ ;
Date: 2026-01-05 11:14:22 Duration: 3ms Database: postgres
-
WITH pre_symbols AS ( /* find relevant symbols */ ;
Date: 2026-01-05 11:14:22 Duration: 2ms Database: postgres
-
WITH pre_symbols AS ( /* find relevant symbols */ ;
Date: 2026-01-05 11:14:22 Duration: 2ms Database: postgres
20 10ms 7 0ms 2ms 1ms with wh_patitioned as ( ;Times Reported Time consuming prepare #20
Day Hour Count Duration Avg duration 11 7 10ms 1ms -
with wh_patitioned as ( ;
Date: 2026-01-05 11:21:42 Duration: 2ms Database: postgres
-
with wh_patitioned as ( ;
Date: 2026-01-05 11:25:44 Duration: 2ms Database: postgres
-
with wh_patitioned as ( ;
Date: 2026-01-05 11:02:07 Duration: 1ms Database: postgres
Time consuming bind
Rank Total duration Times executed Min duration Max duration Avg duration Query 1 30s970ms 4,888 0ms 53ms 6ms WITH rar_max as ( ;Times Reported Time consuming bind #1
Day Hour Count Duration Avg duration Jan 05 11 4,888 30s970ms 6ms -
WITH rar_max as ( ;
Date: 2026-01-05 11:11:08 Duration: 53ms Database: postgres parameters: $1 = 't', $2 = '489', $3 = '0', $4 = '0', $5 = '0', $6 = '', $7 = '0', $8 = '', $9 = '0', $10 = '', $11 = '0', $12 = '0', $13 = '0', $14 = '0', $15 = '0', $16 = 't', $17 = '0', $18 = '0'
-
WITH rar_max as ( ;
Date: 2026-01-05 11:32:19 Duration: 46ms Database: postgres parameters: $1 = '607450462432900301', $2 = '607450462432900301', $3 = '607450462432900301'
-
WITH rar_max as ( ;
Date: 2026-01-05 11:32:19 Duration: 44ms Database: postgres parameters: $1 = '607450482841105303', $2 = '607450482841105303', $3 = '607450482841105303'
2 8s680ms 25,766 0ms 8ms 0ms SELECT ;Times Reported Time consuming bind #2
Day Hour Count Duration Avg duration 11 25,766 8s680ms 0ms -
SELECT ;
Date: 2026-01-05 11:45:04 Duration: 8ms Database: postgres parameters: $1 = '515840243273647300'
-
SELECT ;
Date: 2026-01-05 11:00:04 Duration: 7ms Database: postgres parameters: $1 = '515840243257534300'
-
SELECT ;
Date: 2026-01-05 11:32:19 Duration: 7ms Database: postgres parameters: $1 = '689', $2 = '689', $3 = '515840246005449300'
3 2s575ms 1,172 0ms 7ms 2ms SELECT symbolid, ;Times Reported Time consuming bind #3
Day Hour Count Duration Avg duration 11 1,172 2s575ms 2ms -
SELECT symbolid, ;
Date: 2026-01-05 11:01:25 Duration: 7ms Database: postgres parameters: $1 = 'BDSWISS', $2 = '60', $3 = '#ADBE'
-
SELECT symbolid, ;
Date: 2026-01-05 11:06:02 Duration: 5ms Database: postgres parameters: $1 = 'ICMARKETS-AU-MT5', $2 = '15', $3 = 'GBPSGD', $4 = 'IT40', $5 = 'JP225', $6 = 'GBPSEK', $7 = 'GBPUSD', $8 = 'NOKJPY'
-
SELECT symbolid, ;
Date: 2026-01-05 11:00:55 Duration: 3ms Database: postgres parameters: $1 = 'PEPPERSTONE', $2 = '15', $3 = 'NZDCAD', $4 = 'NETH25', $5 = 'LTCUSD', $6 = 'LINKUSD', $7 = 'NAS100', $8 = 'NZDJPY', $9 = 'NZDCHF', $10 = 'NOKSEK', $11 = 'NOKJPY'
4 865ms 538 1ms 2ms 1ms SELECT s.symbolid, dss.downloadfrequency, dss.downloadersymbol;Times Reported Time consuming bind #4
Day Hour Count Duration Avg duration 11 538 865ms 1ms -
SELECT s.symbolid, dss.downloadfrequency, dss.downloadersymbol;
Date: 2026-01-05 11:15:27 Duration: 2ms Database: postgres parameters: $1 = 'Alpari'
-
SELECT s.symbolid, dss.downloadfrequency, dss.downloadersymbol;
Date: 2026-01-05 11:45:45 Duration: 2ms Database: postgres parameters: $1 = 'AXIORY'
-
SELECT s.symbolid, dss.downloadfrequency, dss.downloadersymbol;
Date: 2026-01-05 11:48:48 Duration: 2ms Database: postgres parameters: $1 = 'PEPPERSTONEMT5'
5 711ms 29 0ms 48ms 24ms with wh_patitioned as ( ;Times Reported Time consuming bind #5
Day Hour Count Duration Avg duration 11 29 711ms 24ms -
with wh_patitioned as ( ;
Date: 2026-01-05 11:00:03 Duration: 48ms Database: postgres parameters: $1 = '558', $2 = '558', $3 = '558', $4 = '558', $5 = '558', $6 = '558', $7 = '558', $8 = '558', $9 = '558'
-
with wh_patitioned as ( ;
Date: 2026-01-05 11:21:43 Duration: 47ms Database: postgres parameters: $1 = '627', $2 = '627', $3 = '627', $4 = '627', $5 = '627', $6 = '627', $7 = '627', $8 = '627', $9 = '627'
-
with wh_patitioned as ( ;
Date: 2026-01-05 11:42:28 Duration: 47ms Database: postgres parameters: $1 = '558', $2 = '558', $3 = '558', $4 = '558', $5 = '558', $6 = '558', $7 = '558', $8 = '558', $9 = '558'
6 497ms 48 0ms 19ms 10ms WITH /*Latest.JapSticks*/ all_results AS ( SELECT ;Times Reported Time consuming bind #6
Day Hour Count Duration Avg duration 11 48 497ms 10ms -
WITH /*Latest.JapSticks*/ all_results AS ( SELECT ;
Date: 2026-01-05 11:28:16 Duration: 19ms Database: postgres parameters: $1 = '667', $2 = '0', $3 = '0', $4 = '0', $5 = '', $6 = '0', $7 = '', $8 = '0', $9 = '', $10 = '0', $11 = '0'
-
WITH /*Latest.JapSticks*/ all_results AS ( SELECT ;
Date: 2026-01-05 11:06:49 Duration: 19ms Database: postgres parameters: $1 = '489', $2 = '0', $3 = '0', $4 = '0', $5 = '', $6 = '0', $7 = '', $8 = '0', $9 = '', $10 = '0', $11 = '0'
-
WITH /*Latest.JapSticks*/ all_results AS ( SELECT ;
Date: 2026-01-05 11:12:28 Duration: 19ms Database: postgres parameters: $1 = '667', $2 = '0', $3 = '0', $4 = '0', $5 = '', $6 = '0', $7 = '', $8 = '0', $9 = '', $10 = '0', $11 = '0'
7 424ms 60 4ms 17ms 7ms WITH last_candle AS ( ;Times Reported Time consuming bind #7
Day Hour Count Duration Avg duration 11 60 424ms 7ms -
WITH last_candle AS ( ;
Date: 2026-01-05 11:12:04 Duration: 17ms Database: postgres parameters: $1 = '538', $2 = '538'
-
WITH last_candle AS ( ;
Date: 2026-01-05 11:04:00 Duration: 14ms Database: postgres parameters: $1 = '558', $2 = '558'
-
WITH last_candle AS ( ;
Date: 2026-01-05 11:32:00 Duration: 12ms Database: postgres parameters: $1 = '558', $2 = '558'
8 400ms 12 28ms 47ms 33ms with sym_info as ( ;Times Reported Time consuming bind #8
Day Hour Count Duration Avg duration 11 12 400ms 33ms -
with sym_info as ( ;
Date: 2026-01-05 11:06:46 Duration: 47ms Database: postgres parameters: $1 = '620', $2 = 'Forex', $3 = 'Forex', $4 = '620', $5 = 'Forex', $6 = '620', $7 = '620', $8 = 'Forex', $9 = '620'
-
with sym_info as ( ;
Date: 2026-01-05 11:51:57 Duration: 44ms Database: postgres parameters: $1 = '692', $2 = 'Forex', $3 = 'Forex', $4 = '692', $5 = 'Forex', $6 = '692', $7 = '692', $8 = 'Forex', $9 = '692'
-
with sym_info as ( ;
Date: 2026-01-05 11:06:57 Duration: 40ms Database: postgres parameters: $1 = '692', $2 = 'Forex', $3 = 'Forex', $4 = '692', $5 = 'Forex', $6 = '692', $7 = '692', $8 = 'Forex', $9 = '692'
9 394ms 21,849 0ms 3ms 0ms select 1;Times Reported Time consuming bind #9
Day Hour Count Duration Avg duration 11 21,849 394ms 0ms -
select 1;
Date: 2026-01-05 11:17:31 Duration: 3ms Database: postgres
-
select 1;
Date: 2026-01-05 11:02:33 Duration: 2ms Database: postgres
-
select 1;
Date: 2026-01-05 11:02:33 Duration: 1ms Database: postgres
10 266ms 6,676 0ms 0ms 0ms INSERT INTO T15 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;Times Reported Time consuming bind #10
Day Hour Count Duration Avg duration 11 6,676 266ms 0ms -
INSERT INTO T15 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;
Date: 2026-01-05 11:17:58 Duration: 0ms Database: postgres parameters: $1 = '2026-01-02 22:45:00', $2 = '226.15', $3 = '226.68', $4 = '226.01', $5 = '226.515', $6 = '439', $7 = '515840249418515300', $8 = '0', $9 = '2026-01-05 11:17:58.973', $10 = '2026-01-05 11:17:58.939', $11 = '226.15', $12 = '226.68', $13 = '226.01', $14 = '226.515', $15 = '439', $16 = '0', $17 = '2026-01-05 11:17:58.973', $18 = '2026-01-05 11:17:58.939'
-
INSERT INTO T15 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;
Date: 2026-01-05 11:26:40 Duration: 0ms Database: postgres parameters: $1 = '2026-01-05 11:00:00', $2 = '8714.9', $3 = '8724.9', $4 = '8714.3', $5 = '8718.9', $6 = '1207', $7 = '515840248015086300', $8 = '0', $9 = '2026-01-05 11:26:40.773', $10 = '2026-01-05 11:26:40.696', $11 = '8714.9', $12 = '8724.9', $13 = '8714.3', $14 = '8718.9', $15 = '1207', $16 = '0', $17 = '2026-01-05 11:26:40.773', $18 = '2026-01-05 11:26:40.696'
-
INSERT INTO T15 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;
Date: 2026-01-05 11:41:40 Duration: 0ms Database: postgres parameters: $1 = '2026-01-05 11:15:00', $2 = '8718.55', $3 = '8719.9', $4 = '8713.25', $5 = '8719.4', $6 = '1499', $7 = '515840248015086300', $8 = '0', $9 = '2026-01-05 11:41:40.685', $10 = '2026-01-05 11:41:40.606', $11 = '8718.55', $12 = '8719.9', $13 = '8713.25', $14 = '8719.4', $15 = '1499', $16 = '0', $17 = '2026-01-05 11:41:40.685', $18 = '2026-01-05 11:41:40.606'
11 249ms 3,737 0ms 0ms 0ms INSERT INTO T30 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;Times Reported Time consuming bind #11
Day Hour Count Duration Avg duration 11 3,737 249ms 0ms -
INSERT INTO T30 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;
Date: 2026-01-05 11:11:55 Duration: 0ms Database: postgres parameters: $1 = '2026-01-05 10:30:00', $2 = '6871.5', $3 = '6874.55', $4 = '6869.55', $5 = '6873.35', $6 = '5640', $7 = '515840248032224300', $8 = '0', $9 = '2026-01-05 11:11:55.774', $10 = '2026-01-05 11:11:55.685', $11 = '6871.5', $12 = '6874.55', $13 = '6869.55', $14 = '6873.35', $15 = '5640', $16 = '0', $17 = '2026-01-05 11:11:55.774', $18 = '2026-01-05 11:11:55.685'
-
INSERT INTO T30 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;
Date: 2026-01-05 11:11:39 Duration: 0ms Database: postgres parameters: $1 = '2026-01-05 10:30:00', $2 = '8716.7', $3 = '8720.55', $4 = '8713.3', $5 = '8714.4', $6 = '3242', $7 = '515840248015340300', $8 = '0', $9 = '2026-01-05 11:11:39.665', $10 = '2026-01-05 11:11:39.546', $11 = '8716.7', $12 = '8720.55', $13 = '8713.3', $14 = '8714.4', $15 = '3242', $16 = '0', $17 = '2026-01-05 11:11:39.665', $18 = '2026-01-05 11:11:39.546'
-
INSERT INTO T30 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;
Date: 2026-01-05 11:41:40 Duration: 0ms Database: postgres parameters: $1 = '2026-01-05 11:00:00', $2 = '8714.9', $3 = '8724.9', $4 = '8713.25', $5 = '8719.4', $6 = '2706', $7 = '515840248015340300', $8 = '0', $9 = '2026-01-05 11:41:40.704', $10 = '2026-01-05 11:41:40.616', $11 = '8714.9', $12 = '8724.9', $13 = '8713.25', $14 = '8719.4', $15 = '2706', $16 = '0', $17 = '2026-01-05 11:41:40.704', $18 = '2026-01-05 11:41:40.616'
12 186ms 2,387 0ms 0ms 0ms INSERT INTO T60 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;Times Reported Time consuming bind #12
Day Hour Count Duration Avg duration 11 2,387 186ms 0ms -
INSERT INTO T60 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;
Date: 2026-01-05 11:02:26 Duration: 0ms Database: postgres parameters: $1 = '2026-01-02 21:00:00', $2 = '336.08', $3 = '336.65', $4 = '335.05', $5 = '335.07', $6 = '1842', $7 = '515840247899857300', $8 = '0', $9 = '2026-01-05 11:02:26.182', $10 = '2026-01-05 11:02:26.14', $11 = '336.08', $12 = '336.65', $13 = '335.05', $14 = '335.07', $15 = '1842', $16 = '0', $17 = '2026-01-05 11:02:26.182', $18 = '2026-01-05 11:02:26.14'
-
INSERT INTO T60 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;
Date: 2026-01-05 11:11:27 Duration: 0ms Database: postgres parameters: $1 = '2026-01-02 21:00:00', $2 = '336.08', $3 = '336.65', $4 = '335.05', $5 = '335.07', $6 = '1842', $7 = '515840247899857300', $8 = '0', $9 = '2026-01-05 11:11:27.438', $10 = '2026-01-05 11:11:27.392', $11 = '336.08', $12 = '336.65', $13 = '335.05', $14 = '335.07', $15 = '1842', $16 = '0', $17 = '2026-01-05 11:11:27.438', $18 = '2026-01-05 11:11:27.392'
-
INSERT INTO T60 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;
Date: 2026-01-05 11:10:55 Duration: 0ms Database: postgres parameters: $1 = '2026-01-05 09:00:00', $2 = '25304.67', $3 = '25313.3', $4 = '25259.8', $5 = '25287.67', $6 = '11168', $7 = '515840248039327300', $8 = '0', $9 = '2026-01-05 11:10:55.299', $10 = '2026-01-05 11:10:55.219', $11 = '25304.67', $12 = '25313.3', $13 = '25259.8', $14 = '25287.67', $15 = '11168', $16 = '0', $17 = '2026-01-05 11:10:55.299', $18 = '2026-01-05 11:10:55.219'
13 179ms 474 0ms 2ms 0ms SELECT NULL AS TABLE_CAT, n.nspname AS TABLE_SCHEM, c.relname AS TABLE_NAME, CASE n.nspname ~ '^pg_' OR n.nspname = 'information_schema' WHEN true THEN CASE WHEN n.nspname = 'pg_catalog' OR n.nspname = 'information_schema' THEN CASE c.relkind WHEN 'r' THEN 'SYSTEM TABLE' WHEN 'v' THEN 'SYSTEM VIEW' WHEN 'i' THEN 'SYSTEM INDEX' ELSE NULL END WHEN n.nspname = 'pg_toast' THEN CASE c.relkind WHEN 'r' THEN 'SYSTEM TOAST TABLE' WHEN 'i' THEN 'SYSTEM TOAST INDEX' ELSE NULL END ELSE CASE c.relkind WHEN 'r' THEN 'TEMPORARY TABLE' WHEN 'p' THEN 'TEMPORARY TABLE' WHEN 'i' THEN 'TEMPORARY INDEX' WHEN 'S' THEN 'TEMPORARY SEQUENCE' WHEN 'v' THEN 'TEMPORARY VIEW' ELSE NULL END END WHEN false THEN CASE c.relkind WHEN 'r' THEN 'TABLE' WHEN 'p' THEN 'PARTITIONED TABLE' WHEN 'i' THEN 'INDEX' WHEN 'S' THEN 'SEQUENCE' WHEN 'v' THEN 'VIEW' WHEN 'c' THEN 'TYPE' WHEN 'f' THEN 'FOREIGN TABLE' WHEN 'm' THEN 'MATERIALIZED VIEW' ELSE NULL END ELSE NULL END AS TABLE_TYPE, d.description AS REMARKS, '' as TYPE_CAT, '' as TYPE_SCHEM, '' as TYPE_NAME, '' AS SELF_REFERENCING_COL_NAME, '' AS REF_GENERATION FROM pg_catalog.pg_namespace n, pg_catalog.pg_class c LEFT JOIN pg_catalog.pg_description d ON (c.oid = d.objoid AND d.objsubid = 0) LEFT JOIN pg_catalog.pg_class dc ON (d.classoid = dc.oid AND dc.relname = 'pg_class') LEFT JOIN pg_catalog.pg_namespace dn ON (dn.oid = dc.relnamespace AND dn.nspname = 'pg_catalog') WHERE c.relnamespace = n.oid AND c.relname LIKE 'PROBABLYNOT' AND (false OR (c.relkind = 'r' AND n.nspname !~ '^pg_' AND n.nspname <> 'information_schema')) ORDER BY TABLE_TYPE, TABLE_SCHEM, TABLE_NAME;Times Reported Time consuming bind #13
Day Hour Count Duration Avg duration 11 474 179ms 0ms -
SELECT NULL AS TABLE_CAT, n.nspname AS TABLE_SCHEM, c.relname AS TABLE_NAME, CASE n.nspname ~ '^pg_' OR n.nspname = 'information_schema' WHEN true THEN CASE WHEN n.nspname = 'pg_catalog' OR n.nspname = 'information_schema' THEN CASE c.relkind WHEN 'r' THEN 'SYSTEM TABLE' WHEN 'v' THEN 'SYSTEM VIEW' WHEN 'i' THEN 'SYSTEM INDEX' ELSE NULL END WHEN n.nspname = 'pg_toast' THEN CASE c.relkind WHEN 'r' THEN 'SYSTEM TOAST TABLE' WHEN 'i' THEN 'SYSTEM TOAST INDEX' ELSE NULL END ELSE CASE c.relkind WHEN 'r' THEN 'TEMPORARY TABLE' WHEN 'p' THEN 'TEMPORARY TABLE' WHEN 'i' THEN 'TEMPORARY INDEX' WHEN 'S' THEN 'TEMPORARY SEQUENCE' WHEN 'v' THEN 'TEMPORARY VIEW' ELSE NULL END END WHEN false THEN CASE c.relkind WHEN 'r' THEN 'TABLE' WHEN 'p' THEN 'PARTITIONED TABLE' WHEN 'i' THEN 'INDEX' WHEN 'S' THEN 'SEQUENCE' WHEN 'v' THEN 'VIEW' WHEN 'c' THEN 'TYPE' WHEN 'f' THEN 'FOREIGN TABLE' WHEN 'm' THEN 'MATERIALIZED VIEW' ELSE NULL END ELSE NULL END AS TABLE_TYPE, d.description AS REMARKS, '' as TYPE_CAT, '' as TYPE_SCHEM, '' as TYPE_NAME, '' AS SELF_REFERENCING_COL_NAME, '' AS REF_GENERATION FROM pg_catalog.pg_namespace n, pg_catalog.pg_class c LEFT JOIN pg_catalog.pg_description d ON (c.oid = d.objoid AND d.objsubid = 0) LEFT JOIN pg_catalog.pg_class dc ON (d.classoid = dc.oid AND dc.relname = 'pg_class') LEFT JOIN pg_catalog.pg_namespace dn ON (dn.oid = dc.relnamespace AND dn.nspname = 'pg_catalog') WHERE c.relnamespace = n.oid AND c.relname LIKE 'PROBABLYNOT' AND (false OR (c.relkind = 'r' AND n.nspname !~ '^pg_' AND n.nspname <> 'information_schema')) ORDER BY TABLE_TYPE, TABLE_SCHEM, TABLE_NAME;
Date: 2026-01-05 11:14:22 Duration: 2ms Database: postgres
-
SELECT NULL AS TABLE_CAT, n.nspname AS TABLE_SCHEM, c.relname AS TABLE_NAME, CASE n.nspname ~ '^pg_' OR n.nspname = 'information_schema' WHEN true THEN CASE WHEN n.nspname = 'pg_catalog' OR n.nspname = 'information_schema' THEN CASE c.relkind WHEN 'r' THEN 'SYSTEM TABLE' WHEN 'v' THEN 'SYSTEM VIEW' WHEN 'i' THEN 'SYSTEM INDEX' ELSE NULL END WHEN n.nspname = 'pg_toast' THEN CASE c.relkind WHEN 'r' THEN 'SYSTEM TOAST TABLE' WHEN 'i' THEN 'SYSTEM TOAST INDEX' ELSE NULL END ELSE CASE c.relkind WHEN 'r' THEN 'TEMPORARY TABLE' WHEN 'p' THEN 'TEMPORARY TABLE' WHEN 'i' THEN 'TEMPORARY INDEX' WHEN 'S' THEN 'TEMPORARY SEQUENCE' WHEN 'v' THEN 'TEMPORARY VIEW' ELSE NULL END END WHEN false THEN CASE c.relkind WHEN 'r' THEN 'TABLE' WHEN 'p' THEN 'PARTITIONED TABLE' WHEN 'i' THEN 'INDEX' WHEN 'S' THEN 'SEQUENCE' WHEN 'v' THEN 'VIEW' WHEN 'c' THEN 'TYPE' WHEN 'f' THEN 'FOREIGN TABLE' WHEN 'm' THEN 'MATERIALIZED VIEW' ELSE NULL END ELSE NULL END AS TABLE_TYPE, d.description AS REMARKS, '' as TYPE_CAT, '' as TYPE_SCHEM, '' as TYPE_NAME, '' AS SELF_REFERENCING_COL_NAME, '' AS REF_GENERATION FROM pg_catalog.pg_namespace n, pg_catalog.pg_class c LEFT JOIN pg_catalog.pg_description d ON (c.oid = d.objoid AND d.objsubid = 0) LEFT JOIN pg_catalog.pg_class dc ON (d.classoid = dc.oid AND dc.relname = 'pg_class') LEFT JOIN pg_catalog.pg_namespace dn ON (dn.oid = dc.relnamespace AND dn.nspname = 'pg_catalog') WHERE c.relnamespace = n.oid AND c.relname LIKE 'PROBABLYNOT' AND (false OR (c.relkind = 'r' AND n.nspname !~ '^pg_' AND n.nspname <> 'information_schema')) ORDER BY TABLE_TYPE, TABLE_SCHEM, TABLE_NAME;
Date: 2026-01-05 11:14:23 Duration: 0ms Database: postgres
-
SELECT NULL AS TABLE_CAT, n.nspname AS TABLE_SCHEM, c.relname AS TABLE_NAME, CASE n.nspname ~ '^pg_' OR n.nspname = 'information_schema' WHEN true THEN CASE WHEN n.nspname = 'pg_catalog' OR n.nspname = 'information_schema' THEN CASE c.relkind WHEN 'r' THEN 'SYSTEM TABLE' WHEN 'v' THEN 'SYSTEM VIEW' WHEN 'i' THEN 'SYSTEM INDEX' ELSE NULL END WHEN n.nspname = 'pg_toast' THEN CASE c.relkind WHEN 'r' THEN 'SYSTEM TOAST TABLE' WHEN 'i' THEN 'SYSTEM TOAST INDEX' ELSE NULL END ELSE CASE c.relkind WHEN 'r' THEN 'TEMPORARY TABLE' WHEN 'p' THEN 'TEMPORARY TABLE' WHEN 'i' THEN 'TEMPORARY INDEX' WHEN 'S' THEN 'TEMPORARY SEQUENCE' WHEN 'v' THEN 'TEMPORARY VIEW' ELSE NULL END END WHEN false THEN CASE c.relkind WHEN 'r' THEN 'TABLE' WHEN 'p' THEN 'PARTITIONED TABLE' WHEN 'i' THEN 'INDEX' WHEN 'S' THEN 'SEQUENCE' WHEN 'v' THEN 'VIEW' WHEN 'c' THEN 'TYPE' WHEN 'f' THEN 'FOREIGN TABLE' WHEN 'm' THEN 'MATERIALIZED VIEW' ELSE NULL END ELSE NULL END AS TABLE_TYPE, d.description AS REMARKS, '' as TYPE_CAT, '' as TYPE_SCHEM, '' as TYPE_NAME, '' AS SELF_REFERENCING_COL_NAME, '' AS REF_GENERATION FROM pg_catalog.pg_namespace n, pg_catalog.pg_class c LEFT JOIN pg_catalog.pg_description d ON (c.oid = d.objoid AND d.objsubid = 0) LEFT JOIN pg_catalog.pg_class dc ON (d.classoid = dc.oid AND dc.relname = 'pg_class') LEFT JOIN pg_catalog.pg_namespace dn ON (dn.oid = dc.relnamespace AND dn.nspname = 'pg_catalog') WHERE c.relnamespace = n.oid AND c.relname LIKE 'PROBABLYNOT' AND (false OR (c.relkind = 'r' AND n.nspname !~ '^pg_' AND n.nspname <> 'information_schema')) ORDER BY TABLE_TYPE, TABLE_SCHEM, TABLE_NAME;
Date: 2026-01-05 11:14:23 Duration: 0ms Database: postgres
14 82ms 93 0ms 1ms 0ms SELECT timegranularity FROM brokersymbollist bsl INNER JOIN symbols s ON bsl.symbolid = s.symbolid INNER JOIN downloadersymbolsettings dss on s.symbolid = dss.symbolid LEFT OUTER JOIN brokerinstrumentmapping bdfi ON bdfi.brokerid = $1 AND dss.datafeedinstrumentid = bdfi.datafeedinstrumentid WHERE s.nonliquid = 0 and s.deleted = 0 and dss.enabled = 1 AND s.symbol ILIKE $2 AND bsl.brokerid = $3 AND timegranularity >= 15 ORDER BY timegranularity LIMIT 1;Times Reported Time consuming bind #14
Day Hour Count Duration Avg duration 11 93 82ms 0ms -
SELECT timegranularity FROM brokersymbollist bsl INNER JOIN symbols s ON bsl.symbolid = s.symbolid INNER JOIN downloadersymbolsettings dss on s.symbolid = dss.symbolid LEFT OUTER JOIN brokerinstrumentmapping bdfi ON bdfi.brokerid = $1 AND dss.datafeedinstrumentid = bdfi.datafeedinstrumentid WHERE s.nonliquid = 0 and s.deleted = 0 and dss.enabled = 1 AND s.symbol ILIKE $2 AND bsl.brokerid = $3 AND timegranularity >= 15 ORDER BY timegranularity LIMIT 1;
Date: 2026-01-05 11:46:01 Duration: 1ms Database: postgres parameters: $1 = '558', $2 = 'USDCAD', $3 = '558'
-
SELECT timegranularity FROM brokersymbollist bsl INNER JOIN symbols s ON bsl.symbolid = s.symbolid INNER JOIN downloadersymbolsettings dss on s.symbolid = dss.symbolid LEFT OUTER JOIN brokerinstrumentmapping bdfi ON bdfi.brokerid = $1 AND dss.datafeedinstrumentid = bdfi.datafeedinstrumentid WHERE s.nonliquid = 0 and s.deleted = 0 and dss.enabled = 1 AND s.symbol ILIKE $2 AND bsl.brokerid = $3 AND timegranularity >= 15 ORDER BY timegranularity LIMIT 1;
Date: 2026-01-05 11:16:05 Duration: 1ms Database: postgres parameters: $1 = '538', $2 = 'EURUSDb', $3 = '538'
-
SELECT timegranularity FROM brokersymbollist bsl INNER JOIN symbols s ON bsl.symbolid = s.symbolid INNER JOIN downloadersymbolsettings dss on s.symbolid = dss.symbolid LEFT OUTER JOIN brokerinstrumentmapping bdfi ON bdfi.brokerid = $1 AND dss.datafeedinstrumentid = bdfi.datafeedinstrumentid WHERE s.nonliquid = 0 and s.deleted = 0 and dss.enabled = 1 AND s.symbol ILIKE $2 AND bsl.brokerid = $3 AND timegranularity >= 15 ORDER BY timegranularity LIMIT 1;
Date: 2026-01-05 11:45:35 Duration: 1ms Database: postgres parameters: $1 = '621', $2 = 'USDJPY', $3 = '621'
15 48ms 437 0ms 1ms 0ms select category, ;Times Reported Time consuming bind #15
Day Hour Count Duration Avg duration 11 437 48ms 0ms -
select category, ;
Date: 2026-01-05 11:23:30 Duration: 1ms Database: postgres parameters: $1 = '604104683406582307', $2 = 'symbol', $3 = 'AUDJPY', $4 = 'CADJPY', $5 = 'GBPJPY', $6 = 'NZDJPY', $7 = 'CHFJPY', $8 = 'EURJPY', $9 = 'GBPAUD', $10 = 'GBPNZD', $11 = 'EURAUD', $12 = 'EURNZD', $13 = 'GBPCAD', $14 = 'EURGBP', $15 = 'EURCAD', $16 = 'GBPCAD', $17 = 'USDSGD', $18 = 'CADCHF', $19 = 'EURNZD', $20 = 'CADJPY', $21 = 'EURCAD', $22 = 'AUDNZD', $23 = 'NZDUSD', $24 = 'GBPCHF', $25 = 'AUDJPY', $26 = 'EURJPY', $27 = 'EURCHF', $28 = 'NZDCHF', $29 = 'AUDCHF', $30 = 'NZDJPY', $31 = 'GBPJPY', $32 = 'GBPCHF', $33 = 'GBPAUD', $34 = 'AUDCAD', $35 = 'NZDCAD', $36 = 'NZDCAD', $37 = 'GBPNZD', $38 = 'USDSGD', $39 = 'EURAUD', $40 = 'CHFJPY', $41 = 'AUDCAD', $42 = 'NZDUSD', $43 = 'AUDCHF', $44 = 'EURCHF', $45 = 'CADCHF', $46 = 'NZDCHF', $47 = 'AUDNZD', $48 = 'EURGBP', $49 = '604104683406582307', $50 = 'symbol', $51 = 'AUDJPY', $52 = 'CADJPY', $53 = 'GBPJPY', $54 = 'NZDJPY', $55 = 'CHFJPY', $56 = 'EURJPY', $57 = 'GBPAUD', $58 = 'GBPNZD', $59 = 'EURAUD', $60 = 'EURNZD', $61 = 'GBPCAD', $62 = 'EURGBP', $63 = 'EURCAD', $64 = 'GBPCAD', $65 = 'USDSGD', $66 = 'CADCHF', $67 = 'EURNZD', $68 = 'CADJPY', $69 = 'EURCAD', $70 = 'AUDNZD', $71 = 'NZDUSD', $72 = 'GBPCHF', $73 = 'AUDJPY', $74 = 'EURJPY', $75 = 'EURCHF', $76 = 'NZDCHF', $77 = 'AUDCHF', $78 = 'NZDJPY', $79 = 'GBPJPY', $80 = 'GBPCHF', $81 = 'GBPAUD', $82 = 'AUDCAD', $83 = 'NZDCAD', $84 = 'NZDCAD', $85 = 'GBPNZD', $86 = 'USDSGD', $87 = 'EURAUD', $88 = 'CHFJPY', $89 = 'AUDCAD', $90 = 'NZDUSD', $91 = 'AUDCHF', $92 = 'EURCHF', $93 = 'CADCHF', $94 = 'NZDCHF', $95 = 'AUDNZD', $96 = 'EURGBP'
-
select category, ;
Date: 2026-01-05 11:51:07 Duration: 0ms Database: postgres parameters: $1 = '605634061978417307', $2 = 'symbol', $3 = 'AUDJPY', $4 = 'USDJPY', $5 = 'EURMXN', $6 = 'EURJPY', $7 = 'GBPJPY', $8 = 'USDMXN', $9 = 'EURAUD', $10 = 'GBPUSD', $11 = 'USDCAD', $12 = 'EURUSD', $13 = 'NZDUSD', $14 = 'AUDUSD', $15 = 'EURCHF', $16 = 'AUDCAD', $17 = 'USDCHF', $18 = 'USDCAD', $19 = 'EURGBP', $20 = 'NZDUSD', $21 = 'EURGBP', $22 = 'USDMXN', $23 = 'GBPUSD', $24 = 'EURCHF', $25 = 'GBPJPY', $26 = 'EURJPY', $27 = 'USDCHF', $28 = 'AUDCAD', $29 = 'USDJPY', $30 = 'AUDUSD', $31 = 'EURMXN', $32 = 'EURAUD', $33 = 'AUDJPY', $34 = 'EURUSD', $35 = '605634061978417307', $36 = 'symbol', $37 = 'AUDJPY', $38 = 'USDJPY', $39 = 'EURMXN', $40 = 'EURJPY', $41 = 'GBPJPY', $42 = 'USDMXN', $43 = 'EURAUD', $44 = 'GBPUSD', $45 = 'USDCAD', $46 = 'EURUSD', $47 = 'NZDUSD', $48 = 'AUDUSD', $49 = 'EURCHF', $50 = 'AUDCAD', $51 = 'USDCHF', $52 = 'USDCAD', $53 = 'EURGBP', $54 = 'NZDUSD', $55 = 'EURGBP', $56 = 'USDMXN', $57 = 'GBPUSD', $58 = 'EURCHF', $59 = 'GBPJPY', $60 = 'EURJPY', $61 = 'USDCHF', $62 = 'AUDCAD', $63 = 'USDJPY', $64 = 'AUDUSD', $65 = 'EURMXN', $66 = 'EURAUD', $67 = 'AUDJPY', $68 = 'EURUSD'
-
select category, ;
Date: 2026-01-05 11:23:42 Duration: 0ms Database: postgres parameters: $1 = '604104683407304307', $2 = 'symbol', $3 = 'UK100', $4 = 'JP225', $5 = 'US2000', $6 = 'US30', $7 = 'US500', $8 = 'USTEC', $9 = 'AUS200', $10 = 'F40', $11 = 'CHINA50', $12 = 'STOXX50', $13 = 'US2000', $14 = 'IT40', $15 = 'HK50', $16 = 'UK100', $17 = 'US500', $18 = 'AUS200', $19 = 'US30', $20 = 'JP225', $21 = 'USTEC', $22 = 'CHINA50', $23 = 'ES35', $24 = 'STOXX50', $25 = 'F40', $26 = 'HK50', $27 = 'IT40', $28 = 'ES35', $29 = '604104683407304307', $30 = 'symbol', $31 = 'UK100', $32 = 'JP225', $33 = 'US2000', $34 = 'US30', $35 = 'US500', $36 = 'USTEC', $37 = 'AUS200', $38 = 'F40', $39 = 'CHINA50', $40 = 'STOXX50', $41 = 'US2000', $42 = 'IT40', $43 = 'HK50', $44 = 'UK100', $45 = 'US500', $46 = 'AUS200', $47 = 'US30', $48 = 'JP225', $49 = 'USTEC', $50 = 'CHINA50', $51 = 'ES35', $52 = 'STOXX50', $53 = 'F40', $54 = 'HK50', $55 = 'IT40', $56 = 'ES35'
16 40ms 1 40ms 40ms 40ms with maxwhid as ( ;Times Reported Time consuming bind #16
Day Hour Count Duration Avg duration 11 1 40ms 40ms -
with maxwhid as ( ;
Date: 2026-01-05 11:12:14 Duration: 40ms Database: postgres parameters: $1 = '335', $2 = '621', $3 = '637', $4 = '642', $5 = '660', $6 = '666', $7 = '643', $8 = '630', $9 = '680', $10 = '641', $11 = '431', $12 = '622', $13 = '489', $14 = '529', $15 = '576', $16 = '665', $17 = '667', $18 = '558', $19 = '620', $20 = '125', $21 = '488', $22 = '567', $23 = '689', $24 = '700', $25 = '758', $26 = '763', $27 = '765', $28 = '817', $29 = '914', $30 = '972'
17 39ms 8 3ms 6ms 4ms SELECT DISTINCT ON (basegroupname, symbol) ;Times Reported Time consuming bind #17
Day Hour Count Duration Avg duration 11 8 39ms 4ms -
SELECT DISTINCT ON (basegroupname, symbol) ;
Date: 2026-01-05 11:11:35 Duration: 6ms Database: postgres parameters: $1 = '958', $2 = '958'
-
SELECT DISTINCT ON (basegroupname, symbol) ;
Date: 2026-01-05 11:11:00 Duration: 5ms Database: postgres parameters: $1 = '627', $2 = '627'
-
SELECT DISTINCT ON (basegroupname, symbol) ;
Date: 2026-01-05 11:06:02 Duration: 5ms Database: postgres parameters: $1 = '627', $2 = '627'
18 35ms 8 2ms 7ms 4ms WITH pre_symbols AS ( /* find relevant symbols */ ;Times Reported Time consuming bind #18
Day Hour Count Duration Avg duration 11 8 35ms 4ms -
WITH pre_symbols AS ( /* find relevant symbols */ ;
Date: 2026-01-05 11:14:22 Duration: 7ms Database: postgres parameters: $1 = '1018', $2 = 'ICMARKETS-AU-MT5', $3 = 'AAPL.NAS', $4 = 'ABBV.NYSE', $5 = 'AMCR.NYSE', $6 = 'AMP.NYSE', $7 = 'AMZN.NAS', $8 = 'ANZ.ASX', $9 = 'AUDJPY', $10 = 'AUDUSD', $11 = 'AUS200', $12 = 'BABA.NYSE', $13 = 'BIIB.NAS', $14 = 'BXB.ASX', $15 = 'CBA.ASX', $16 = 'CHINA50', $17 = 'CSL.ASX', $18 = 'DE30', $19 = 'ES35', $20 = 'EURCHF', $21 = 'EURGBP', $22 = 'EURUSD', $23 = 'F40', $24 = 'FMG.ASX', $25 = 'GBPJPY', $26 = 'GBPUSD', $27 = 'GOOG.NAS', $28 = 'HK50', $29 = 'IT40', $30 = 'JP225', $31 = 'KO.NYSE', $32 = 'MQG.ASX', $33 = 'MSFT.NAS', $34 = 'NAB.ASX', $35 = 'NFLX.NAS', $36 = 'PYPL.NAS', $37 = 'QBE.ASX', $38 = 'STOXX50', $39 = 'SUN.ASX', $40 = 'TCL.ASX', $41 = 'TLS.ASX', $42 = 'TSLA.NAS', $43 = 'UK100', $44 = 'UNH.NYSE', $45 = 'US2000', $46 = 'US30', $47 = 'US500', $48 = 'USDCAD', $49 = 'USDCHF', $50 = 'USDCNH', $51 = 'USDJPY', $52 = 'USTEC', $53 = 'WBC.ASX', $54 = 'WES.ASX', $55 = 'WOW.ASX', $56 = 'WPL.ASX', $57 = 'XAUEUR', $58 = 'XAUUSD', $59 = 'XBRUSD', $60 = 'XTIUSD', $61 = 'AAPL.NAS', $62 = 'ABBV.NYSE', $63 = 'AMCR.NYSE', $64 = 'AMP.NYSE', $65 = 'AMZN.NAS', $66 = 'ANZ.ASX', $67 = 'AUDJPY', $68 = 'AUDUSD', $69 = 'AUS200', $70 = 'BABA.NYSE', $71 = 'BIIB.NAS', $72 = 'BXB.ASX', $73 = 'CBA.ASX', $74 = 'CHINA50', $75 = 'CSL.ASX', $76 = 'DE30', $77 = 'ES35', $78 = 'EURCHF', $79 = 'EURGBP', $80 = 'EURUSD', $81 = 'F40', $82 = 'FMG.ASX', $83 = 'GBPJPY', $84 = 'GBPUSD', $85 = 'GOOG.NAS', $86 = 'HK50', $87 = 'IT40', $88 = 'JP225', $89 = 'KO.NYSE', $90 = 'MQG.ASX', $91 = 'MSFT.NAS', $92 = 'NAB.ASX', $93 = 'NFLX.NAS', $94 = 'PYPL.NAS', $95 = 'QBE.ASX', $96 = 'STOXX50', $97 = 'SUN.ASX', $98 = 'TCL.ASX', $99 = 'TLS.ASX', $100 = 'TSLA.NAS', $101 = 'UK100', $102 = 'UNH.NYSE', $103 = 'US2000', $104 = 'US30', $105 = 'US500', $106 = 'USDCAD', $107 = 'USDCHF', $108 = 'USDCNH', $109 = 'USDJPY', $110 = 'USTEC', $111 = 'WBC.ASX', $112 = 'WES.ASX', $113 = 'WOW.ASX', $114 = 'WPL.ASX', $115 = 'XAUEUR', $116 = 'XAUUSD', $117 = 'XBRUSD', $118 = 'XTIUSD', $119 = '5'
-
WITH pre_symbols AS ( /* find relevant symbols */ ;
Date: 2026-01-05 11:14:22 Duration: 6ms Database: postgres parameters: $1 = '1018', $2 = 'ICMARKETS-AU-MT5', $3 = 'AAPL.NAS', $4 = 'ABBV.NYSE', $5 = 'AMCR.NYSE', $6 = 'AMP.NYSE', $7 = 'AMZN.NAS', $8 = 'ANZ.ASX', $9 = 'AUDJPY', $10 = 'AUDUSD', $11 = 'AUS200', $12 = 'BABA.NYSE', $13 = 'BIIB.NAS', $14 = 'BXB.ASX', $15 = 'CBA.ASX', $16 = 'CHINA50', $17 = 'CSL.ASX', $18 = 'DE30', $19 = 'ES35', $20 = 'EURCHF', $21 = 'EURGBP', $22 = 'EURUSD', $23 = 'F40', $24 = 'FMG.ASX', $25 = 'GBPJPY', $26 = 'GBPUSD', $27 = 'GOOG.NAS', $28 = 'HK50', $29 = 'IT40', $30 = 'JP225', $31 = 'KO.NYSE', $32 = 'MQG.ASX', $33 = 'MSFT.NAS', $34 = 'NAB.ASX', $35 = 'NFLX.NAS', $36 = 'PYPL.NAS', $37 = 'QBE.ASX', $38 = 'STOXX50', $39 = 'SUN.ASX', $40 = 'TCL.ASX', $41 = 'TLS.ASX', $42 = 'TSLA.NAS', $43 = 'UK100', $44 = 'UNH.NYSE', $45 = 'US2000', $46 = 'US30', $47 = 'US500', $48 = 'USDCAD', $49 = 'USDCHF', $50 = 'USDCNH', $51 = 'USDJPY', $52 = 'USTEC', $53 = 'WBC.ASX', $54 = 'WES.ASX', $55 = 'WOW.ASX', $56 = 'WPL.ASX', $57 = 'XAUEUR', $58 = 'XAUUSD', $59 = 'XBRUSD', $60 = 'XTIUSD', $61 = 'AAPL.NAS', $62 = 'ABBV.NYSE', $63 = 'AMCR.NYSE', $64 = 'AMP.NYSE', $65 = 'AMZN.NAS', $66 = 'ANZ.ASX', $67 = 'AUDJPY', $68 = 'AUDUSD', $69 = 'AUS200', $70 = 'BABA.NYSE', $71 = 'BIIB.NAS', $72 = 'BXB.ASX', $73 = 'CBA.ASX', $74 = 'CHINA50', $75 = 'CSL.ASX', $76 = 'DE30', $77 = 'ES35', $78 = 'EURCHF', $79 = 'EURGBP', $80 = 'EURUSD', $81 = 'F40', $82 = 'FMG.ASX', $83 = 'GBPJPY', $84 = 'GBPUSD', $85 = 'GOOG.NAS', $86 = 'HK50', $87 = 'IT40', $88 = 'JP225', $89 = 'KO.NYSE', $90 = 'MQG.ASX', $91 = 'MSFT.NAS', $92 = 'NAB.ASX', $93 = 'NFLX.NAS', $94 = 'PYPL.NAS', $95 = 'QBE.ASX', $96 = 'STOXX50', $97 = 'SUN.ASX', $98 = 'TCL.ASX', $99 = 'TLS.ASX', $100 = 'TSLA.NAS', $101 = 'UK100', $102 = 'UNH.NYSE', $103 = 'US2000', $104 = 'US30', $105 = 'US500', $106 = 'USDCAD', $107 = 'USDCHF', $108 = 'USDCNH', $109 = 'USDJPY', $110 = 'USTEC', $111 = 'WBC.ASX', $112 = 'WES.ASX', $113 = 'WOW.ASX', $114 = 'WPL.ASX', $115 = 'XAUEUR', $116 = 'XAUUSD', $117 = 'XBRUSD', $118 = 'XTIUSD', $119 = '5'
-
WITH pre_symbols AS ( /* find relevant symbols */ ;
Date: 2026-01-05 11:14:22 Duration: 6ms Database: postgres parameters: $1 = '1018', $2 = 'ICMARKETS-AU-MT5', $3 = 'AAPL.NAS', $4 = 'ABBV.NYSE', $5 = 'AMCR.NYSE', $6 = 'AMP.NYSE', $7 = 'AMZN.NAS', $8 = 'ANZ.ASX', $9 = 'AUDJPY', $10 = 'AUDUSD', $11 = 'AUS200', $12 = 'BABA.NYSE', $13 = 'BIIB.NAS', $14 = 'BXB.ASX', $15 = 'CBA.ASX', $16 = 'CHINA50', $17 = 'CSL.ASX', $18 = 'DE30', $19 = 'ES35', $20 = 'EURCHF', $21 = 'EURGBP', $22 = 'EURUSD', $23 = 'F40', $24 = 'FMG.ASX', $25 = 'GBPJPY', $26 = 'GBPUSD', $27 = 'GOOG.NAS', $28 = 'HK50', $29 = 'IT40', $30 = 'JP225', $31 = 'KO.NYSE', $32 = 'MQG.ASX', $33 = 'MSFT.NAS', $34 = 'NAB.ASX', $35 = 'NFLX.NAS', $36 = 'PYPL.NAS', $37 = 'QBE.ASX', $38 = 'STOXX50', $39 = 'SUN.ASX', $40 = 'TCL.ASX', $41 = 'TLS.ASX', $42 = 'TSLA.NAS', $43 = 'UK100', $44 = 'UNH.NYSE', $45 = 'US2000', $46 = 'US30', $47 = 'US500', $48 = 'USDCAD', $49 = 'USDCHF', $50 = 'USDCNH', $51 = 'USDJPY', $52 = 'USTEC', $53 = 'WBC.ASX', $54 = 'WES.ASX', $55 = 'WOW.ASX', $56 = 'WPL.ASX', $57 = 'XAUEUR', $58 = 'XAUUSD', $59 = 'XBRUSD', $60 = 'XTIUSD', $61 = 'AAPL.NAS', $62 = 'ABBV.NYSE', $63 = 'AMCR.NYSE', $64 = 'AMP.NYSE', $65 = 'AMZN.NAS', $66 = 'ANZ.ASX', $67 = 'AUDJPY', $68 = 'AUDUSD', $69 = 'AUS200', $70 = 'BABA.NYSE', $71 = 'BIIB.NAS', $72 = 'BXB.ASX', $73 = 'CBA.ASX', $74 = 'CHINA50', $75 = 'CSL.ASX', $76 = 'DE30', $77 = 'ES35', $78 = 'EURCHF', $79 = 'EURGBP', $80 = 'EURUSD', $81 = 'F40', $82 = 'FMG.ASX', $83 = 'GBPJPY', $84 = 'GBPUSD', $85 = 'GOOG.NAS', $86 = 'HK50', $87 = 'IT40', $88 = 'JP225', $89 = 'KO.NYSE', $90 = 'MQG.ASX', $91 = 'MSFT.NAS', $92 = 'NAB.ASX', $93 = 'NFLX.NAS', $94 = 'PYPL.NAS', $95 = 'QBE.ASX', $96 = 'STOXX50', $97 = 'SUN.ASX', $98 = 'TCL.ASX', $99 = 'TLS.ASX', $100 = 'TSLA.NAS', $101 = 'UK100', $102 = 'UNH.NYSE', $103 = 'US2000', $104 = 'US30', $105 = 'US500', $106 = 'USDCAD', $107 = 'USDCHF', $108 = 'USDCNH', $109 = 'USDJPY', $110 = 'USTEC', $111 = 'WBC.ASX', $112 = 'WES.ASX', $113 = 'WOW.ASX', $114 = 'WPL.ASX', $115 = 'XAUEUR', $116 = 'XAUUSD', $117 = 'XBRUSD', $118 = 'XTIUSD', $119 = '5'
19 23ms 6 3ms 4ms 3ms with rankedmt4 as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors ), last_feed_entry as ( select * from rankedmt4 where r = 1 ), ok_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where status = 'OK' ), earliest_entry_after_ok as ( select m.datafeedname, min(m.eventtimestamp) as eventtimestamp from mt4datafeederrors m left outer join ( select datafeedname, eventtimestamp from ok_entries where r = 1) oo on m.datafeedname = oo.datafeedname where m.eventtimestamp > coalesce(oo.eventtimestamp, '1900-01-01'::timestamp without time zone) group by m.datafeedname ), notified_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where notified is not null and notified <> '' ), broker as ( select *, row_number() over (partition by feedname order by brokerid) r from ( select distinct b.brokerid, b.name as brokername, dss.classname as feedname from downloadersymbolsettings dss inner join brokersymbollist bsl on dss.symbolid = bsl.symbolid inner join broker b on bsl.brokerid = b.brokerid where dss.enabled = 1) a ) select last.id, last.datafeedname, last.eventtimestamp, last.status, last.errordescription, last.serveraddress, last.username, note.notified, note.eventtimestamp, broker.brokername from last_feed_entry last inner join earliest_entry_after_ok after_ok on last.datafeedname = after_ok.datafeedname inner join broker on last.datafeedname = broker.feedname left outer join ok_entries ok on ok.datafeedname = last.datafeedname left outer join notified_entries note on note.datafeedname = last.datafeedname and note.r = 1 where (ok.r is null or ok.r = 1) and last.datafeedname not in ( select distinct datafeedname from last_feed_entry where status = 'OK') and extract(epoch from (last.eventtimestamp - after_ok.eventtimestamp)) > 60 * 60 and last.eventtimestamp > current_timestamp - interval '1 day' and (note.eventtimestamp is null or note.eventtimestamp < current_timestamp - interval '10 hours') and last.eventtimestamp > current_timestamp - interval '1 hour' and broker.r = 1;Times Reported Time consuming bind #19
Day Hour Count Duration Avg duration 11 6 23ms 3ms -
with rankedmt4 as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors ), last_feed_entry as ( select * from rankedmt4 where r = 1 ), ok_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where status = 'OK' ), earliest_entry_after_ok as ( select m.datafeedname, min(m.eventtimestamp) as eventtimestamp from mt4datafeederrors m left outer join ( select datafeedname, eventtimestamp from ok_entries where r = 1) oo on m.datafeedname = oo.datafeedname where m.eventtimestamp > coalesce(oo.eventtimestamp, '1900-01-01'::timestamp without time zone) group by m.datafeedname ), notified_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where notified is not null and notified <> '' ), broker as ( select *, row_number() over (partition by feedname order by brokerid) r from ( select distinct b.brokerid, b.name as brokername, dss.classname as feedname from downloadersymbolsettings dss inner join brokersymbollist bsl on dss.symbolid = bsl.symbolid inner join broker b on bsl.brokerid = b.brokerid where dss.enabled = 1) a ) select last.id, last.datafeedname, last.eventtimestamp, last.status, last.errordescription, last.serveraddress, last.username, note.notified, note.eventtimestamp, broker.brokername from last_feed_entry last inner join earliest_entry_after_ok after_ok on last.datafeedname = after_ok.datafeedname inner join broker on last.datafeedname = broker.feedname left outer join ok_entries ok on ok.datafeedname = last.datafeedname left outer join notified_entries note on note.datafeedname = last.datafeedname and note.r = 1 where (ok.r is null or ok.r = 1) and last.datafeedname not in ( select distinct datafeedname from last_feed_entry where status = 'OK') and extract(epoch from (last.eventtimestamp - after_ok.eventtimestamp)) > 60 * 60 and last.eventtimestamp > current_timestamp - interval '1 day' and (note.eventtimestamp is null or note.eventtimestamp < current_timestamp - interval '10 hours') and last.eventtimestamp > current_timestamp - interval '1 hour' and broker.r = 1;
Date: 2026-01-05 11:40:02 Duration: 4ms Database: postgres
-
with rankedmt4 as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors ), last_feed_entry as ( select * from rankedmt4 where r = 1 ), ok_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where status = 'OK' ), earliest_entry_after_ok as ( select m.datafeedname, min(m.eventtimestamp) as eventtimestamp from mt4datafeederrors m left outer join ( select datafeedname, eventtimestamp from ok_entries where r = 1) oo on m.datafeedname = oo.datafeedname where m.eventtimestamp > coalesce(oo.eventtimestamp, '1900-01-01'::timestamp without time zone) group by m.datafeedname ), notified_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where notified is not null and notified <> '' ), broker as ( select *, row_number() over (partition by feedname order by brokerid) r from ( select distinct b.brokerid, b.name as brokername, dss.classname as feedname from downloadersymbolsettings dss inner join brokersymbollist bsl on dss.symbolid = bsl.symbolid inner join broker b on bsl.brokerid = b.brokerid where dss.enabled = 1) a ) select last.id, last.datafeedname, last.eventtimestamp, last.status, last.errordescription, last.serveraddress, last.username, note.notified, note.eventtimestamp, broker.brokername from last_feed_entry last inner join earliest_entry_after_ok after_ok on last.datafeedname = after_ok.datafeedname inner join broker on last.datafeedname = broker.feedname left outer join ok_entries ok on ok.datafeedname = last.datafeedname left outer join notified_entries note on note.datafeedname = last.datafeedname and note.r = 1 where (ok.r is null or ok.r = 1) and last.datafeedname not in ( select distinct datafeedname from last_feed_entry where status = 'OK') and extract(epoch from (last.eventtimestamp - after_ok.eventtimestamp)) > 60 * 60 and last.eventtimestamp > current_timestamp - interval '1 day' and (note.eventtimestamp is null or note.eventtimestamp < current_timestamp - interval '10 hours') and last.eventtimestamp > current_timestamp - interval '1 hour' and broker.r = 1;
Date: 2026-01-05 11:30:02 Duration: 3ms Database: postgres
-
with rankedmt4 as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors ), last_feed_entry as ( select * from rankedmt4 where r = 1 ), ok_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where status = 'OK' ), earliest_entry_after_ok as ( select m.datafeedname, min(m.eventtimestamp) as eventtimestamp from mt4datafeederrors m left outer join ( select datafeedname, eventtimestamp from ok_entries where r = 1) oo on m.datafeedname = oo.datafeedname where m.eventtimestamp > coalesce(oo.eventtimestamp, '1900-01-01'::timestamp without time zone) group by m.datafeedname ), notified_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where notified is not null and notified <> '' ), broker as ( select *, row_number() over (partition by feedname order by brokerid) r from ( select distinct b.brokerid, b.name as brokername, dss.classname as feedname from downloadersymbolsettings dss inner join brokersymbollist bsl on dss.symbolid = bsl.symbolid inner join broker b on bsl.brokerid = b.brokerid where dss.enabled = 1) a ) select last.id, last.datafeedname, last.eventtimestamp, last.status, last.errordescription, last.serveraddress, last.username, note.notified, note.eventtimestamp, broker.brokername from last_feed_entry last inner join earliest_entry_after_ok after_ok on last.datafeedname = after_ok.datafeedname inner join broker on last.datafeedname = broker.feedname left outer join ok_entries ok on ok.datafeedname = last.datafeedname left outer join notified_entries note on note.datafeedname = last.datafeedname and note.r = 1 where (ok.r is null or ok.r = 1) and last.datafeedname not in ( select distinct datafeedname from last_feed_entry where status = 'OK') and extract(epoch from (last.eventtimestamp - after_ok.eventtimestamp)) > 60 * 60 and last.eventtimestamp > current_timestamp - interval '1 day' and (note.eventtimestamp is null or note.eventtimestamp < current_timestamp - interval '10 hours') and last.eventtimestamp > current_timestamp - interval '1 hour' and broker.r = 1;
Date: 2026-01-05 11:50:03 Duration: 3ms Database: postgres
20 18ms 238 0ms 0ms 0ms /*server.CPResult*/ SELECT patternid, resy0, resy1, supporty0, supporty1, predictiontimeto, patternstarttime, s.symbolid, resx0, resx1, supportx0, supportx1, symbol, longname, shortname, timegranularity, patternendtime, pattern, a.direction, trendchange, patternlengthbars, patternquality, resultuid as uid, breakout, initialtrend, volumeincrease, symmetry as uniformity, predictionpricefrom, predictionpriceto, noise, exchange, breakout, dtt.absolutetimezoneoffset as tzOs, dtt.timezone as tz FROM autochartist_results a INNER JOIN downloadersymbolsettings dss on a.symbolid = dss.symbolid INNER JOIN datafeedstimetable dtt ON dss.classname = dtt.classname inner join symbols s on a.symbolid = s.symbolid inner join patterns p on p.patternname = a.pattern where resultuid = $1;Times Reported Time consuming bind #20
Day Hour Count Duration Avg duration 11 238 18ms 0ms -
/*server.CPResult*/ SELECT patternid, resy0, resy1, supporty0, supporty1, predictiontimeto, patternstarttime, s.symbolid, resx0, resx1, supportx0, supportx1, symbol, longname, shortname, timegranularity, patternendtime, pattern, a.direction, trendchange, patternlengthbars, patternquality, resultuid as uid, breakout, initialtrend, volumeincrease, symmetry as uniformity, predictionpricefrom, predictionpriceto, noise, exchange, breakout, dtt.absolutetimezoneoffset as tzOs, dtt.timezone as tz FROM autochartist_results a INNER JOIN downloadersymbolsettings dss on a.symbolid = dss.symbolid INNER JOIN datafeedstimetable dtt ON dss.classname = dtt.classname inner join symbols s on a.symbolid = s.symbolid inner join patterns p on p.patternname = a.pattern where resultuid = $1;
Date: 2026-01-05 11:42:11 Duration: 0ms Database: postgres parameters: $1 = '607449992070181301'
-
/*server.CPResult*/ SELECT patternid, resy0, resy1, supporty0, supporty1, predictiontimeto, patternstarttime, s.symbolid, resx0, resx1, supportx0, supportx1, symbol, longname, shortname, timegranularity, patternendtime, pattern, a.direction, trendchange, patternlengthbars, patternquality, resultuid as uid, breakout, initialtrend, volumeincrease, symmetry as uniformity, predictionpricefrom, predictionpriceto, noise, exchange, breakout, dtt.absolutetimezoneoffset as tzOs, dtt.timezone as tz FROM autochartist_results a INNER JOIN downloadersymbolsettings dss on a.symbolid = dss.symbolid INNER JOIN datafeedstimetable dtt ON dss.classname = dtt.classname inner join symbols s on a.symbolid = s.symbolid inner join patterns p on p.patternname = a.pattern where resultuid = $1;
Date: 2026-01-05 11:30:45 Duration: 0ms Database: postgres parameters: $1 = '607425242760606301'
-
/*server.CPResult*/ SELECT patternid, resy0, resy1, supporty0, supporty1, predictiontimeto, patternstarttime, s.symbolid, resx0, resx1, supportx0, supportx1, symbol, longname, shortname, timegranularity, patternendtime, pattern, a.direction, trendchange, patternlengthbars, patternquality, resultuid as uid, breakout, initialtrend, volumeincrease, symmetry as uniformity, predictionpricefrom, predictionpriceto, noise, exchange, breakout, dtt.absolutetimezoneoffset as tzOs, dtt.timezone as tz FROM autochartist_results a INNER JOIN downloadersymbolsettings dss on a.symbolid = dss.symbolid INNER JOIN datafeedstimetable dtt ON dss.classname = dtt.classname inner join symbols s on a.symbolid = s.symbolid inner join patterns p on p.patternname = a.pattern where resultuid = $1;
Date: 2026-01-05 11:37:03 Duration: 0ms Database: postgres parameters: $1 = '607450345911688301'
-
Events
Log levels
Key values
- 423,835 Log entries
Events distribution
Key values
- 0 PANIC entries
- 0 FATAL entries
- 2 ERROR entries
- 0 WARNING entries
Most Frequent Errors/Events
Key values
- 1 Max number of times the same event was reported
- 2 Total events found
Rank Times reported Error 1 1 ERROR: relation "..." does not exist
Times Reported Most Frequent Error / Event #1
Day Hour Count Jan 05 11 1 - ERROR: relation "t0" does not exist at character 83
Statement: SELECT * FROM ( SELECT PriceDateTime, Open, High, Low, Close, Volume, BSF FROM T0 WHERE symbolid = $1 AND (BSF = 0 OR BSF IS NULL) ORDER BY PriceDateTime DESC LIMIT 1050 ) a ORDER BY PriceDateTime ASC
Date: 2026-01-05 11:19:49
2 1 LOG: process ... still waiting for ShareLock on transaction ... after ... ms
Times Reported Most Frequent Error / Event #2
Day Hour Count Jan 05 11 1 - LOG: process 27554 still waiting for ShareLock on transaction 771860617 after 1000.048 ms
Detail: Process holding the lock: 27553. Wait queue: 27554.
Context: while inserting index tuple (15064,61) in relation "t60"
Statement: INSERT INTO T60 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1,$2,$3,$4,$5,$6,$7,$8,$9,$10) ON CONFLICT (pricedatetime,symbolid) DO UPDATE SET open=$11, high=$12, low=$13, close=$14, volume=$15, bsf=$16, sastdatetimewritten=$17, sastdatetimereceived=$18Date: 2026-01-05 11:06:15