-
Global information
- Generated on Mon Jan 26 02:59:56 2026
- Log file: /home/postgres/pg_data/data/pg_log/postgresql-2026-01-26_040000.log
- Parsed 2,646,141 log entries in 55s
- Log start from 2026-01-26 04:00:00 to 2026-01-26 04:59:50
-
Overview
Global Stats
- 254 Number of unique normalized queries
- 254,182 Number of queries
- 1h34m31s Total query duration
- 2026-01-26 04:00:00 First query
- 2026-01-26 04:59:50 Last query
- 1,915 queries/s at 2026-01-26 04:00:04 Query peak
- 1h34m31s Total query duration
- 25s149ms Prepare/parse total duration
- 1m27s Bind total duration
- 1h32m38s Execute total duration
- 36 Number of events
- 3 Number of unique normalized events
- 34 Max number of times the same event was reported
- 0 Number of cancellation
- 40 Total number of automatic vacuums
- 57 Total number of automatic analyzes
- 596 Number temporary file
- 385.39 MiB Max size of temporary file
- 9.62 MiB Average size of temporary file
- 9,976 Total number of sessions
- 16 sessions at 2026-01-26 04:52:10 Session peak
- 1d22h12m30s Total duration of sessions
- 16s675ms Average duration of sessions
- 25 Average queries per session
- 568ms Average queries duration per session
- 16s106ms Average idle time per session
- 9,976 Total number of connections
- 48 connections/s at 2026-01-26 04:23:38 Connection peak
- 3 Total number of databases
SQL Traffic
Key values
- 1,915 queries/s Query Peak
- 2026-01-26 04:00:04 Date
SELECT Traffic
Key values
- 946 queries/s Query Peak
- 2026-01-26 04:00:04 Date
INSERT/UPDATE/DELETE Traffic
Key values
- 216 queries/s Query Peak
- 2026-01-26 04:00:53 Date
Queries duration
Key values
- 1h34m31s Total query duration
Prepared queries ratio
Key values
- 0.00 Ratio of bind vs prepare
- 0.00 % Ratio between prepared and "usual" statements
General Activity
↑ Back to the top of the General Activity tableDay Hour Count Min duration Max duration Avg duration Latency Percentile(90) Latency Percentile(95) Latency Percentile(99) Jan 26 04 254,182 0ms 32s277ms 21ms 3m25s 3m44s 4m21s Day Hour SELECT COPY TO Average Duration Latency Percentile(90) Latency Percentile(95) Latency Percentile(99) Jan 26 04 70,362 26 0ms 0ms 0ms 0ms Day Hour INSERT UPDATE DELETE COPY FROM Average Duration Latency Percentile(90) Latency Percentile(95) Latency Percentile(99) Jan 26 04 28,770 1,884 16 96 0ms 0ms 0ms 0ms Day Hour Prepare Bind Bind/Prepare Percentage of prepare Jan 26 04 60,301 99,986 1.66 55.47% Day Hour Count Average / Second Jan 26 04 9,976 2.77/s Day Hour Count Average Duration Average idle time Jan 26 04 9,976 16s675ms 16s117ms -
Connections
Established Connections
Key values
- 48 connections Connection Peak
- 2026-01-26 04:23:38 Date
Connections per database
Key values
- acaweb_fx Main Database
- 9,976 connections Total
Connections per user
Key values
- postgres Main User
- 9,976 connections Total
Connections per host
Key values
- 192.168.1.15 Main host with 4423 connections
- 9,976 Total connections
Host Count 127.0.0.1 111 192.168.0.114 3 192.168.0.216 101 192.168.0.74 3,010 192.168.1.145 102 192.168.1.15 4,423 192.168.1.20 135 192.168.1.239 4 192.168.1.90 64 192.168.2.126 62 192.168.2.182 12 192.168.2.82 48 192.168.3.199 36 192.168.4.142 1,133 192.168.4.144 1 192.168.4.145 4 192.168.4.150 10 192.168.4.238 14 192.168.4.33 90 192.168.4.41 6 192.168.4.98 330 [local] 277 -
Sessions
Simultaneous sessions
Key values
- 16 sessions Session Peak
- 2026-01-26 04:52:10 Date
Histogram of session times
Key values
- 8,702 0-500ms duration
Sessions per database
Key values
- acaweb_fx Main Database
- 9,976 sessions Total
Sessions per user
Key values
- postgres Main User
- 9,976 sessions Total
Sessions per host
Key values
- 192.168.1.15 Main Host
- 9,976 sessions Total
Host Count Total Duration Average Duration 127.0.0.1 111 13s141ms 118ms 192.168.0.114 3 15m4s 5m1s 192.168.0.216 101 1m2s 616ms 192.168.0.74 3,010 7h6m24s 8s499ms 192.168.1.145 102 1h47m22s 1m3s 192.168.1.15 4,423 2h44m4s 2s225ms 192.168.1.20 135 13h47m3s 6m7s 192.168.1.239 4 203ms 50ms 192.168.1.90 64 35s607ms 556ms 192.168.2.126 62 6s545ms 105ms 192.168.2.182 12 833ms 69ms 192.168.2.82 48 13s255ms 276ms 192.168.3.199 36 1s451ms 40ms 192.168.4.142 1,133 10m28s 554ms 192.168.4.144 1 252ms 252ms 192.168.4.145 4 22s644ms 5s661ms 192.168.4.150 10 20h15m1s 2h1m30s 192.168.4.238 14 15s908ms 1s136ms 192.168.4.33 90 1m13s 821ms 192.168.4.41 6 68ms 11ms 192.168.4.98 330 16s199ms 49ms [local] 277 2m39s 575ms -
Checkpoints / Restartpoints
Checkpoints Buffers
Key values
- 14,912 buffers Checkpoint Peak
- 2026-01-26 04:07:38 Date
- 210.070 seconds Highest write time
- 0.129 seconds Sync time
Checkpoints Wal files
Key values
- 8 files Wal files usage Peak
- 2026-01-26 04:07:38 Date
Checkpoints distance
Key values
- 239.70 Mo Distance Peak
- 2026-01-26 04:07:38 Date
Checkpoints Activity
↑ Back to the top of the Checkpoint Activity tableDay Hour Written buffers Write time Sync time Total time Jan 26 04 42,885 1,697.638s 0.161s 1,698.109s Day Hour Added Removed Recycled Synced files Longest sync Average sync Jan 26 04 0 0 22 1,807 0.124s 0.001s Day Hour Count Avg time (sec) Jan 26 04 0 0s Day Hour Mean distance Mean estimate Jan 26 04 29,927.92 kB 86,998.08 kB -
Temporary Files
Size of temporary files
Key values
- 385.39 MiB Temp Files size Peak
- 2026-01-26 04:05:08 Date
Number of temporary files
Key values
- 30 per second Temp Files Peak
- 2026-01-26 04:47:08 Date
Temporary Files Activity
↑ Back to the top of the Temporary Files Activity tableDay Hour Count Total size Average size Jan 26 04 596 5.60 GiB 9.62 MiB Queries generating the most temporary files (N)
Rank Count Total size Min size Max size Avg size Query 1 29 1.65 GiB 4.16 MiB 171.21 MiB 58.13 MiB with rankedmt4 as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors ), last_feed_entry as ( select * from rankedmt4 where r = ? ), ok_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where status = ? ), earliest_entry_after_ok as ( select m.datafeedname, min(m.eventtimestamp) as eventtimestamp from mt4datafeederrors m left outer join ( select datafeedname, eventtimestamp from ok_entries where r = ?) oo on m.datafeedname = oo.datafeedname where m.eventtimestamp > coalesce(oo.eventtimestamp, ?::timestamp without time zone) group by m.datafeedname ), notified_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where notified is not null and notified <> ? ), broker as ( select *, row_number() over (partition by feedname order by brokerid) r from ( select distinct b.brokerid, b.name as brokername, dss.classname as feedname from downloadersymbolsettings dss inner join brokersymbollist bsl on dss.symbolid = bsl.symbolid inner join broker b on bsl.brokerid = b.brokerid where dss.enabled = ?) a ) select last.id, last.datafeedname, last.eventtimestamp, last.status, last.errordescription, last.serveraddress, last.username, note.notified, note.eventtimestamp, broker.brokername from last_feed_entry last inner join earliest_entry_after_ok after_ok on last.datafeedname = after_ok.datafeedname inner join broker on last.datafeedname = broker.feedname left outer join ok_entries ok on ok.datafeedname = last.datafeedname left outer join notified_entries note on note.datafeedname = last.datafeedname and note.r = ? where (ok.r is null or ok.r = ?) and last.datafeedname not in ( select distinct datafeedname from last_feed_entry where status = ?) and extract(epoch from (last.eventtimestamp - after_ok.eventtimestamp)) > ? * ? and last.eventtimestamp > current_timestamp - interval ? and (note.eventtimestamp is null or note.eventtimestamp < current_timestamp - interval ?) and last.eventtimestamp > current_timestamp - interval ? and broker.r = ?;-
with rankedmt4 as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors ), last_feed_entry as ( select * from rankedmt4 where r = 1 ), ok_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where status = 'OK' ), earliest_entry_after_ok as ( select m.datafeedname, min(m.eventtimestamp) as eventtimestamp from mt4datafeederrors m left outer join ( select datafeedname, eventtimestamp from ok_entries where r = 1) oo on m.datafeedname = oo.datafeedname where m.eventtimestamp > coalesce(oo.eventtimestamp, '1900-01-01'::timestamp without time zone) group by m.datafeedname ), notified_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where notified is not null and notified <> '' ), broker as ( select *, row_number() over (partition by feedname order by brokerid) r from ( select distinct b.brokerid, b.name as brokername, dss.classname as feedname from downloadersymbolsettings dss inner join brokersymbollist bsl on dss.symbolid = bsl.symbolid inner join broker b on bsl.brokerid = b.brokerid where dss.enabled = 1) a ) select last.id, last.datafeedname, last.eventtimestamp, last.status, last.errordescription, last.serveraddress, last.username, note.notified, note.eventtimestamp, broker.brokername from last_feed_entry last inner join earliest_entry_after_ok after_ok on last.datafeedname = after_ok.datafeedname inner join broker on last.datafeedname = broker.feedname left outer join ok_entries ok on ok.datafeedname = last.datafeedname left outer join notified_entries note on note.datafeedname = last.datafeedname and note.r = 1 where (ok.r is null or ok.r = 1) and last.datafeedname not in ( select distinct datafeedname from last_feed_entry where status = 'OK') and extract(epoch from (last.eventtimestamp - after_ok.eventtimestamp)) > 60 * 60 and last.eventtimestamp > current_timestamp - interval '1 day' and (note.eventtimestamp is null or note.eventtimestamp < current_timestamp - interval '10 hours') and last.eventtimestamp > current_timestamp - interval '1 hour' and broker.r = 1;
Date: 2026-01-26 04:00:08 Duration: 0ms
2 24 175.25 MiB 7.29 MiB 7.31 MiB 7.30 MiB jr.resultuid as resultuid, jr.direction as direction, jr.patternendtime as identified, jr.patternlengthbars as length, jr.patternstarttime as patternstarttime, case when jr.trendchangeid = ? then ? else ? end as trendchange, s.exchange as exchange, s.symbolid as symbolid, coalesce(bim.code, s.symbol) as symbol_code, s.longname as symbol_name, s.timegranularity as interval, jp.patternname as pattern_name, dtt.timezone as timezone, ? as age, cps.pip, g.basegroupname from japsticks_results jr inner join brokersymbollist bsl on bsl.brokerid = ? and bsl.symbolid = jr.symbolid inner join relevance_japsticks_results rar on rar.resultuid = jr.resultuid inner join symbols s on jr.symbolid = s.symbolid and s.nonliquid = ? inner join japsticks_patterns jp on jr.patternid = jp.id inner join downloadersymbolsettings dss on jr.symbolid = dss.symbolid inner join datafeedstimetable dtt on dss.classname = dtt.classname and dtt.dayofweek = ? inner join symbolgroup sg on s.symbolid = sg.symbolid inner join groups g on sg.groupid = g.groupid inner join brokergroups bg on g.groupid = bg.groupid and bsl.brokerid = bg.brokerid left join currencypips cps on cps.symbol = s.symbol left outer join brokerinstrumentmap bim on dss.datafeedinstrumentid = bim.datafeedinstrumentid and bim.brokerid = bsl.brokerid and bim.type = ? where jr.gmttimefound > now() - interval ? and s.deleted = ? and (jr.simulation = ? or jr.simulation is null) and (rar.relevant = ?) --and (semicolon_age = ? or rar.age <= semicolon_age) and (? = ? or s.timegranularity in (...)) and (? = ? or s.exchange in (...)) and (? = ? or coalesce(bim.code, s.symbol) in (...)) and (? = ? or jp.patternname in (...)) and (? = ? or jr.patternlengthbars <= ?) ), results as ( select distinct on (symbolid) * from all_results order by symbolid, resultuid ) select * from results order by identified desc, length desc ;-
jr.resultuid AS resultuid, jr.direction AS direction, jr.patternendtime AS identified, jr.patternlengthbars AS length, jr.patternstarttime AS patternstarttime, case when jr.trendchangeid = 1 then 'Continuation' else 'Reversal' end AS trendchange, s.exchange AS exchange, s.symbolid AS symbolid, coalesce(bim.code, s.symbol) AS symbol_code, s.longname AS symbol_name, s.timegranularity AS interval, jp.patternname AS pattern_name, dtt.timezone AS timezone, 0 AS age, cps.pip, g.basegroupname FROM japsticks_results jr INNER JOIN brokersymbollist bsl ON bsl.brokerid = $1 AND bsl.symbolid = jr.symbolid INNER JOIN relevance_japsticks_results rar ON rar.resultuid = jr.resultuid INNER JOIN symbols s ON jr.symbolid = s.symbolid AND s.nonliquid = 0 INNER JOIN japsticks_patterns jp ON jr.patternid = jp.id INNER JOIN downloadersymbolsettings dss ON jr.symbolid = dss.symbolid INNER JOIN datafeedstimetable dtt ON dss.classname = dtt.classname AND dtt.dayofweek = 3 INNER JOIN symbolgroup sg on s.symbolid = sg.symbolid INNER JOIN groups g ON sg.groupid = g.groupid INNER JOIN brokergroups bg on g.groupid = bg.groupid AND bsl.brokerid = bg.brokerid LEFT JOIN currencypips cps on cps.symbol = s.symbol LEFT OUTER JOIN brokerinstrumentmap bim ON dss.datafeedinstrumentid = bim.datafeedinstrumentid AND bim.brokerid = bsl.brokerid AND bim.TYPE = 'OUTBOUND' WHERE jr.gmttimefound > now() - INTERVAL '7 DAYS' AND s.deleted = 0 AND (jr.simulation = 0 OR jr.simulation IS NULL) AND (rar.relevant = 1) --AND (semicolon_age = 0 OR rar.age <= semicolon_age) AND ($2 = 0 OR s.timegranularity in ($3)) AND ($4 = 0 OR s.exchange in ($5)) AND ($6 = 0 OR coalesce(bim.code, s.symbol) in ($7)) AND ($8 = 0 OR jp.patternname in ($9)) AND ($10 = 0 OR jr.patternlengthbars <= $11)), results AS ( SELECT DISTINCT ON (symbolid) * FROM all_results ORDER BY symbolid, resultuid ) SELECT * from results ORDER BY identified DESC, length DESC;
Date: 2026-01-26 04:00:33 Duration: 0ms
3 16 737.25 MiB 46.08 MiB 46.08 MiB 46.08 MiB update solr_relevance_old set new_hod_correct = sub.hod_correct, new_hod_percent = sub.hod_percent, new_hod_total = sub.hod_total, new_pattern_correct = sub.pattern_correct, new_pattern_percent = sub.pattern_percent, new_pattern_total = sub.pattern_total, new_percent = sub.percent, new_symbol_correct = sub.symbol_correct, new_symbol_percent = sub.symbol_percent, new_symbol_total = sub.symbol_total from ( select distinct resultuid, hod_correct, hod_percent, hod_total, hod, pattern_correct, pattern_percent, pattern_total, percent, symbol_correct, symbol_percent, symbol_total from whatshot_probability where type = ?) sub where result_uid = sub.resultuid;-
UPDATE solr_relevance_old SET new_hod_correct = sub.hod_correct, new_hod_percent = sub.hod_percent, new_hod_total = sub.hod_total, new_pattern_correct = sub.pattern_correct, new_pattern_percent = sub.pattern_percent, new_pattern_total = sub.pattern_total, new_percent = sub.percent, new_symbol_correct = sub.symbol_correct, new_symbol_percent = sub.symbol_percent, new_symbol_total = sub.symbol_total FROM ( select distinct resultuid, hod_correct, hod_percent, hod_total, hod, pattern_correct, pattern_percent, pattern_total, percent, symbol_correct, symbol_percent, symbol_total FROM whatshot_probability WHERE type = 'cp') sub WHERE result_uid = sub.resultuid;
Date: 2026-01-26 04:01:13 Duration: 0ms
4 16 1.22 GiB 78.19 MiB 78.19 MiB 78.19 MiB with max_ra as ( select resultuid from relevance_keylevels_results order by resultuid desc limit ?) update solr_relevance_old set newrelevant = sub.relevant, newage = sub.age from ( select so.uuid, case when ra.relevant is not null then ra.relevant when so.result_uid < max_ra.resultuid then ? else ? end as relevant, case when ra.age is not null then ra.age when so.result_uid < max_ra.resultuid then ? else ? end as age, so.result_uid from max_ra, solr_relevance_old so inner join keylevels_results k on so.result_uid = k.resultuid and so.uuid ilike ? inner join downloadersymbolsettings dss on k.symbolid = dss.symbolid left outer join relevance_keylevels_results ra on so.result_uid = ra.resultuid and so.uuid ilike ?) sub where solr_relevance_old.result_uid = sub.result_uid and solr_relevance_old.uuid ilike ?; update solr_relevance_old set newrelevant = ? where result_uid in ( select result_uid from solr_relevance_old s left outer join keylevels_results a on a.resultuid = s.result_uid where s.uuid ilike ? and a.resultuid is null); update solr_relevance_old set new_hod_correct = sub.hod_correct, new_hod_percent = sub.hod_percent, new_hod_total = sub.hod_total, new_pattern_correct = sub.pattern_correct, new_pattern_percent = sub.pattern_percent, new_pattern_total = sub.pattern_total, new_percent = sub.percent, new_symbol_correct = sub.symbol_correct, new_symbol_percent = sub.symbol_percent, new_symbol_total = sub.symbol_total from ( select distinct resultuid, hod_correct, hod_percent, hod_total, hod, pattern_correct, pattern_percent, pattern_total, percent, symbol_correct, symbol_percent, symbol_total from whatshot_probability where type in (...)) sub where result_uid = sub.resultuid;-
with max_ra as ( select resultuid from relevance_keylevels_results order by resultuid desc limit 1) update solr_relevance_old set newrelevant = sub.relevant, newage = sub.age from ( select so.uuid, case when ra.relevant is not null then ra.relevant when so.result_uid < max_ra.resultuid then 0 else 1 end as relevant, case when ra.age is not null then ra.age when so.result_uid < max_ra.resultuid then 11 else 0 end as age, so.result_uid from max_ra, solr_relevance_old so inner join keylevels_results k on so.result_uid = k.resultuid and so.uuid ilike 'kl_%' inner join downloadersymbolsettings dss on k.symbolid = dss.symbolid left outer join relevance_keylevels_results ra on so.result_uid = ra.resultuid and so.uuid ilike 'kl_%') sub where solr_relevance_old.result_uid = sub.result_uid and solr_relevance_old.uuid ilike 'kl_%'; update solr_relevance_old set newrelevant = 0 where result_uid in ( select result_uid from solr_relevance_old s left outer join keylevels_results a on a.resultuid = s.result_uid where s.uuid ilike 'kl_%' and a.resultuid is null); UPDATE solr_relevance_old SET new_hod_correct = sub.hod_correct, new_hod_percent = sub.hod_percent, new_hod_total = sub.hod_total, new_pattern_correct = sub.pattern_correct, new_pattern_percent = sub.pattern_percent, new_pattern_total = sub.pattern_total, new_percent = sub.percent, new_symbol_correct = sub.symbol_correct, new_symbol_percent = sub.symbol_percent, new_symbol_total = sub.symbol_total FROM ( select distinct resultuid, hod_correct, hod_percent, hod_total, hod, pattern_correct, pattern_percent, pattern_total, percent, symbol_correct, symbol_percent, symbol_total FROM whatshot_probability WHERE type in ('kl', 'ekl')) sub WHERE result_uid = sub.resultuid;
Date: 2026-01-26 04:01:15 Duration: 0ms
5 8 1.02 GiB 130.38 MiB 130.42 MiB 130.39 MiB select updateresultsmaterializedview ();-
select updateresultsmaterializedview ();
Date: 2026-01-26 04:02:15 Duration: 0ms
6 4 355.86 MiB 88.92 MiB 89.02 MiB 88.96 MiB select updateageforrelevantresults ();-
select updateageforrelevantresults ();
Date: 2026-01-26 04:02:05 Duration: 0ms
Queries generating the largest temporary files
Rank Size Query 1 171.21 MiB with rankedmt4 as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors ), last_feed_entry as ( select * from rankedmt4 where r = 1 ), ok_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where status = 'OK' ), earliest_entry_after_ok as ( select m.datafeedname, min(m.eventtimestamp) as eventtimestamp from mt4datafeederrors m left outer join ( select datafeedname, eventtimestamp from ok_entries where r = 1) oo on m.datafeedname = oo.datafeedname where m.eventtimestamp > coalesce(oo.eventtimestamp, '1900-01-01'::timestamp without time zone) group by m.datafeedname ), notified_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where notified is not null and notified <> '' ), broker as ( select *, row_number() over (partition by feedname order by brokerid) r from ( select distinct b.brokerid, b.name as brokername, dss.classname as feedname from downloadersymbolsettings dss inner join brokersymbollist bsl on dss.symbolid = bsl.symbolid inner join broker b on bsl.brokerid = b.brokerid where dss.enabled = 1) a ) select last.id, last.datafeedname, last.eventtimestamp, last.status, last.errordescription, last.serveraddress, last.username, note.notified, note.eventtimestamp, broker.brokername from last_feed_entry last inner join earliest_entry_after_ok after_ok on last.datafeedname = after_ok.datafeedname inner join broker on last.datafeedname = broker.feedname left outer join ok_entries ok on ok.datafeedname = last.datafeedname left outer join notified_entries note on note.datafeedname = last.datafeedname and note.r = 1 where (ok.r is null or ok.r = 1) and last.datafeedname not in ( select distinct datafeedname from last_feed_entry where status = 'OK') and extract(epoch from (last.eventtimestamp - after_ok.eventtimestamp)) > 60 * 60 and last.eventtimestamp > current_timestamp - interval '1 day' and (note.eventtimestamp is null or note.eventtimestamp < current_timestamp - interval '10 hours') and last.eventtimestamp > current_timestamp - interval '1 hour' and broker.r = 1;[ Date: 2026-01-26 04:50:05 ]
2 159.90 MiB with rankedmt4 as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors ), last_feed_entry as ( select * from rankedmt4 where r = 1 ), ok_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where status = 'OK' ), earliest_entry_after_ok as ( select m.datafeedname, min(m.eventtimestamp) as eventtimestamp from mt4datafeederrors m left outer join ( select datafeedname, eventtimestamp from ok_entries where r = 1) oo on m.datafeedname = oo.datafeedname where m.eventtimestamp > coalesce(oo.eventtimestamp, '1900-01-01'::timestamp without time zone) group by m.datafeedname ), notified_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where notified is not null and notified <> '' ), broker as ( select *, row_number() over (partition by feedname order by brokerid) r from ( select distinct b.brokerid, b.name as brokername, dss.classname as feedname from downloadersymbolsettings dss inner join brokersymbollist bsl on dss.symbolid = bsl.symbolid inner join broker b on bsl.brokerid = b.brokerid where dss.enabled = 1) a ) select last.id, last.datafeedname, last.eventtimestamp, last.status, last.errordescription, last.serveraddress, last.username, note.notified, note.eventtimestamp, broker.brokername from last_feed_entry last inner join earliest_entry_after_ok after_ok on last.datafeedname = after_ok.datafeedname inner join broker on last.datafeedname = broker.feedname left outer join ok_entries ok on ok.datafeedname = last.datafeedname left outer join notified_entries note on note.datafeedname = last.datafeedname and note.r = 1 where (ok.r is null or ok.r = 1) and last.datafeedname not in ( select distinct datafeedname from last_feed_entry where status = 'OK') and extract(epoch from (last.eventtimestamp - after_ok.eventtimestamp)) > 60 * 60 and last.eventtimestamp > current_timestamp - interval '1 day' and (note.eventtimestamp is null or note.eventtimestamp < current_timestamp - interval '10 hours') and last.eventtimestamp > current_timestamp - interval '1 hour' and broker.r = 1;[ Date: 2026-01-26 04:30:05 ]
3 159.41 MiB with rankedmt4 as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors ), last_feed_entry as ( select * from rankedmt4 where r = 1 ), ok_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where status = 'OK' ), earliest_entry_after_ok as ( select m.datafeedname, min(m.eventtimestamp) as eventtimestamp from mt4datafeederrors m left outer join ( select datafeedname, eventtimestamp from ok_entries where r = 1) oo on m.datafeedname = oo.datafeedname where m.eventtimestamp > coalesce(oo.eventtimestamp, '1900-01-01'::timestamp without time zone) group by m.datafeedname ), notified_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where notified is not null and notified <> '' ), broker as ( select *, row_number() over (partition by feedname order by brokerid) r from ( select distinct b.brokerid, b.name as brokername, dss.classname as feedname from downloadersymbolsettings dss inner join brokersymbollist bsl on dss.symbolid = bsl.symbolid inner join broker b on bsl.brokerid = b.brokerid where dss.enabled = 1) a ) select last.id, last.datafeedname, last.eventtimestamp, last.status, last.errordescription, last.serveraddress, last.username, note.notified, note.eventtimestamp, broker.brokername from last_feed_entry last inner join earliest_entry_after_ok after_ok on last.datafeedname = after_ok.datafeedname inner join broker on last.datafeedname = broker.feedname left outer join ok_entries ok on ok.datafeedname = last.datafeedname left outer join notified_entries note on note.datafeedname = last.datafeedname and note.r = 1 where (ok.r is null or ok.r = 1) and last.datafeedname not in ( select distinct datafeedname from last_feed_entry where status = 'OK') and extract(epoch from (last.eventtimestamp - after_ok.eventtimestamp)) > 60 * 60 and last.eventtimestamp > current_timestamp - interval '1 day' and (note.eventtimestamp is null or note.eventtimestamp < current_timestamp - interval '10 hours') and last.eventtimestamp > current_timestamp - interval '1 hour' and broker.r = 1;[ Date: 2026-01-26 04:10:06 ]
4 156.33 MiB with rankedmt4 as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors ), last_feed_entry as ( select * from rankedmt4 where r = 1 ), ok_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where status = 'OK' ), earliest_entry_after_ok as ( select m.datafeedname, min(m.eventtimestamp) as eventtimestamp from mt4datafeederrors m left outer join ( select datafeedname, eventtimestamp from ok_entries where r = 1) oo on m.datafeedname = oo.datafeedname where m.eventtimestamp > coalesce(oo.eventtimestamp, '1900-01-01'::timestamp without time zone) group by m.datafeedname ), notified_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where notified is not null and notified <> '' ), broker as ( select *, row_number() over (partition by feedname order by brokerid) r from ( select distinct b.brokerid, b.name as brokername, dss.classname as feedname from downloadersymbolsettings dss inner join brokersymbollist bsl on dss.symbolid = bsl.symbolid inner join broker b on bsl.brokerid = b.brokerid where dss.enabled = 1) a ) select last.id, last.datafeedname, last.eventtimestamp, last.status, last.errordescription, last.serveraddress, last.username, note.notified, note.eventtimestamp, broker.brokername from last_feed_entry last inner join earliest_entry_after_ok after_ok on last.datafeedname = after_ok.datafeedname inner join broker on last.datafeedname = broker.feedname left outer join ok_entries ok on ok.datafeedname = last.datafeedname left outer join notified_entries note on note.datafeedname = last.datafeedname and note.r = 1 where (ok.r is null or ok.r = 1) and last.datafeedname not in ( select distinct datafeedname from last_feed_entry where status = 'OK') and extract(epoch from (last.eventtimestamp - after_ok.eventtimestamp)) > 60 * 60 and last.eventtimestamp > current_timestamp - interval '1 day' and (note.eventtimestamp is null or note.eventtimestamp < current_timestamp - interval '10 hours') and last.eventtimestamp > current_timestamp - interval '1 hour' and broker.r = 1;[ Date: 2026-01-26 04:00:08 ]
5 130.42 MiB select updateresultsmaterializedview ();[ Date: 2026-01-26 04:02:15 ]
6 130.40 MiB select updateresultsmaterializedview ();[ Date: 2026-01-26 04:47:13 ]
7 130.40 MiB select updateresultsmaterializedview ();[ Date: 2026-01-26 04:32:14 ]
8 130.39 MiB select updateresultsmaterializedview ();[ Date: 2026-01-26 04:50:32 ]
9 130.39 MiB select updateresultsmaterializedview ();[ Date: 2026-01-26 04:35:32 ]
10 130.39 MiB select updateresultsmaterializedview ();[ Date: 2026-01-26 04:17:14 ]
11 130.38 MiB select updateresultsmaterializedview ();[ Date: 2026-01-26 04:05:32 ]
12 130.38 MiB select updateresultsmaterializedview ();[ Date: 2026-01-26 04:20:32 ]
13 112.13 MiB with rankedmt4 as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors ), last_feed_entry as ( select * from rankedmt4 where r = 1 ), ok_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where status = 'OK' ), earliest_entry_after_ok as ( select m.datafeedname, min(m.eventtimestamp) as eventtimestamp from mt4datafeederrors m left outer join ( select datafeedname, eventtimestamp from ok_entries where r = 1) oo on m.datafeedname = oo.datafeedname where m.eventtimestamp > coalesce(oo.eventtimestamp, '1900-01-01'::timestamp without time zone) group by m.datafeedname ), notified_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where notified is not null and notified <> '' ), broker as ( select *, row_number() over (partition by feedname order by brokerid) r from ( select distinct b.brokerid, b.name as brokername, dss.classname as feedname from downloadersymbolsettings dss inner join brokersymbollist bsl on dss.symbolid = bsl.symbolid inner join broker b on bsl.brokerid = b.brokerid where dss.enabled = 1) a ) select last.id, last.datafeedname, last.eventtimestamp, last.status, last.errordescription, last.serveraddress, last.username, note.notified, note.eventtimestamp, broker.brokername from last_feed_entry last inner join earliest_entry_after_ok after_ok on last.datafeedname = after_ok.datafeedname inner join broker on last.datafeedname = broker.feedname left outer join ok_entries ok on ok.datafeedname = last.datafeedname left outer join notified_entries note on note.datafeedname = last.datafeedname and note.r = 1 where (ok.r is null or ok.r = 1) and last.datafeedname not in ( select distinct datafeedname from last_feed_entry where status = 'OK') and extract(epoch from (last.eventtimestamp - after_ok.eventtimestamp)) > 60 * 60 and last.eventtimestamp > current_timestamp - interval '1 day' and (note.eventtimestamp is null or note.eventtimestamp < current_timestamp - interval '10 hours') and last.eventtimestamp > current_timestamp - interval '1 hour' and broker.r = 1;[ Date: 2026-01-26 04:40:04 ]
14 98.30 MiB with rankedmt4 as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors ), last_feed_entry as ( select * from rankedmt4 where r = 1 ), ok_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where status = 'OK' ), earliest_entry_after_ok as ( select m.datafeedname, min(m.eventtimestamp) as eventtimestamp from mt4datafeederrors m left outer join ( select datafeedname, eventtimestamp from ok_entries where r = 1) oo on m.datafeedname = oo.datafeedname where m.eventtimestamp > coalesce(oo.eventtimestamp, '1900-01-01'::timestamp without time zone) group by m.datafeedname ), notified_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where notified is not null and notified <> '' ), broker as ( select *, row_number() over (partition by feedname order by brokerid) r from ( select distinct b.brokerid, b.name as brokername, dss.classname as feedname from downloadersymbolsettings dss inner join brokersymbollist bsl on dss.symbolid = bsl.symbolid inner join broker b on bsl.brokerid = b.brokerid where dss.enabled = 1) a ) select last.id, last.datafeedname, last.eventtimestamp, last.status, last.errordescription, last.serveraddress, last.username, note.notified, note.eventtimestamp, broker.brokername from last_feed_entry last inner join earliest_entry_after_ok after_ok on last.datafeedname = after_ok.datafeedname inner join broker on last.datafeedname = broker.feedname left outer join ok_entries ok on ok.datafeedname = last.datafeedname left outer join notified_entries note on note.datafeedname = last.datafeedname and note.r = 1 where (ok.r is null or ok.r = 1) and last.datafeedname not in ( select distinct datafeedname from last_feed_entry where status = 'OK') and extract(epoch from (last.eventtimestamp - after_ok.eventtimestamp)) > 60 * 60 and last.eventtimestamp > current_timestamp - interval '1 day' and (note.eventtimestamp is null or note.eventtimestamp < current_timestamp - interval '10 hours') and last.eventtimestamp > current_timestamp - interval '1 hour' and broker.r = 1;[ Date: 2026-01-26 04:20:04 ]
15 97.13 MiB with rankedmt4 as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors ), last_feed_entry as ( select * from rankedmt4 where r = 1 ), ok_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where status = 'OK' ), earliest_entry_after_ok as ( select m.datafeedname, min(m.eventtimestamp) as eventtimestamp from mt4datafeederrors m left outer join ( select datafeedname, eventtimestamp from ok_entries where r = 1) oo on m.datafeedname = oo.datafeedname where m.eventtimestamp > coalesce(oo.eventtimestamp, '1900-01-01'::timestamp without time zone) group by m.datafeedname ), notified_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where notified is not null and notified <> '' ), broker as ( select *, row_number() over (partition by feedname order by brokerid) r from ( select distinct b.brokerid, b.name as brokername, dss.classname as feedname from downloadersymbolsettings dss inner join brokersymbollist bsl on dss.symbolid = bsl.symbolid inner join broker b on bsl.brokerid = b.brokerid where dss.enabled = 1) a ) select last.id, last.datafeedname, last.eventtimestamp, last.status, last.errordescription, last.serveraddress, last.username, note.notified, note.eventtimestamp, broker.brokername from last_feed_entry last inner join earliest_entry_after_ok after_ok on last.datafeedname = after_ok.datafeedname inner join broker on last.datafeedname = broker.feedname left outer join ok_entries ok on ok.datafeedname = last.datafeedname left outer join notified_entries note on note.datafeedname = last.datafeedname and note.r = 1 where (ok.r is null or ok.r = 1) and last.datafeedname not in ( select distinct datafeedname from last_feed_entry where status = 'OK') and extract(epoch from (last.eventtimestamp - after_ok.eventtimestamp)) > 60 * 60 and last.eventtimestamp > current_timestamp - interval '1 day' and (note.eventtimestamp is null or note.eventtimestamp < current_timestamp - interval '10 hours') and last.eventtimestamp > current_timestamp - interval '1 hour' and broker.r = 1;[ Date: 2026-01-26 04:20:04 ]
16 89.02 MiB select updateageforrelevantresults ();[ Date: 2026-01-26 04:02:05 ]
17 88.96 MiB select updateageforrelevantresults ();[ Date: 2026-01-26 04:32:05 ]
18 88.95 MiB select updateageforrelevantresults ();[ Date: 2026-01-26 04:47:04 ]
19 88.92 MiB select updateageforrelevantresults ();[ Date: 2026-01-26 04:17:05 ]
20 78.19 MiB with max_ra as ( select resultuid from relevance_keylevels_results order by resultuid desc limit 1) update solr_relevance_old set newrelevant = sub.relevant, newage = sub.age from ( select so.uuid, case when ra.relevant is not null then ra.relevant when so.result_uid < max_ra.resultuid then 0 else 1 end as relevant, case when ra.age is not null then ra.age when so.result_uid < max_ra.resultuid then 11 else 0 end as age, so.result_uid from max_ra, solr_relevance_old so inner join keylevels_results k on so.result_uid = k.resultuid and so.uuid ilike 'kl_%' inner join downloadersymbolsettings dss on k.symbolid = dss.symbolid left outer join relevance_keylevels_results ra on so.result_uid = ra.resultuid and so.uuid ilike 'kl_%') sub where solr_relevance_old.result_uid = sub.result_uid and solr_relevance_old.uuid ilike 'kl_%'; update solr_relevance_old set newrelevant = 0 where result_uid in ( select result_uid from solr_relevance_old s left outer join keylevels_results a on a.resultuid = s.result_uid where s.uuid ilike 'kl_%' and a.resultuid is null); UPDATE solr_relevance_old SET new_hod_correct = sub.hod_correct, new_hod_percent = sub.hod_percent, new_hod_total = sub.hod_total, new_pattern_correct = sub.pattern_correct, new_pattern_percent = sub.pattern_percent, new_pattern_total = sub.pattern_total, new_percent = sub.percent, new_symbol_correct = sub.symbol_correct, new_symbol_percent = sub.symbol_percent, new_symbol_total = sub.symbol_total FROM ( select distinct resultuid, hod_correct, hod_percent, hod_total, hod, pattern_correct, pattern_percent, pattern_total, percent, symbol_correct, symbol_percent, symbol_total FROM whatshot_probability WHERE type in ('kl', 'ekl')) sub WHERE result_uid = sub.resultuid;[ Date: 2026-01-26 04:01:15 ]
-
Vacuums
Vacuums / Analyzes Distribution
Key values
- 0 sec Highest CPU-cost vacuum
Table
Database - Date
- 0 sec Highest CPU-cost analyze
Table
Database - Date
Analyzes per table
Key values
- public.solr_relevance_old (16) Main table analyzed (database acaweb_fx)
- 57 analyzes Total
Table Number of analyzes acaweb_fx.public.solr_relevance_old 16 acaweb_fx.pg_catalog.pg_attribute 5 acaweb_fx.public.datafeeds_latestrun 4 acaweb_fx.public.relevance_keylevels_results 4 acaweb_fx.pg_catalog.pg_class 4 acaweb_fx.public.relevance_fibonacci_results 4 acaweb_fx.public.relevance_autochartist_results 4 acaweb_fx.pg_catalog.pg_type 3 acaweb_fx.pg_catalog.pg_index 2 acaweb_fx.public.latest_t15_candle_view 2 acaweb_fx.public.autochartist_symbolupdates 1 acaweb_fx.public.solr_imports 1 socialmedia.public.processes 1 acaweb_fx.pg_catalog.pg_depend 1 acaweb_fx.public.latest_candle_datetime_per_receng 1 acaweb_fx.public.sa_hist_bigmove 1 acaweb_fx.public.relevance_consecutivecandles_results 1 acaweb_fx.public.symbollatestupdatetime 1 acaweb_fx.public.bigmovement_results_underlying 1 Total 57 Vacuums per table
Key values
- public.solr_relevance_old (16) Main table vacuumed on database acaweb_fx
- 40 vacuums Total
Index Buffer usage Skipped WAL usage Table Vacuums scans hits misses dirtied pins frozen records full page bytes acaweb_fx.public.solr_relevance_old 16 16 11,153 0 53 0 0 8,165 16 1,492,912 acaweb_fx.public.datafeeds_latestrun 4 0 437 0 36 0 0 56 32 68,636 acaweb_fx.pg_toast.pg_toast_2619 2 2 288 0 68 0 0 231 66 277,144 acaweb_fx.pg_catalog.pg_attribute 2 2 1,623 0 357 0 134 745 272 1,643,423 acaweb_fx.public.latest_t15_candle_view 2 2 159 0 2 0 0 12 2 18,158 acaweb_fx.public.relevance_fibonacci_results 2 2 2,501 0 58 1 120 365 43 110,524 acaweb_fx.public.relevance_autochartist_results 2 2 6,692 0 280 2 496 1,344 266 655,742 acaweb_fx.pg_catalog.pg_index 1 1 89 0 12 0 0 28 11 80,699 acaweb_fx.public.autochartist_symbolupdates 1 1 23,024 0 2,920 10 38,252 7,068 2,827 1,118,926 acaweb_fx.pg_catalog.pg_depend 1 1 384 0 80 0 59 172 64 347,185 acaweb_fx.public.relevance_keylevels_results 1 1 3,986 0 184 2 81 1,126 176 432,372 acaweb_fx.pg_catalog.pg_class 1 1 472 0 44 0 0 163 42 217,061 acaweb_fx.public.sa_hist_bigmove 1 1 4,191 0 854 0 0 1,502 851 2,585,514 acaweb_fx.pg_catalog.pg_type 1 1 132 0 27 0 0 52 22 131,707 acaweb_fx.pg_catalog.pg_statistic 1 1 1,056 0 177 0 551 497 169 607,614 acaweb_fx.public.relevance_consecutivecandles_results 1 1 77 0 7 0 0 27 4 22,311 acaweb_fx.public.symbollatestupdatetime 1 1 1,884 0 138 0 638 1,007 118 437,607 Total 40 36 58,148 48,790 5,297 15 40,331 22,560 4,981 10,247,535 Tuples removed per table
Key values
- public.solr_relevance_old (67070) Main table with removed tuples on database acaweb_fx
- 109958 tuples Total removed
Index Tuples Pages Table Vacuums scans removed remain not yet removable removed remain acaweb_fx.public.solr_relevance_old 16 16 67,070 80,440 0 0 2,652 acaweb_fx.public.symbollatestupdatetime 1 1 17,890 93,865 784 0 1,714 acaweb_fx.public.sa_hist_bigmove 1 1 12,380 48,792 0 0 1,100 acaweb_fx.public.autochartist_symbolupdates 1 1 4,578 47,281 336 0 40,691 acaweb_fx.pg_catalog.pg_attribute 2 2 2,943 21,858 0 38 480 acaweb_fx.public.relevance_autochartist_results 2 2 1,584 14,854 0 0 760 acaweb_fx.public.relevance_keylevels_results 1 1 1,167 10,650 0 0 279 acaweb_fx.pg_catalog.pg_depend 1 1 652 14,650 0 0 135 acaweb_fx.pg_catalog.pg_statistic 1 1 538 3,614 0 0 1,194 acaweb_fx.public.relevance_fibonacci_results 2 2 273 2,557 0 0 204 acaweb_fx.public.datafeeds_latestrun 4 0 240 75 19 0 64 acaweb_fx.pg_toast.pg_toast_2619 2 2 163 348 10 0 105 acaweb_fx.pg_catalog.pg_class 1 1 147 1,649 0 0 150 acaweb_fx.public.latest_t15_candle_view 2 2 126 28 0 0 2 acaweb_fx.pg_catalog.pg_type 1 1 122 1,446 0 0 39 acaweb_fx.public.relevance_consecutivecandles_results 1 1 68 307 0 0 8 acaweb_fx.pg_catalog.pg_index 1 1 17 813 0 0 22 Total 40 36 109,958 343,227 1,149 38 49,599 Pages removed per table
Key values
- pg_catalog.pg_attribute (38) Main table with removed pages on database acaweb_fx
- 38 pages Total removed
Table Number of vacuums Index scans Tuples removed Pages removed acaweb_fx.pg_catalog.pg_attribute 2 2 2943 38 acaweb_fx.pg_catalog.pg_index 1 1 17 0 acaweb_fx.pg_toast.pg_toast_2619 2 2 163 0 acaweb_fx.public.autochartist_symbolupdates 1 1 4578 0 acaweb_fx.public.datafeeds_latestrun 4 0 240 0 acaweb_fx.pg_catalog.pg_depend 1 1 652 0 acaweb_fx.public.latest_t15_candle_view 2 2 126 0 acaweb_fx.public.relevance_keylevels_results 1 1 1167 0 acaweb_fx.pg_catalog.pg_class 1 1 147 0 acaweb_fx.public.relevance_fibonacci_results 2 2 273 0 acaweb_fx.public.sa_hist_bigmove 1 1 12380 0 acaweb_fx.pg_catalog.pg_type 1 1 122 0 acaweb_fx.pg_catalog.pg_statistic 1 1 538 0 acaweb_fx.public.relevance_consecutivecandles_results 1 1 68 0 acaweb_fx.public.symbollatestupdatetime 1 1 17890 0 acaweb_fx.public.relevance_autochartist_results 2 2 1584 0 acaweb_fx.public.solr_relevance_old 16 16 67070 0 Total 40 36 109,958 38 Autovacuum Activity
↑ Back to the top of the Autovacuum Activity tableDay Hour VACUUMs ANALYZEs Jan 26 04 40 57 - 0 sec Highest CPU-cost vacuum
-
Locks
Locks by types
Key values
- unknown Main Lock Type
- 0 locks Total
Most frequent waiting queries (N)
Rank Count Total time Min time Max time Avg duration Query NO DATASET
Queries that waited the most
Rank Wait time Query NO DATASET
-
Queries
Queries by type
Key values
- 70,362 Total read queries
- 38,353 Total write queries
Queries by database
Key values
- unknown Main database
- 253,170 Requests
- 1h32m38s (unknown)
- Main time consuming database
Database Request type Count Duration acaweb_fx Total 921 0ms copy from 80 0ms copy to 26 0ms cte 104 0ms ddl 16 0ms delete 16 0ms others 206 0ms select 104 0ms tcl 332 0ms update 37 0ms socialmedia Total 91 0ms select 90 0ms tcl 1 0ms unknown Total 253,170 1h32m38s copy from 16 0ms cte 6,747 0ms insert 28,770 0ms others 18,187 0ms select 70,168 0ms tcl 361 0ms update 1,847 0ms Queries by user
Key values
- unknown Main user
- 253,170 Requests
User Request type Count Duration postgres Total 1,012 0ms copy from 80 0ms copy to 26 0ms cte 104 0ms ddl 16 0ms delete 16 0ms others 206 0ms select 194 0ms tcl 333 0ms update 37 0ms unknown Total 253,170 1h32m38s copy from 16 0ms cte 6,747 0ms insert 28,770 0ms others 18,187 0ms select 70,168 0ms tcl 361 0ms update 1,847 0ms Duration by user
Key values
- 1h32m38s (unknown) Main time consuming user
User Request type Count Duration postgres Total 1,012 0ms copy from 80 0ms copy to 26 0ms cte 104 0ms ddl 16 0ms delete 16 0ms others 206 0ms select 194 0ms tcl 333 0ms update 37 0ms unknown Total 253,170 1h32m38s copy from 16 0ms cte 6,747 0ms insert 28,770 0ms others 18,187 0ms select 70,168 0ms tcl 361 0ms update 1,847 0ms Queries by host
Key values
- unknown Main host
- 254,182 Requests
- 1h32m38s (unknown)
- Main time consuming host
Queries by application
Key values
- unknown Main application
- 253,794 Requests
- 1h32m38s (unknown)
- Main time consuming application
Number of cancelled queries
Key values
- 0 per second Cancelled query Peak
- 2026-01-26 04:04:25 Date
Number of cancelled queries (5 minutes period)
NO DATASET
-
Top Queries
Histogram of query times
Key values
- 84,959 0-1ms duration
Slowest individual queries
Rank Duration Query NO DATASET
Time consuming queries
Rank Total duration Times executed Min duration Max duration Avg duration Query 1 0ms 1 0ms 0ms 0ms insert into t30 (symbolid, pricedatetime, open, high, low, close, volume, bsf, sastdatetimereceived) values (?, ?::timestamp without time zone, ?.?, ?.?, ?.?, ?, ?, ?, ?::timestamp without time zone) on conflict (symbolid, pricedatetime) do nothing;Times Reported Time consuming queries #1
Day Hour Count Duration Avg duration Jan 26 04 1 0ms 0ms 2 0ms 30 0ms 0ms 0ms select key, value from datasources ds inner join datasourceparams dsp on ds.id = dsp.datasourceid where ds.name = ?;Times Reported Time consuming queries #2
Day Hour Count Duration Avg duration Jan 26 04 30 0ms 0ms 3 0ms 224 0ms 0ms 0ms with rar_max as ( select resultuid from relevance_consecutivecandles_results order by resultuid desc limit ? ), all_results as ( select ccr.resultuid as resultuid, ccr.direction as direction, s.exchange as exchange, s.symbolid as symbolid, coalesce(bim.code, s.symbol) as symbol_code, s.longname as symbol_name, s.timegranularity as interval, ccr.patternendtime as identified, dtt.timezone as timezone, ccr.qtyconsecutivecandles as length, g.basegroupname, case when rcr.age is not null then rcr.age when ccr.resultuid <= rm.resultuid then ? else ? end as age, case when rcr.relevant is not null then rcr.relevant when ccr.resultuid <= rm.resultuid then ? else ? end as relevant, cps.pip, newlevels.filtered from consecutivecandles_results ccr inner join brokersymbollist bsl on bsl.brokerid = ? and bsl.symbolid = ccr.symbolid inner join symbols s on ccr.symbolid = s.symbolid and s.nonliquid = ? inner join downloadersymbolsettings dss on ccr.symbolid = dss.symbolid inner join datafeedstimetable dtt on dss.classname = dtt.classname and dtt.dayofweek = ? inner join symbolgroup sg on ccr.symbolid = sg.symbolid inner join groups g on sg.groupid = g.groupid inner join brokergroups bg on g.groupid = bg.groupid and bsl.brokerid = bg.brokerid inner join rar_max rm on ? = ? left outer join relevance_consecutivecandles_results rcr on rcr.resultuid = ccr.resultuid left join currencypips cps on cps.symbol = s.symbol left outer join brokerinstrumentmap bim on dss.datafeedinstrumentid = bim.datafeedinstrumentid and bim.brokerid = bsl.brokerid and bim.type = ? left join lateral calc_cc_signal_filter (ccr.resultuid) newlevels on true where ccr.gmttimefound > now() - interval ? and s.deleted = ? and (ccr.simulation = ? or ccr.simulation is null) and (? = ? or s.timegranularity in (...)) and (? = ? or s.exchange in (...)) and (? = ? or coalesce(bim.code, s.symbol) in (...)) and (? = ? or ccr.patternlengthbars <= ?) ), results as ( select distinct on (symbolid) * from all_results where (false = ? or relevant = ?) and (? = ? or age <= ?) order by symbolid, resultuid ) select * from results order by identified desc, length desc;Times Reported Time consuming queries #3
Day Hour Count Duration Avg duration Jan 26 04 224 0ms 0ms 4 0ms 3 0ms 0ms 0ms select trumpetsymbolid as sid, trumpettimegranularity as tg from brokersymbollist bsl left join powerstats_symboldata psd on bsl.symbolid = psd.symbolid left join downloadersymbolsettings dss on psd.symbolid = dss.symbolid left join symbols s on dss.symbolid = s.symbolid left outer join brokerinstrumentmap bdfi on code = ? and bdfi.brokerid = ? and dss.datafeedinstrumentid = bdfi.datafeedinstrumentid where (code = ? or s.symbol = ?) and bsl.brokerid = ? and dss.classname <> ? group by trumpetsymbolid, trumpettimegranularity;Times Reported Time consuming queries #4
Day Hour Count Duration Avg duration Jan 26 04 3 0ms 0ms 5 0ms 1 0ms 0ms 0ms insert into t15 (symbolid, pricedatetime, open, high, low, close, volume, bsf, sastdatetimereceived) values (?, ?::timestamp without time zone, ?, ?, ?, ?.?, ?, ?, ?::timestamp without time zone) on conflict (symbolid, pricedatetime) do nothing;Times Reported Time consuming queries #5
Day Hour Count Duration Avg duration Jan 26 04 1 0ms 0ms 6 0ms 200 0ms 0ms 0ms with rar_max as ( select resultuid from relevance_bigmovement_results order by resultuid desc limit ? ) select bmr.symbolid, patternstarttime, patternendtime, timegranularity, ? as direction, case when bmr.old_resultuid = ? then bmr.old_resultuid else bmr.resultuid end as uid, s.exchange, s.symbol, s.longname, s.shortname, dtt.timezone, bmr.patternmovement, bmr.statisticalmovement, bmr.fromprice, bmr.toprice, bmr.percentile, bmr.patternlengthbars, case when rbr.age is not null then rbr.age when bmr.resultuid <= rm.resultuid then ? else ? end as age, case when rbr.relevant is not null then rbr.relevant when bmr.resultuid <= rm.resultuid then ? else ? end as relevant, cps.pip from bigmovement_results bmr inner join downloadersymbolsettings dss on bmr.symbolid = dss.symbolid inner join datafeedstimetable dtt on dss.classname = dtt.classname inner join symbols s on bmr.symbolid = s.symbolid inner join rar_max rm on ? = ? left outer join relevance_bigmovement_results rbr on rbr.resultuid = bmr.resultuid left join currencypips cps on cps.symbol = s.symbol where (bmr.old_resultuid = ? or bmr.resultuid = ?) and dtt.dayofweek = ?;Times Reported Time consuming queries #6
Day Hour Count Duration Avg duration Jan 26 04 200 0ms 0ms 7 0ms 2,220 0ms 0ms 0ms insert into t60 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) values (?, ?, ?, ?, ?, ?, ?, ?, ?, ?) on conflict (pricedatetime, symbolid) do update set open = ?, high = ?, low = ?, close = ?, volume = ?, bsf = ?, sastdatetimewritten = ?, sastdatetimereceived = ?;Times Reported Time consuming queries #7
Day Hour Count Duration Avg duration Jan 26 04 2,220 0ms 0ms 8 0ms 48 0ms 0ms 0ms select count(*) from datafeeds_latestrun where feedname ilike ? and ((latestrxtime > current_timestamp - interval ? and latestdbwritetime > current_timestamp - interval ?) or (latestdbwritetime > current_timestamp - interval ? and lateststartuptime > current_timestamp - interval ?));Times Reported Time consuming queries #8
Day Hour Count Duration Avg duration Jan 26 04 48 0ms 0ms 9 0ms 1 0ms 0ms 0ms insert into t60 (symbolid, pricedatetime, open, high, low, close, volume, bsf, sastdatetimereceived) values (?, ?::timestamp without time zone, ?.?, ?.?, ?, ?, ?, ?, ?::timestamp without time zone) on conflict (symbolid, pricedatetime) do nothing;Times Reported Time consuming queries #9
Day Hour Count Duration Avg duration Jan 26 04 1 0ms 0ms 10 0ms 4 0ms 0ms 0ms select updaterelevantforrelevantresults ();Times Reported Time consuming queries #10
Day Hour Count Duration Avg duration Jan 26 04 4 0ms 0ms 11 0ms 62 0ms 0ms 0ms select case when a.old_resultuid = ? then a.old_resultuid else a.resultuid end as resultuid, s.symbol, timegranularity as interval, direction as direction, patternendtime as patternendtime, patternstartprice as psp, patternendprice as pep, target03 as t03, target16 as t16, patternlengthbars as length, p.patternname as patternname, dtt.timezone, cps.pip from fibonacci_results a inner join downloadersymbolsettings dss on a.symbolid = dss.symbolid inner join datafeedstimetable dtt on dss.classname = dtt.classname inner join symbols s on a.symbolid = s.symbolid inner join fibonaccipatterns p on a.pattern = p.patternname left join currencypips cps on cps.symbol = s.symbol where (a.old_resultuid = ? or a.resultuid = ?) and dtt.dayofweek = ?;Times Reported Time consuming queries #11
Day Hour Count Duration Avg duration Jan 26 04 62 0ms 0ms 12 0ms 1 0ms 0ms 0ms select count(*) from autochartist_stocklist where base_engine = recognitionengine and base_engine ilike ?;Times Reported Time consuming queries #12
Day Hour Count Duration Avg duration Jan 26 04 1 0ms 0ms 13 0ms 1,298 0ms 0ms 0ms update patternresultsrelevance set relevant = ?, saxo_relevant = ?, notrelevantpricedatetime = ?, reason = ? where uniqueindex = ? and relevant = ?;Times Reported Time consuming queries #13
Day Hour Count Duration Avg duration Jan 26 04 1,298 0ms 0ms 14 0ms 2 0ms 0ms 0ms update correlating_signals set filtered = true where id in (...) returning id;Times Reported Time consuming queries #14
Day Hour Count Duration Avg duration Jan 26 04 2 0ms 0ms 15 0ms 1 0ms 0ms 0ms insert into t15 (symbolid, pricedatetime, open, high, low, close, volume, bsf, sastdatetimereceived) values (?, ?::timestamp without time zone, ?.?, ?, ?.?, ?, ?, ?, ?::timestamp without time zone) on conflict (symbolid, pricedatetime) do nothing;Times Reported Time consuming queries #15
Day Hour Count Duration Avg duration Jan 26 04 1 0ms 0ms 16 0ms 240 0ms 0ms 0ms select v.datname, c.relname, v.phase, v.heap_blks_total, v.heap_blks_scanned, v.heap_blks_vacuumed, v.index_vacuum_count, v.max_dead_tuples, v.num_dead_tuples from pg_stat_progress_vacuum as v join pg_class c on c.oid = v.relid;Times Reported Time consuming queries #16
Day Hour Count Duration Avg duration Jan 26 04 240 0ms 0ms 17 0ms 18 0ms 0ms 0ms select cast(count(*) / cast(setting as numeric) * ? as int) from pg_stat_activity, pg_settings where name = ? group by setting;Times Reported Time consuming queries #17
Day Hour Count Duration Avg duration Jan 26 04 18 0ms 0ms 18 0ms 65 0ms 0ms 0ms select symbolid, pricedatetime, classname, downloadfrequency, downloadersymbol, open, high, low, close, volume, bsf, sastdatetimereceived from ( select pricedatetime, dss.classname, dss.downloadfrequency, dss.symbolid, dss.downloadersymbol, t.open, t.high, t.low, t.close, t.volume, t.bsf, t.sastdatetimereceived, row_number() over (partition by t.symbolid order by t.pricedatetime desc) as rn from t240 t, downloadersymbolsettings dss, symbols s where dss.classname = ? and dss.downloadfrequency = ? and dss.symbolid = t.symbolid and s.symbolid = dss.symbolid and dss.enabled = ? and s.deleted = ? and dss.downloadersymbol in (...) and t.pricedatetime > now() - interval ?) as ranked_candles_table where rn = ?;Times Reported Time consuming queries #18
Day Hour Count Duration Avg duration Jan 26 04 65 0ms 0ms 19 0ms 1 0ms 0ms 0ms select count(*) from autochartist_stocklist where base_engine = recognitionengine and enabled = ?;Times Reported Time consuming queries #19
Day Hour Count Duration Avg duration Jan 26 04 1 0ms 0ms 20 0ms 11 0ms 0ms 0ms select * from ( select pricedatetime, open, high, low, close, volume, bsf from t60 where symbolid = ? and (bsf = ? or bsf is null) order by pricedatetime desc limit ?) a order by pricedatetime asc;Times Reported Time consuming queries #20
Day Hour Count Duration Avg duration Jan 26 04 11 0ms 0ms Most frequent queries (N)
Rank Times executed Total duration Min duration Max duration Avg duration Query 1 32,080 0ms 0ms 0ms 0ms select ?;Times Reported Time consuming queries #1
Day Hour Count Duration Avg duration Jan 26 04 32,080 0ms 0ms 2 23,186 0ms 0ms 0ms 0ms select distinct on (coalesce(bim.code, s.symbol) , s.exchange, s.timegranularity, df.timezone) s.symbolid as id, coalesce(bim.code, s.symbol) as name, s.symbol as symbol, dss.downloadersymbol as ticker, s.exchange as exchange, s.timegranularity as interval, df.timezone as timezone from symbols s inner join downloadersymbolsettings dss on dss.symbolid = s.symbolid inner join datafeedstimetable df on df.classname ilike dss.classname left join brokersymbollist bsl on bsl.brokerid = ? and bsl.symbolid = s.symbolid left outer join brokerinstrumentmap bim on dss.datafeedinstrumentid = bim.datafeedinstrumentid and bim.brokerid = ? and bim.type = ? where s.symbolid = ?;Times Reported Time consuming queries #2
Day Hour Count Duration Avg duration Jan 26 04 23,186 0ms 0ms 3 8,876 0ms 0ms 0ms 0ms set extra_float_digits = ?;Times Reported Time consuming queries #3
Day Hour Count Duration Avg duration Jan 26 04 8,876 0ms 0ms 4 8,850 0ms 0ms 0ms 0ms set application_name = ?;Times Reported Time consuming queries #4
Day Hour Count Duration Avg duration Jan 26 04 8,850 0ms 0ms 5 7,860 0ms 0ms 0ms 0ms insert into executionlogs (executionid, status, message, details, detailtype) values (null, ?, ?, null, null);Times Reported Time consuming queries #5
Day Hour Count Duration Avg duration Jan 26 04 7,860 0ms 0ms 6 5,558 0ms 0ms 0ms 0ms insert into t15 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) values (?, ?, ?, ?, ?, ?, ?, ?, ?, ?) on conflict (pricedatetime, symbolid) do update set open = ?, high = ?, low = ?, close = ?, volume = ?, bsf = ?, sastdatetimewritten = ?, sastdatetimereceived = ?;Times Reported Time consuming queries #6
Day Hour Count Duration Avg duration Jan 26 04 5,558 0ms 0ms 7 4,023 0ms 0ms 0ms 0ms insert into autochartist_results (resultid, symbolid, bandwidth, pattern, qtytp, gmttimefound, direction, initialtrend, breakout, volumeincrease, noise, symmetry, predictionpricefrom, predictionpriceto, predictiontimefrom, predictiontimeto, patternstarttime, patternendtime, patternstartprice, patternendprice, resx0, resx1, supportx0, supportx1, resy0, resy1, supporty0, supporty1, supportgradient, resgradient, riskreward, patternquality, trendchange, maxmovementafterbreakout, latestbaratbreakouttime, latestbaratbreakoutprice, patternlengthbars, temporarypattern, relevancestartdistance, simulation, writtendatetime) values (?, ?, ?.?, ?, ?, ?::timestamp without time zone, ?, ?.?, ?.?, ?.?, ?.?, ?.?, ?.?, ?.?, ?::timestamp without time zone, ?::timestamp without time zone, ?::timestamp without time zone, ?::timestamp without time zone, ?.?, ?.?, ?::timestamp without time zone, ?::timestamp without time zone, ?::timestamp without time zone, ?::timestamp without time zone, ?.?, ?.?, ?.?, ?.?, ?.?, ?.?, ?.?, ?.?, ?, ?.?, ?::timestamp without time zone, ?.?, ?, ?, ?.?, ?, current_timestamp::timestamp without time zone) on conflict do nothing;Times Reported Time consuming queries #7
Day Hour Count Duration Avg duration Jan 26 04 4,023 0ms 0ms 8 3,247 0ms 0ms 0ms 0ms insert into t30 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) values (?, ?, ?, ?, ?, ?, ?, ?, ?, ?) on conflict (pricedatetime, symbolid) do update set open = ?, high = ?, low = ?, close = ?, volume = ?, bsf = ?, sastdatetimewritten = ?, sastdatetimereceived = ?;Times Reported Time consuming queries #8
Day Hour Count Duration Avg duration Jan 26 04 3,247 0ms 0ms 9 2,656 0ms 0ms 0ms 0ms select s.symbolid as id, s.symbol as name, s.exchange as exchange, s.timegranularity as interval, dtt.timezone as timezone from symbols s inner join downloadersymbolsettings dss on dss.symbolid = s.symbolid inner join datafeedstimetable dtt on dss.classname = dtt.classname and dtt.dayofweek = ? inner join brokersymbollist bsl on bsl.symbolid = s.symbolid where bsl.brokerid = ? and (? = ? or s.timegranularity = ?) and (s.symbol = ? or dss.downloadersymbol = ?) and dss.enabled = ?;Times Reported Time consuming queries #9
Day Hour Count Duration Avg duration Jan 26 04 2,656 0ms 0ms 10 2,586 0ms 0ms 0ms 0ms with rar_max as ( select resultuid from relevance_autochartist_results order by resultuid desc limit ? ) select a.symbolid, pattern, patternid, resy0, resy1, resx0, resx1, supporty0, supporty1, supportx0, supportx1, predictiontimeto, patternstarttime, timegranularity, patternendtime, direction, trendchange, patternlengthbars, patternquality, case when a.old_resultuid = ? then a.old_resultuid else a.resultuid end as uid, breakout, initialtrend, volumeincrease, symmetry as uniformity, predictionpricefrom, predictionpriceto, noise, s.exchange, s.symbol, s.longname, s.shortname, breakout, dtt.timezone, patternstartprice, patternendprice, qtytp, newlevels.profit, newlevels.stop, newlevels.filtered, case when rar.age is not null then rar.age when a.resultuid <= rm.resultuid then ? else ? end as age, case when rar.relevant is not null then rar.relevant when a.resultuid <= rm.resultuid then ? else ? end as relevant, cps.pip from autochartist_results a inner join downloadersymbolsettings dss on a.symbolid = dss.symbolid inner join datafeedstimetable dtt on dss.classname = dtt.classname inner join symbols s on a.symbolid = s.symbolid inner join patterns p on p.patternname = a.pattern inner join rar_max rm on ? = ? left outer join relevance_autochartist_results rar on rar.resultuid = a.resultuid left join lateral calc_cp_signal (a.resultuid) newlevels on true left join currencypips cps on cps.symbol = s.symbol where (a.old_resultuid = ? or a.resultuid = ?) and dtt.dayofweek = ?;Times Reported Time consuming queries #10
Day Hour Count Duration Avg duration Jan 26 04 2,586 0ms 0ms 11 2,220 0ms 0ms 0ms 0ms insert into t60 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) values (?, ?, ?, ?, ?, ?, ?, ?, ?, ?) on conflict (pricedatetime, symbolid) do update set open = ?, high = ?, low = ?, close = ?, volume = ?, bsf = ?, sastdatetimewritten = ?, sastdatetimereceived = ?;Times Reported Time consuming queries #11
Day Hour Count Duration Avg duration Jan 26 04 2,220 0ms 0ms 12 2,067 0ms 0ms 0ms 0ms insert into fibonacci_results (bandwidth, pattern, gmttimefound, direction, patternstarttime, patternendtime, patternstartprice, patternendprice, qtytp, pricex, timex, pricea, timea, priceb, timeb, pricec, timec, priced, timed, averagequality, timequality, errormargin, patternlengthbars, target10, target06, target16, target07, target12, target05, target03, symbolid, noise, ratiosfound, temporarypattern, uniqueindex, completed, simulation, writtendatetime) values (?.?, ?, ?::timestamp without time zone, ?, ?::timestamp without time zone, ?::timestamp without time zone, ?.?, ?.?, ?, ?.?, ?::timestamp without time zone, ?.?, ?::timestamp without time zone, ?.?, ?::timestamp without time zone, ?.?, ?::timestamp without time zone, ?.?, ?::timestamp without time zone, ?.?, ?.?, ?.?, ?, ?.?, ?.?, ?.?, ?.?, ?.?, ?.?, ?.?, ?, ?.?, ?, ?, ?, ?, ?, current_timestamp::timestamp without time zone) on conflict do nothing;Times Reported Time consuming queries #12
Day Hour Count Duration Avg duration Jan 26 04 2,067 0ms 0ms 13 1,408 0ms 0ms 0ms 0ms insert into keylevels_results (bandwidth, breakout, patternid, gmttimefound, approachingtimestamp, approachingregion, qtytp, patternlengthbars, patternprice, x0, x1, x2, x3, x4, x5, x6, x7, x8, x9, breakoutbars, breakoutprice, patternendtime, atbaridentified, atpriceidentified, errormargin, direction, symbolid, predictionpricefrom, predictionpriceto, predictiontimefrom, predictiontimebars, uniquepointsvalue, furthestprice, relevancestartdistance, patternclassid, patternstarttime, stoplosslevel, simulation, writtendatetime) values (?.?, ?, ?, ?::timestamp without time zone, ?, ?.?, ?, ?, ?.?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?.?, ?::timestamp without time zone, ?, ?.?, ?.?, ?, ?, ?.?, ?.?, ?::timestamp without time zone, ?, ?, ?.?, ?.?, ?, ?, ?.?, ?, current_timestamp::timestamp without time zone) on conflict do nothing;Times Reported Time consuming queries #13
Day Hour Count Duration Avg duration Jan 26 04 1,408 0ms 0ms 14 1,405 0ms 0ms 0ms 0ms with rar_max as ( select resultuid from relevance_keylevels_results order by resultuid desc limit ? ) select case when a.old_resultuid = ? then a.old_resultuid else a.resultuid end as ruid, s.symbolid as sid, s.symbol as sym, longname, shortname, exchange as e, timegranularity as tg, a.patternid as pid, a.direction as d, a.patternprice as pp, atbaridentified as pet, case when (x9 != ?) then x9 when (x8 != ?) then x8 when (x7 != ?) then x7 when (x6 != ?) then x6 when (x5 != ?) then x5 when (x4 != ?) then x4 when (x3 != ?) then x3 when (x2 != ?) then x2 end as pst, patternprice as patp, x0, x1, x2, case when (x3 != ?) then x3 else ? end as x3, case when (x4 != ?) then x4 else ? end as x4, case when (x5 != ?) then x5 else ? end as x5, case when (x6 != ?) then x6 else ? end as x6, case when (x7 != ?) then x7 else ? end as x7, case when (x8 != ?) then x8 else ? end as x8, errormargin as erm, breakoutprice as pe, breakoutbars as be, breakout, atbaridentified as atbar, atpriceidentified as atprice, patternlengthbars as l, bandwidth as bw, qtytp as qtp, p.patternname as patternname, dtt.absolutetimezoneoffset as tzos, dtt.timezone as timezone, approachingtimestamp as apt, approachingregion as apr, predictionpricefrom as ppf, predictionpriceto as ppt, predictiontimefrom as ptf, predictiontimebars as ptb, furthestprice as fp, newlevels.filtered, a.uniquepointsvalue as upv, case when rar.age is not null then rar.age when a.resultuid <= rm.resultuid then ? else ? end as age, case when rar.relevant is not null then rar.relevant when a.resultuid <= rm.resultuid then ? else ? end as relevant, cps.pip from keylevels_results a inner join downloadersymbolsettings dss on a.symbolid = dss.symbolid inner join datafeedstimetable dtt on dss.classname = dtt.classname inner join symbols s on a.symbolid = s.symbolid inner join hrspatterns p on a.patternid = p.patternid inner join rar_max rm on ? = ? left outer join relevance_keylevels_results rar on a.resultuid = rar.resultuid left join lateral calc_kl_signal_filter (a.resultuid) newlevels on true left join currencypips cps on cps.symbol = s.symbol where (a.old_resultuid = ? or a.resultuid = ?) and dtt.dayofweek = ?;Times Reported Time consuming queries #14
Day Hour Count Duration Avg duration Jan 26 04 1,405 0ms 0ms 15 1,298 0ms 0ms 0ms 0ms update patternresultsrelevance set relevant = ?, saxo_relevant = ?, notrelevantpricedatetime = ?, reason = ? where uniqueindex = ? and relevant = ?;Times Reported Time consuming queries #15
Day Hour Count Duration Avg duration Jan 26 04 1,298 0ms 0ms 16 1,244 0ms 0ms 0ms 0ms insert into t240 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) values (?, ?, ?, ?, ?, ?, ?, ?, ?, ?) on conflict (pricedatetime, symbolid) do update set open = ?, high = ?, low = ?, close = ?, volume = ?, bsf = ?, sastdatetimewritten = ?, sastdatetimereceived = ?;Times Reported Time consuming queries #16
Day Hour Count Duration Avg duration Jan 26 04 1,244 0ms 0ms 17 928 0ms 0ms 0ms 0ms select symbolid, pricedatetime, classname, downloadfrequency, downloadersymbol, open, high, low, close, volume, bsf, sastdatetimereceived from ( select pricedatetime, dss.classname, dss.downloadfrequency, dss.symbolid, dss.downloadersymbol, t.open, t.high, t.low, t.close, t.volume, t.bsf, t.sastdatetimereceived, row_number() over (partition by t.symbolid order by t.pricedatetime desc) as rn from t15 t, downloadersymbolsettings dss, symbols s where dss.classname = ? and dss.downloadfrequency = ? and dss.symbolid = t.symbolid and s.symbolid = dss.symbolid and dss.enabled = ? and s.deleted = ? and dss.downloadersymbol in (...) and t.pricedatetime > now() - interval ?) as ranked_candles_table where rn = ?;Times Reported Time consuming queries #17
Day Hour Count Duration Avg duration Jan 26 04 928 0ms 0ms 18 492 0ms 0ms 0ms 0ms select t.pricedatetime as pricedatetime, t.open as open, t.high as high, t.low as low, t.close "..." close, t.volume as volume, t.bsf as bsf from t240 t where t.symbolid = ? and (bsf = ? or bsf is null) and pricedatetime >= ? and pricedatetime <= ? order by pricedatetime desc limit ?;Times Reported Time consuming queries #18
Day Hour Count Duration Avg duration Jan 26 04 492 0ms 0ms 19 437 0ms 0ms 0ms 0ms select downloadersymbol, spike_threshold from price_datafeed_spike_threshold where classname = ?;Times Reported Time consuming queries #19
Day Hour Count Duration Avg duration Jan 26 04 437 0ms 0ms 20 437 0ms 0ms 0ms 0ms select s.symbolid, dss.downloadfrequency, dss.downloadersymbol from downloadersymbolsettings dss inner join symbols s on dss.symbolid = s.symbolid where dss.classname = ? and s.deleted = ? and dss.enabled = ?;Times Reported Time consuming queries #20
Day Hour Count Duration Avg duration Jan 26 04 437 0ms 0ms Normalized slowest queries (N)
Rank Min duration Max duration Avg duration Times executed Total duration Query 1 0ms 0ms 0ms 1 0ms insert into t30 (symbolid, pricedatetime, open, high, low, close, volume, bsf, sastdatetimereceived) values (?, ?::timestamp without time zone, ?.?, ?.?, ?.?, ?, ?, ?, ?::timestamp without time zone) on conflict (symbolid, pricedatetime) do nothing;Times Reported Time consuming queries #1
Day Hour Count Duration Avg duration Jan 26 04 1 0ms 0ms 2 0ms 0ms 0ms 30 0ms select key, value from datasources ds inner join datasourceparams dsp on ds.id = dsp.datasourceid where ds.name = ?;Times Reported Time consuming queries #2
Day Hour Count Duration Avg duration Jan 26 04 30 0ms 0ms 3 0ms 0ms 0ms 224 0ms with rar_max as ( select resultuid from relevance_consecutivecandles_results order by resultuid desc limit ? ), all_results as ( select ccr.resultuid as resultuid, ccr.direction as direction, s.exchange as exchange, s.symbolid as symbolid, coalesce(bim.code, s.symbol) as symbol_code, s.longname as symbol_name, s.timegranularity as interval, ccr.patternendtime as identified, dtt.timezone as timezone, ccr.qtyconsecutivecandles as length, g.basegroupname, case when rcr.age is not null then rcr.age when ccr.resultuid <= rm.resultuid then ? else ? end as age, case when rcr.relevant is not null then rcr.relevant when ccr.resultuid <= rm.resultuid then ? else ? end as relevant, cps.pip, newlevels.filtered from consecutivecandles_results ccr inner join brokersymbollist bsl on bsl.brokerid = ? and bsl.symbolid = ccr.symbolid inner join symbols s on ccr.symbolid = s.symbolid and s.nonliquid = ? inner join downloadersymbolsettings dss on ccr.symbolid = dss.symbolid inner join datafeedstimetable dtt on dss.classname = dtt.classname and dtt.dayofweek = ? inner join symbolgroup sg on ccr.symbolid = sg.symbolid inner join groups g on sg.groupid = g.groupid inner join brokergroups bg on g.groupid = bg.groupid and bsl.brokerid = bg.brokerid inner join rar_max rm on ? = ? left outer join relevance_consecutivecandles_results rcr on rcr.resultuid = ccr.resultuid left join currencypips cps on cps.symbol = s.symbol left outer join brokerinstrumentmap bim on dss.datafeedinstrumentid = bim.datafeedinstrumentid and bim.brokerid = bsl.brokerid and bim.type = ? left join lateral calc_cc_signal_filter (ccr.resultuid) newlevels on true where ccr.gmttimefound > now() - interval ? and s.deleted = ? and (ccr.simulation = ? or ccr.simulation is null) and (? = ? or s.timegranularity in (...)) and (? = ? or s.exchange in (...)) and (? = ? or coalesce(bim.code, s.symbol) in (...)) and (? = ? or ccr.patternlengthbars <= ?) ), results as ( select distinct on (symbolid) * from all_results where (false = ? or relevant = ?) and (? = ? or age <= ?) order by symbolid, resultuid ) select * from results order by identified desc, length desc;Times Reported Time consuming queries #3
Day Hour Count Duration Avg duration Jan 26 04 224 0ms 0ms 4 0ms 0ms 0ms 3 0ms select trumpetsymbolid as sid, trumpettimegranularity as tg from brokersymbollist bsl left join powerstats_symboldata psd on bsl.symbolid = psd.symbolid left join downloadersymbolsettings dss on psd.symbolid = dss.symbolid left join symbols s on dss.symbolid = s.symbolid left outer join brokerinstrumentmap bdfi on code = ? and bdfi.brokerid = ? and dss.datafeedinstrumentid = bdfi.datafeedinstrumentid where (code = ? or s.symbol = ?) and bsl.brokerid = ? and dss.classname <> ? group by trumpetsymbolid, trumpettimegranularity;Times Reported Time consuming queries #4
Day Hour Count Duration Avg duration Jan 26 04 3 0ms 0ms 5 0ms 0ms 0ms 1 0ms insert into t15 (symbolid, pricedatetime, open, high, low, close, volume, bsf, sastdatetimereceived) values (?, ?::timestamp without time zone, ?, ?, ?, ?.?, ?, ?, ?::timestamp without time zone) on conflict (symbolid, pricedatetime) do nothing;Times Reported Time consuming queries #5
Day Hour Count Duration Avg duration Jan 26 04 1 0ms 0ms 6 0ms 0ms 0ms 200 0ms with rar_max as ( select resultuid from relevance_bigmovement_results order by resultuid desc limit ? ) select bmr.symbolid, patternstarttime, patternendtime, timegranularity, ? as direction, case when bmr.old_resultuid = ? then bmr.old_resultuid else bmr.resultuid end as uid, s.exchange, s.symbol, s.longname, s.shortname, dtt.timezone, bmr.patternmovement, bmr.statisticalmovement, bmr.fromprice, bmr.toprice, bmr.percentile, bmr.patternlengthbars, case when rbr.age is not null then rbr.age when bmr.resultuid <= rm.resultuid then ? else ? end as age, case when rbr.relevant is not null then rbr.relevant when bmr.resultuid <= rm.resultuid then ? else ? end as relevant, cps.pip from bigmovement_results bmr inner join downloadersymbolsettings dss on bmr.symbolid = dss.symbolid inner join datafeedstimetable dtt on dss.classname = dtt.classname inner join symbols s on bmr.symbolid = s.symbolid inner join rar_max rm on ? = ? left outer join relevance_bigmovement_results rbr on rbr.resultuid = bmr.resultuid left join currencypips cps on cps.symbol = s.symbol where (bmr.old_resultuid = ? or bmr.resultuid = ?) and dtt.dayofweek = ?;Times Reported Time consuming queries #6
Day Hour Count Duration Avg duration Jan 26 04 200 0ms 0ms 7 0ms 0ms 0ms 2,220 0ms insert into t60 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) values (?, ?, ?, ?, ?, ?, ?, ?, ?, ?) on conflict (pricedatetime, symbolid) do update set open = ?, high = ?, low = ?, close = ?, volume = ?, bsf = ?, sastdatetimewritten = ?, sastdatetimereceived = ?;Times Reported Time consuming queries #7
Day Hour Count Duration Avg duration Jan 26 04 2,220 0ms 0ms 8 0ms 0ms 0ms 48 0ms select count(*) from datafeeds_latestrun where feedname ilike ? and ((latestrxtime > current_timestamp - interval ? and latestdbwritetime > current_timestamp - interval ?) or (latestdbwritetime > current_timestamp - interval ? and lateststartuptime > current_timestamp - interval ?));Times Reported Time consuming queries #8
Day Hour Count Duration Avg duration Jan 26 04 48 0ms 0ms 9 0ms 0ms 0ms 1 0ms insert into t60 (symbolid, pricedatetime, open, high, low, close, volume, bsf, sastdatetimereceived) values (?, ?::timestamp without time zone, ?.?, ?.?, ?, ?, ?, ?, ?::timestamp without time zone) on conflict (symbolid, pricedatetime) do nothing;Times Reported Time consuming queries #9
Day Hour Count Duration Avg duration Jan 26 04 1 0ms 0ms 10 0ms 0ms 0ms 4 0ms select updaterelevantforrelevantresults ();Times Reported Time consuming queries #10
Day Hour Count Duration Avg duration Jan 26 04 4 0ms 0ms 11 0ms 0ms 0ms 62 0ms select case when a.old_resultuid = ? then a.old_resultuid else a.resultuid end as resultuid, s.symbol, timegranularity as interval, direction as direction, patternendtime as patternendtime, patternstartprice as psp, patternendprice as pep, target03 as t03, target16 as t16, patternlengthbars as length, p.patternname as patternname, dtt.timezone, cps.pip from fibonacci_results a inner join downloadersymbolsettings dss on a.symbolid = dss.symbolid inner join datafeedstimetable dtt on dss.classname = dtt.classname inner join symbols s on a.symbolid = s.symbolid inner join fibonaccipatterns p on a.pattern = p.patternname left join currencypips cps on cps.symbol = s.symbol where (a.old_resultuid = ? or a.resultuid = ?) and dtt.dayofweek = ?;Times Reported Time consuming queries #11
Day Hour Count Duration Avg duration Jan 26 04 62 0ms 0ms 12 0ms 0ms 0ms 1 0ms select count(*) from autochartist_stocklist where base_engine = recognitionengine and base_engine ilike ?;Times Reported Time consuming queries #12
Day Hour Count Duration Avg duration Jan 26 04 1 0ms 0ms 13 0ms 0ms 0ms 1,298 0ms update patternresultsrelevance set relevant = ?, saxo_relevant = ?, notrelevantpricedatetime = ?, reason = ? where uniqueindex = ? and relevant = ?;Times Reported Time consuming queries #13
Day Hour Count Duration Avg duration Jan 26 04 1,298 0ms 0ms 14 0ms 0ms 0ms 2 0ms update correlating_signals set filtered = true where id in (...) returning id;Times Reported Time consuming queries #14
Day Hour Count Duration Avg duration Jan 26 04 2 0ms 0ms 15 0ms 0ms 0ms 1 0ms insert into t15 (symbolid, pricedatetime, open, high, low, close, volume, bsf, sastdatetimereceived) values (?, ?::timestamp without time zone, ?.?, ?, ?.?, ?, ?, ?, ?::timestamp without time zone) on conflict (symbolid, pricedatetime) do nothing;Times Reported Time consuming queries #15
Day Hour Count Duration Avg duration Jan 26 04 1 0ms 0ms 16 0ms 0ms 0ms 240 0ms select v.datname, c.relname, v.phase, v.heap_blks_total, v.heap_blks_scanned, v.heap_blks_vacuumed, v.index_vacuum_count, v.max_dead_tuples, v.num_dead_tuples from pg_stat_progress_vacuum as v join pg_class c on c.oid = v.relid;Times Reported Time consuming queries #16
Day Hour Count Duration Avg duration Jan 26 04 240 0ms 0ms 17 0ms 0ms 0ms 18 0ms select cast(count(*) / cast(setting as numeric) * ? as int) from pg_stat_activity, pg_settings where name = ? group by setting;Times Reported Time consuming queries #17
Day Hour Count Duration Avg duration Jan 26 04 18 0ms 0ms 18 0ms 0ms 0ms 65 0ms select symbolid, pricedatetime, classname, downloadfrequency, downloadersymbol, open, high, low, close, volume, bsf, sastdatetimereceived from ( select pricedatetime, dss.classname, dss.downloadfrequency, dss.symbolid, dss.downloadersymbol, t.open, t.high, t.low, t.close, t.volume, t.bsf, t.sastdatetimereceived, row_number() over (partition by t.symbolid order by t.pricedatetime desc) as rn from t240 t, downloadersymbolsettings dss, symbols s where dss.classname = ? and dss.downloadfrequency = ? and dss.symbolid = t.symbolid and s.symbolid = dss.symbolid and dss.enabled = ? and s.deleted = ? and dss.downloadersymbol in (...) and t.pricedatetime > now() - interval ?) as ranked_candles_table where rn = ?;Times Reported Time consuming queries #18
Day Hour Count Duration Avg duration Jan 26 04 65 0ms 0ms 19 0ms 0ms 0ms 1 0ms select count(*) from autochartist_stocklist where base_engine = recognitionengine and enabled = ?;Times Reported Time consuming queries #19
Day Hour Count Duration Avg duration Jan 26 04 1 0ms 0ms 20 0ms 0ms 0ms 11 0ms select * from ( select pricedatetime, open, high, low, close, volume, bsf from t60 where symbolid = ? and (bsf = ? or bsf is null) order by pricedatetime desc limit ?) a order by pricedatetime asc;Times Reported Time consuming queries #20
Day Hour Count Duration Avg duration Jan 26 04 11 0ms 0ms Time consuming prepare
Rank Total duration Times executed Min duration Max duration Avg duration Query 1 13s250ms 14,192 0ms 19ms 0ms SELECT ;Times Reported Time consuming prepare #1
Day Hour Count Duration Avg duration Jan 26 04 14,192 13s250ms 0ms -
SELECT ;
Date: 2026-01-26 04:30:49 Duration: 19ms Database: postgres
-
SELECT ;
Date: 2026-01-26 04:50:53 Duration: 13ms Database: postgres
-
SELECT ;
Date: 2026-01-26 04:06:32 Duration: 12ms Database: postgres
2 6s809ms 5,123 0ms 18ms 1ms WITH rar_max as ( ;Times Reported Time consuming prepare #2
Day Hour Count Duration Avg duration 04 5,123 6s809ms 1ms -
WITH rar_max as ( ;
Date: 2026-01-26 04:51:14 Duration: 18ms Database: postgres
-
WITH rar_max as ( ;
Date: 2026-01-26 04:42:36 Duration: 15ms Database: postgres
-
WITH rar_max as ( ;
Date: 2026-01-26 04:52:24 Duration: 14ms Database: postgres
3 1s402ms 1,104 0ms 2ms 1ms SELECT symbolid, ;Times Reported Time consuming prepare #3
Day Hour Count Duration Avg duration 04 1,104 1s402ms 1ms -
SELECT symbolid, ;
Date: 2026-01-26 04:30:48 Duration: 2ms Database: postgres
-
SELECT symbolid, ;
Date: 2026-01-26 04:01:14 Duration: 2ms Database: postgres
-
SELECT symbolid, ;
Date: 2026-01-26 04:10:42 Duration: 2ms Database: postgres
4 1s359ms 8,876 0ms 13ms 0ms SET extra_float_digits = 3;Times Reported Time consuming prepare #4
Day Hour Count Duration Avg duration 04 8,876 1s359ms 0ms -
SET extra_float_digits = 3;
Date: 2026-01-26 04:29:10 Duration: 13ms Database: postgres
-
SET extra_float_digits = 3;
Date: 2026-01-26 04:42:36 Duration: 11ms Database: postgres
-
SET extra_float_digits = 3;
Date: 2026-01-26 04:29:10 Duration: 9ms Database: postgres
5 757ms 14,096 0ms 7ms 0ms select 1;Times Reported Time consuming prepare #5
Day Hour Count Duration Avg duration 04 14,096 757ms 0ms -
select 1;
Date: 2026-01-26 04:12:44 Duration: 7ms Database: postgres
-
select 1;
Date: 2026-01-26 04:20:38 Duration: 7ms Database: postgres
-
select 1;
Date: 2026-01-26 04:34:41 Duration: 6ms Database: postgres
6 450ms 437 0ms 1ms 1ms SELECT s.symbolid, dss.downloadfrequency, dss.downloadersymbol;Times Reported Time consuming prepare #6
Day Hour Count Duration Avg duration 04 437 450ms 1ms -
SELECT s.symbolid, dss.downloadfrequency, dss.downloadersymbol;
Date: 2026-01-26 04:31:10 Duration: 1ms Database: postgres
-
SELECT s.symbolid, dss.downloadfrequency, dss.downloadersymbol;
Date: 2026-01-26 04:45:43 Duration: 1ms Database: postgres
-
SELECT s.symbolid, dss.downloadfrequency, dss.downloadersymbol;
Date: 2026-01-26 04:30:48 Duration: 1ms Database: postgres
7 251ms 3,112 0ms 0ms 0ms INSERT INTO T30 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;Times Reported Time consuming prepare #7
Day Hour Count Duration Avg duration 04 3,112 251ms 0ms -
INSERT INTO T30 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;
Date: 2026-01-26 04:10:58 Duration: 0ms Database: postgres
-
INSERT INTO T30 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;
Date: 2026-01-26 04:30:45 Duration: 0ms Database: postgres
-
INSERT INTO T30 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;
Date: 2026-01-26 04:31:14 Duration: 0ms Database: postgres
8 183ms 2,067 0ms 0ms 0ms INSERT INTO T60 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;Times Reported Time consuming prepare #8
Day Hour Count Duration Avg duration 04 2,067 183ms 0ms -
INSERT INTO T60 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;
Date: 2026-01-26 04:10:54 Duration: 0ms Database: postgres
-
INSERT INTO T60 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;
Date: 2026-01-26 04:11:45 Duration: 0ms Database: postgres
-
INSERT INTO T60 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;
Date: 2026-01-26 04:11:59 Duration: 0ms Database: postgres
9 133ms 951 0ms 0ms 0ms INSERT INTO T15 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;Times Reported Time consuming prepare #9
Day Hour Count Duration Avg duration 04 951 133ms 0ms -
INSERT INTO T15 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;
Date: 2026-01-26 04:47:44 Duration: 0ms Database: postgres
-
INSERT INTO T15 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;
Date: 2026-01-26 04:47:46 Duration: 0ms Database: postgres
-
INSERT INTO T15 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;
Date: 2026-01-26 04:26:44 Duration: 0ms Database: postgres
10 122ms 8,850 0ms 4ms 0ms SET application_name = 'PostgreSQL JDBC Driver';Times Reported Time consuming prepare #10
Day Hour Count Duration Avg duration 04 8,850 122ms 0ms -
SET application_name = 'PostgreSQL JDBC Driver';
Date: 2026-01-26 04:49:18 Duration: 4ms Database: postgres
-
SET application_name = 'PostgreSQL JDBC Driver';
Date: 2026-01-26 04:25:39 Duration: 3ms Database: postgres
-
SET application_name = 'PostgreSQL JDBC Driver';
Date: 2026-01-26 04:16:45 Duration: 2ms Database: postgres
11 107ms 1,029 0ms 0ms 0ms INSERT INTO T240 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;Times Reported Time consuming prepare #11
Day Hour Count Duration Avg duration 04 1,029 107ms 0ms -
INSERT INTO T240 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;
Date: 2026-01-26 04:10:46 Duration: 0ms Database: postgres
-
INSERT INTO T240 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;
Date: 2026-01-26 04:10:56 Duration: 0ms Database: postgres
-
INSERT INTO T240 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;
Date: 2026-01-26 04:02:33 Duration: 0ms Database: postgres
12 71ms 12 4ms 8ms 5ms with sym_info as ( ;Times Reported Time consuming prepare #12
Day Hour Count Duration Avg duration 04 12 71ms 5ms -
with sym_info as ( ;
Date: 2026-01-26 04:06:41 Duration: 8ms Database: postgres
-
with sym_info as ( ;
Date: 2026-01-26 04:51:45 Duration: 7ms Database: postgres
-
with sym_info as ( ;
Date: 2026-01-26 04:36:57 Duration: 6ms Database: postgres
13 43ms 18 1ms 3ms 2ms select cast(count(*) / cast(setting as numeric) * 100 as int) from pg_stat_activity, pg_settings WHERE name = 'max_connections' group by setting;Times Reported Time consuming prepare #13
Day Hour Count Duration Avg duration 04 18 43ms 2ms -
select cast(count(*) / cast(setting as numeric) * 100 as int) from pg_stat_activity, pg_settings WHERE name = 'max_connections' group by setting;
Date: 2026-01-26 04:41:03 Duration: 3ms Database: postgres
-
select cast(count(*) / cast(setting as numeric) * 100 as int) from pg_stat_activity, pg_settings WHERE name = 'max_connections' group by setting;
Date: 2026-01-26 04:01:03 Duration: 2ms Database: postgres
-
select cast(count(*) / cast(setting as numeric) * 100 as int) from pg_stat_activity, pg_settings WHERE name = 'max_connections' group by setting;
Date: 2026-01-26 04:21:01 Duration: 2ms Database: postgres
14 25ms 163 0ms 0ms 0ms SELECT NULL AS TABLE_CAT, n.nspname AS TABLE_SCHEM, c.relname AS TABLE_NAME, CASE n.nspname ~ '^pg_' OR n.nspname = 'information_schema' WHEN true THEN CASE WHEN n.nspname = 'pg_catalog' OR n.nspname = 'information_schema' THEN CASE c.relkind WHEN 'r' THEN 'SYSTEM TABLE' WHEN 'v' THEN 'SYSTEM VIEW' WHEN 'i' THEN 'SYSTEM INDEX' ELSE NULL END WHEN n.nspname = 'pg_toast' THEN CASE c.relkind WHEN 'r' THEN 'SYSTEM TOAST TABLE' WHEN 'i' THEN 'SYSTEM TOAST INDEX' ELSE NULL END ELSE CASE c.relkind WHEN 'r' THEN 'TEMPORARY TABLE' WHEN 'p' THEN 'TEMPORARY TABLE' WHEN 'i' THEN 'TEMPORARY INDEX' WHEN 'S' THEN 'TEMPORARY SEQUENCE' WHEN 'v' THEN 'TEMPORARY VIEW' ELSE NULL END END WHEN false THEN CASE c.relkind WHEN 'r' THEN 'TABLE' WHEN 'p' THEN 'PARTITIONED TABLE' WHEN 'i' THEN 'INDEX' WHEN 'S' THEN 'SEQUENCE' WHEN 'v' THEN 'VIEW' WHEN 'c' THEN 'TYPE' WHEN 'f' THEN 'FOREIGN TABLE' WHEN 'm' THEN 'MATERIALIZED VIEW' ELSE NULL END ELSE NULL END AS TABLE_TYPE, d.description AS REMARKS, '' as TYPE_CAT, '' as TYPE_SCHEM, '' as TYPE_NAME, '' AS SELF_REFERENCING_COL_NAME, '' AS REF_GENERATION FROM pg_catalog.pg_namespace n, pg_catalog.pg_class c LEFT JOIN pg_catalog.pg_description d ON (c.oid = d.objoid AND d.objsubid = 0) LEFT JOIN pg_catalog.pg_class dc ON (d.classoid = dc.oid AND dc.relname = 'pg_class') LEFT JOIN pg_catalog.pg_namespace dn ON (dn.oid = dc.relnamespace AND dn.nspname = 'pg_catalog') WHERE c.relnamespace = n.oid AND c.relname LIKE 'PROBABLYNOT' AND (false OR (c.relkind = 'r' AND n.nspname !~ '^pg_' AND n.nspname <> 'information_schema')) ORDER BY TABLE_TYPE, TABLE_SCHEM, TABLE_NAME;Times Reported Time consuming prepare #14
Day Hour Count Duration Avg duration 04 163 25ms 0ms -
SELECT NULL AS TABLE_CAT, n.nspname AS TABLE_SCHEM, c.relname AS TABLE_NAME, CASE n.nspname ~ '^pg_' OR n.nspname = 'information_schema' WHEN true THEN CASE WHEN n.nspname = 'pg_catalog' OR n.nspname = 'information_schema' THEN CASE c.relkind WHEN 'r' THEN 'SYSTEM TABLE' WHEN 'v' THEN 'SYSTEM VIEW' WHEN 'i' THEN 'SYSTEM INDEX' ELSE NULL END WHEN n.nspname = 'pg_toast' THEN CASE c.relkind WHEN 'r' THEN 'SYSTEM TOAST TABLE' WHEN 'i' THEN 'SYSTEM TOAST INDEX' ELSE NULL END ELSE CASE c.relkind WHEN 'r' THEN 'TEMPORARY TABLE' WHEN 'p' THEN 'TEMPORARY TABLE' WHEN 'i' THEN 'TEMPORARY INDEX' WHEN 'S' THEN 'TEMPORARY SEQUENCE' WHEN 'v' THEN 'TEMPORARY VIEW' ELSE NULL END END WHEN false THEN CASE c.relkind WHEN 'r' THEN 'TABLE' WHEN 'p' THEN 'PARTITIONED TABLE' WHEN 'i' THEN 'INDEX' WHEN 'S' THEN 'SEQUENCE' WHEN 'v' THEN 'VIEW' WHEN 'c' THEN 'TYPE' WHEN 'f' THEN 'FOREIGN TABLE' WHEN 'm' THEN 'MATERIALIZED VIEW' ELSE NULL END ELSE NULL END AS TABLE_TYPE, d.description AS REMARKS, '' as TYPE_CAT, '' as TYPE_SCHEM, '' as TYPE_NAME, '' AS SELF_REFERENCING_COL_NAME, '' AS REF_GENERATION FROM pg_catalog.pg_namespace n, pg_catalog.pg_class c LEFT JOIN pg_catalog.pg_description d ON (c.oid = d.objoid AND d.objsubid = 0) LEFT JOIN pg_catalog.pg_class dc ON (d.classoid = dc.oid AND dc.relname = 'pg_class') LEFT JOIN pg_catalog.pg_namespace dn ON (dn.oid = dc.relnamespace AND dn.nspname = 'pg_catalog') WHERE c.relnamespace = n.oid AND c.relname LIKE 'PROBABLYNOT' AND (false OR (c.relkind = 'r' AND n.nspname !~ '^pg_' AND n.nspname <> 'information_schema')) ORDER BY TABLE_TYPE, TABLE_SCHEM, TABLE_NAME;
Date: 2026-01-26 04:13:19 Duration: 0ms Database: postgres
-
SELECT NULL AS TABLE_CAT, n.nspname AS TABLE_SCHEM, c.relname AS TABLE_NAME, CASE n.nspname ~ '^pg_' OR n.nspname = 'information_schema' WHEN true THEN CASE WHEN n.nspname = 'pg_catalog' OR n.nspname = 'information_schema' THEN CASE c.relkind WHEN 'r' THEN 'SYSTEM TABLE' WHEN 'v' THEN 'SYSTEM VIEW' WHEN 'i' THEN 'SYSTEM INDEX' ELSE NULL END WHEN n.nspname = 'pg_toast' THEN CASE c.relkind WHEN 'r' THEN 'SYSTEM TOAST TABLE' WHEN 'i' THEN 'SYSTEM TOAST INDEX' ELSE NULL END ELSE CASE c.relkind WHEN 'r' THEN 'TEMPORARY TABLE' WHEN 'p' THEN 'TEMPORARY TABLE' WHEN 'i' THEN 'TEMPORARY INDEX' WHEN 'S' THEN 'TEMPORARY SEQUENCE' WHEN 'v' THEN 'TEMPORARY VIEW' ELSE NULL END END WHEN false THEN CASE c.relkind WHEN 'r' THEN 'TABLE' WHEN 'p' THEN 'PARTITIONED TABLE' WHEN 'i' THEN 'INDEX' WHEN 'S' THEN 'SEQUENCE' WHEN 'v' THEN 'VIEW' WHEN 'c' THEN 'TYPE' WHEN 'f' THEN 'FOREIGN TABLE' WHEN 'm' THEN 'MATERIALIZED VIEW' ELSE NULL END ELSE NULL END AS TABLE_TYPE, d.description AS REMARKS, '' as TYPE_CAT, '' as TYPE_SCHEM, '' as TYPE_NAME, '' AS SELF_REFERENCING_COL_NAME, '' AS REF_GENERATION FROM pg_catalog.pg_namespace n, pg_catalog.pg_class c LEFT JOIN pg_catalog.pg_description d ON (c.oid = d.objoid AND d.objsubid = 0) LEFT JOIN pg_catalog.pg_class dc ON (d.classoid = dc.oid AND dc.relname = 'pg_class') LEFT JOIN pg_catalog.pg_namespace dn ON (dn.oid = dc.relnamespace AND dn.nspname = 'pg_catalog') WHERE c.relnamespace = n.oid AND c.relname LIKE 'PROBABLYNOT' AND (false OR (c.relkind = 'r' AND n.nspname !~ '^pg_' AND n.nspname <> 'information_schema')) ORDER BY TABLE_TYPE, TABLE_SCHEM, TABLE_NAME;
Date: 2026-01-26 04:13:20 Duration: 0ms Database: postgres
-
SELECT NULL AS TABLE_CAT, n.nspname AS TABLE_SCHEM, c.relname AS TABLE_NAME, CASE n.nspname ~ '^pg_' OR n.nspname = 'information_schema' WHEN true THEN CASE WHEN n.nspname = 'pg_catalog' OR n.nspname = 'information_schema' THEN CASE c.relkind WHEN 'r' THEN 'SYSTEM TABLE' WHEN 'v' THEN 'SYSTEM VIEW' WHEN 'i' THEN 'SYSTEM INDEX' ELSE NULL END WHEN n.nspname = 'pg_toast' THEN CASE c.relkind WHEN 'r' THEN 'SYSTEM TOAST TABLE' WHEN 'i' THEN 'SYSTEM TOAST INDEX' ELSE NULL END ELSE CASE c.relkind WHEN 'r' THEN 'TEMPORARY TABLE' WHEN 'p' THEN 'TEMPORARY TABLE' WHEN 'i' THEN 'TEMPORARY INDEX' WHEN 'S' THEN 'TEMPORARY SEQUENCE' WHEN 'v' THEN 'TEMPORARY VIEW' ELSE NULL END END WHEN false THEN CASE c.relkind WHEN 'r' THEN 'TABLE' WHEN 'p' THEN 'PARTITIONED TABLE' WHEN 'i' THEN 'INDEX' WHEN 'S' THEN 'SEQUENCE' WHEN 'v' THEN 'VIEW' WHEN 'c' THEN 'TYPE' WHEN 'f' THEN 'FOREIGN TABLE' WHEN 'm' THEN 'MATERIALIZED VIEW' ELSE NULL END ELSE NULL END AS TABLE_TYPE, d.description AS REMARKS, '' as TYPE_CAT, '' as TYPE_SCHEM, '' as TYPE_NAME, '' AS SELF_REFERENCING_COL_NAME, '' AS REF_GENERATION FROM pg_catalog.pg_namespace n, pg_catalog.pg_class c LEFT JOIN pg_catalog.pg_description d ON (c.oid = d.objoid AND d.objsubid = 0) LEFT JOIN pg_catalog.pg_class dc ON (d.classoid = dc.oid AND dc.relname = 'pg_class') LEFT JOIN pg_catalog.pg_namespace dn ON (dn.oid = dc.relnamespace AND dn.nspname = 'pg_catalog') WHERE c.relnamespace = n.oid AND c.relname LIKE 'PROBABLYNOT' AND (false OR (c.relkind = 'r' AND n.nspname !~ '^pg_' AND n.nspname <> 'information_schema')) ORDER BY TABLE_TYPE, TABLE_SCHEM, TABLE_NAME;
Date: 2026-01-26 04:13:21 Duration: 0ms Database: postgres
15 14ms 6 2ms 2ms 2ms with rankedmt4 as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors ), last_feed_entry as ( select * from rankedmt4 where r = 1 ), ok_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where status = 'OK' ), earliest_entry_after_ok as ( select m.datafeedname, min(m.eventtimestamp) as eventtimestamp from mt4datafeederrors m left outer join ( select datafeedname, eventtimestamp from ok_entries where r = 1) oo on m.datafeedname = oo.datafeedname where m.eventtimestamp > coalesce(oo.eventtimestamp, '1900-01-01'::timestamp without time zone) group by m.datafeedname ), notified_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where notified is not null and notified <> '' ), broker as ( select *, row_number() over (partition by feedname order by brokerid) r from ( select distinct b.brokerid, b.name as brokername, dss.classname as feedname from downloadersymbolsettings dss inner join brokersymbollist bsl on dss.symbolid = bsl.symbolid inner join broker b on bsl.brokerid = b.brokerid where dss.enabled = 1) a ) select last.id, last.datafeedname, last.eventtimestamp, last.status, last.errordescription, last.serveraddress, last.username, note.notified, note.eventtimestamp, broker.brokername from last_feed_entry last inner join earliest_entry_after_ok after_ok on last.datafeedname = after_ok.datafeedname inner join broker on last.datafeedname = broker.feedname left outer join ok_entries ok on ok.datafeedname = last.datafeedname left outer join notified_entries note on note.datafeedname = last.datafeedname and note.r = 1 where (ok.r is null or ok.r = 1) and last.datafeedname not in ( select distinct datafeedname from last_feed_entry where status = 'OK') and extract(epoch from (last.eventtimestamp - after_ok.eventtimestamp)) > 60 * 60 and last.eventtimestamp > current_timestamp - interval '1 day' and (note.eventtimestamp is null or note.eventtimestamp < current_timestamp - interval '10 hours') and last.eventtimestamp > current_timestamp - interval '1 hour' and broker.r = 1;Times Reported Time consuming prepare #15
Day Hour Count Duration Avg duration 04 6 14ms 2ms -
with rankedmt4 as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors ), last_feed_entry as ( select * from rankedmt4 where r = 1 ), ok_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where status = 'OK' ), earliest_entry_after_ok as ( select m.datafeedname, min(m.eventtimestamp) as eventtimestamp from mt4datafeederrors m left outer join ( select datafeedname, eventtimestamp from ok_entries where r = 1) oo on m.datafeedname = oo.datafeedname where m.eventtimestamp > coalesce(oo.eventtimestamp, '1900-01-01'::timestamp without time zone) group by m.datafeedname ), notified_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where notified is not null and notified <> '' ), broker as ( select *, row_number() over (partition by feedname order by brokerid) r from ( select distinct b.brokerid, b.name as brokername, dss.classname as feedname from downloadersymbolsettings dss inner join brokersymbollist bsl on dss.symbolid = bsl.symbolid inner join broker b on bsl.brokerid = b.brokerid where dss.enabled = 1) a ) select last.id, last.datafeedname, last.eventtimestamp, last.status, last.errordescription, last.serveraddress, last.username, note.notified, note.eventtimestamp, broker.brokername from last_feed_entry last inner join earliest_entry_after_ok after_ok on last.datafeedname = after_ok.datafeedname inner join broker on last.datafeedname = broker.feedname left outer join ok_entries ok on ok.datafeedname = last.datafeedname left outer join notified_entries note on note.datafeedname = last.datafeedname and note.r = 1 where (ok.r is null or ok.r = 1) and last.datafeedname not in ( select distinct datafeedname from last_feed_entry where status = 'OK') and extract(epoch from (last.eventtimestamp - after_ok.eventtimestamp)) > 60 * 60 and last.eventtimestamp > current_timestamp - interval '1 day' and (note.eventtimestamp is null or note.eventtimestamp < current_timestamp - interval '10 hours') and last.eventtimestamp > current_timestamp - interval '1 hour' and broker.r = 1;
Date: 2026-01-26 04:10:02 Duration: 2ms Database: postgres
-
with rankedmt4 as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors ), last_feed_entry as ( select * from rankedmt4 where r = 1 ), ok_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where status = 'OK' ), earliest_entry_after_ok as ( select m.datafeedname, min(m.eventtimestamp) as eventtimestamp from mt4datafeederrors m left outer join ( select datafeedname, eventtimestamp from ok_entries where r = 1) oo on m.datafeedname = oo.datafeedname where m.eventtimestamp > coalesce(oo.eventtimestamp, '1900-01-01'::timestamp without time zone) group by m.datafeedname ), notified_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where notified is not null and notified <> '' ), broker as ( select *, row_number() over (partition by feedname order by brokerid) r from ( select distinct b.brokerid, b.name as brokername, dss.classname as feedname from downloadersymbolsettings dss inner join brokersymbollist bsl on dss.symbolid = bsl.symbolid inner join broker b on bsl.brokerid = b.brokerid where dss.enabled = 1) a ) select last.id, last.datafeedname, last.eventtimestamp, last.status, last.errordescription, last.serveraddress, last.username, note.notified, note.eventtimestamp, broker.brokername from last_feed_entry last inner join earliest_entry_after_ok after_ok on last.datafeedname = after_ok.datafeedname inner join broker on last.datafeedname = broker.feedname left outer join ok_entries ok on ok.datafeedname = last.datafeedname left outer join notified_entries note on note.datafeedname = last.datafeedname and note.r = 1 where (ok.r is null or ok.r = 1) and last.datafeedname not in ( select distinct datafeedname from last_feed_entry where status = 'OK') and extract(epoch from (last.eventtimestamp - after_ok.eventtimestamp)) > 60 * 60 and last.eventtimestamp > current_timestamp - interval '1 day' and (note.eventtimestamp is null or note.eventtimestamp < current_timestamp - interval '10 hours') and last.eventtimestamp > current_timestamp - interval '1 hour' and broker.r = 1;
Date: 2026-01-26 04:30:02 Duration: 2ms Database: postgres
-
with rankedmt4 as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors ), last_feed_entry as ( select * from rankedmt4 where r = 1 ), ok_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where status = 'OK' ), earliest_entry_after_ok as ( select m.datafeedname, min(m.eventtimestamp) as eventtimestamp from mt4datafeederrors m left outer join ( select datafeedname, eventtimestamp from ok_entries where r = 1) oo on m.datafeedname = oo.datafeedname where m.eventtimestamp > coalesce(oo.eventtimestamp, '1900-01-01'::timestamp without time zone) group by m.datafeedname ), notified_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where notified is not null and notified <> '' ), broker as ( select *, row_number() over (partition by feedname order by brokerid) r from ( select distinct b.brokerid, b.name as brokername, dss.classname as feedname from downloadersymbolsettings dss inner join brokersymbollist bsl on dss.symbolid = bsl.symbolid inner join broker b on bsl.brokerid = b.brokerid where dss.enabled = 1) a ) select last.id, last.datafeedname, last.eventtimestamp, last.status, last.errordescription, last.serveraddress, last.username, note.notified, note.eventtimestamp, broker.brokername from last_feed_entry last inner join earliest_entry_after_ok after_ok on last.datafeedname = after_ok.datafeedname inner join broker on last.datafeedname = broker.feedname left outer join ok_entries ok on ok.datafeedname = last.datafeedname left outer join notified_entries note on note.datafeedname = last.datafeedname and note.r = 1 where (ok.r is null or ok.r = 1) and last.datafeedname not in ( select distinct datafeedname from last_feed_entry where status = 'OK') and extract(epoch from (last.eventtimestamp - after_ok.eventtimestamp)) > 60 * 60 and last.eventtimestamp > current_timestamp - interval '1 day' and (note.eventtimestamp is null or note.eventtimestamp < current_timestamp - interval '10 hours') and last.eventtimestamp > current_timestamp - interval '1 hour' and broker.r = 1;
Date: 2026-01-26 04:50:02 Duration: 2ms Database: postgres
16 14ms 6 2ms 2ms 2ms select client_addr, count(1) from pg_stat_activity, pg_settings WHERE name = 'max_connections' group by client_addr, setting having (client_addr is not null OR (client_addr is null and count(1) > (cast(setting as numeric) / 3 * 2))) order by count desc;Times Reported Time consuming prepare #16
Day Hour Count Duration Avg duration 04 6 14ms 2ms -
select client_addr, count(1) from pg_stat_activity, pg_settings WHERE name = 'max_connections' group by client_addr, setting having (client_addr is not null OR (client_addr is null and count(1) > (cast(setting as numeric) / 3 * 2))) order by count desc;
Date: 2026-01-26 04:10:05 Duration: 2ms Database: postgres
-
select client_addr, count(1) from pg_stat_activity, pg_settings WHERE name = 'max_connections' group by client_addr, setting having (client_addr is not null OR (client_addr is null and count(1) > (cast(setting as numeric) / 3 * 2))) order by count desc;
Date: 2026-01-26 04:30:05 Duration: 2ms Database: postgres
-
select client_addr, count(1) from pg_stat_activity, pg_settings WHERE name = 'max_connections' group by client_addr, setting having (client_addr is not null OR (client_addr is null and count(1) > (cast(setting as numeric) / 3 * 2))) order by count desc;
Date: 2026-01-26 04:50:04 Duration: 2ms Database: postgres
17 13ms 24 0ms 0ms 0ms select count(*) from datafeed_restarter_events where is_current_entry = 1;Times Reported Time consuming prepare #17
Day Hour Count Duration Avg duration 04 24 13ms 0ms -
select count(*) from datafeed_restarter_events where is_current_entry = 1;
Date: 2026-01-26 04:20:03 Duration: 0ms Database: postgres
-
select count(*) from datafeed_restarter_events where is_current_entry = 1;
Date: 2026-01-26 04:30:01 Duration: 0ms Database: postgres
-
select count(*) from datafeed_restarter_events where is_current_entry = 1;
Date: 2026-01-26 04:35:00 Duration: 0ms Database: postgres
18 12ms 8 0ms 3ms 1ms WITH pre_symbols AS ( /* find relevant symbols */ ;Times Reported Time consuming prepare #18
Day Hour Count Duration Avg duration 04 8 12ms 1ms -
WITH pre_symbols AS ( /* find relevant symbols */ ;
Date: 2026-01-26 04:13:19 Duration: 3ms Database: postgres
-
WITH pre_symbols AS ( /* find relevant symbols */ ;
Date: 2026-01-26 04:13:19 Duration: 2ms Database: postgres
-
WITH pre_symbols AS ( /* find relevant symbols */ ;
Date: 2026-01-26 04:13:18 Duration: 2ms Database: postgres
19 12ms 12 0ms 1ms 1ms select distinct classname, to_char(created_datetime, 'yyyy-mm-dd HH24:MI'), to_char(cleared_datetime, 'yyyy-mm-dd HH24:MI'), action_to_take, description, created_datetime from datafeed_restarter_events where (is_current_entry = 1 OR cleared_datetime > current_timestamp - interval '17 hour') order by created_datetime desc;Times Reported Time consuming prepare #19
Day Hour Count Duration Avg duration 04 12 12ms 1ms -
select distinct classname, to_char(created_datetime, 'yyyy-mm-dd HH24:MI'), to_char(cleared_datetime, 'yyyy-mm-dd HH24:MI'), action_to_take, description, created_datetime from datafeed_restarter_events where (is_current_entry = 1 OR cleared_datetime > current_timestamp - interval '17 hour') order by created_datetime desc;
Date: 2026-01-26 04:17:10 Duration: 1ms Database: postgres
-
select distinct classname, to_char(created_datetime, 'yyyy-mm-dd HH24:MI'), to_char(cleared_datetime, 'yyyy-mm-dd HH24:MI'), action_to_take, description, created_datetime from datafeed_restarter_events where (is_current_entry = 1 OR cleared_datetime > current_timestamp - interval '17 hour') order by created_datetime desc;
Date: 2026-01-26 04:47:16 Duration: 1ms Database: postgres
-
select distinct classname, to_char(created_datetime, 'yyyy-mm-dd HH24:MI'), to_char(cleared_datetime, 'yyyy-mm-dd HH24:MI'), action_to_take, description, created_datetime from datafeed_restarter_events where (is_current_entry = 1 OR cleared_datetime > current_timestamp - interval '17 hour') order by created_datetime desc;
Date: 2026-01-26 04:42:15 Duration: 1ms Database: postgres
20 10ms 12 0ms 1ms 0ms select feedname, to_char(latestrxtime, 'yyyy-mm-dd HH24:MI'), to_char(LatestDBWriteTime, 'yyyy-mm-dd HH24:MI'), to_char(LatestStartupTime, 'yyyy-mm-dd HH24:MI'), StartupTimeInMinutes, dm.source_type, dm.transport_type, case when latestrxtime < (CURRENT_TIMESTAMP - 5 * interval '1 minute') then 'X' else 'OK' end, case when (feedname ilike '%_EOD' OR feedname ilike 'IQFEED_DAILIES' or feedname ilike 'YAHOO%' or feedname ilike 'QUANDL_FUTURES%' or feedname ilike 'BAR_CHART') then case when LatestDBWriteTime < (CURRENT_TIMESTAMP - 24 * interval '1 hour') then 'X' else 'OK' end else case when (LatestDBWriteTime < (CURRENT_TIMESTAMP - 15 * interval '1 minute') and LatestStartupTime < (CURRENT_TIMESTAMP - 30 * interval '1 minute')) OR latestrxtime < CURRENT_TIMESTAMP - interval '2 hour' then 'X' else 'OK' end end as statusDB, comment from datafeeds_latestrun dlr left outer join datafeeds df on dlr.feedname ilike df.name inner join datafeeds_metadata dm on df.metadata_id = dm.id order by feedname;Times Reported Time consuming prepare #20
Day Hour Count Duration Avg duration 04 12 10ms 0ms -
select feedname, to_char(latestrxtime, 'yyyy-mm-dd HH24:MI'), to_char(LatestDBWriteTime, 'yyyy-mm-dd HH24:MI'), to_char(LatestStartupTime, 'yyyy-mm-dd HH24:MI'), StartupTimeInMinutes, dm.source_type, dm.transport_type, case when latestrxtime < (CURRENT_TIMESTAMP - 5 * interval '1 minute') then 'X' else 'OK' end, case when (feedname ilike '%_EOD' OR feedname ilike 'IQFEED_DAILIES' or feedname ilike 'YAHOO%' or feedname ilike 'QUANDL_FUTURES%' or feedname ilike 'BAR_CHART') then case when LatestDBWriteTime < (CURRENT_TIMESTAMP - 24 * interval '1 hour') then 'X' else 'OK' end else case when (LatestDBWriteTime < (CURRENT_TIMESTAMP - 15 * interval '1 minute') and LatestStartupTime < (CURRENT_TIMESTAMP - 30 * interval '1 minute')) OR latestrxtime < CURRENT_TIMESTAMP - interval '2 hour' then 'X' else 'OK' end end as statusDB, comment from datafeeds_latestrun dlr left outer join datafeeds df on dlr.feedname ilike df.name inner join datafeeds_metadata dm on df.metadata_id = dm.id order by feedname;
Date: 2026-01-26 04:07:09 Duration: 1ms Database: postgres
-
select feedname, to_char(latestrxtime, 'yyyy-mm-dd HH24:MI'), to_char(LatestDBWriteTime, 'yyyy-mm-dd HH24:MI'), to_char(LatestStartupTime, 'yyyy-mm-dd HH24:MI'), StartupTimeInMinutes, dm.source_type, dm.transport_type, case when latestrxtime < (CURRENT_TIMESTAMP - 5 * interval '1 minute') then 'X' else 'OK' end, case when (feedname ilike '%_EOD' OR feedname ilike 'IQFEED_DAILIES' or feedname ilike 'YAHOO%' or feedname ilike 'QUANDL_FUTURES%' or feedname ilike 'BAR_CHART') then case when LatestDBWriteTime < (CURRENT_TIMESTAMP - 24 * interval '1 hour') then 'X' else 'OK' end else case when (LatestDBWriteTime < (CURRENT_TIMESTAMP - 15 * interval '1 minute') and LatestStartupTime < (CURRENT_TIMESTAMP - 30 * interval '1 minute')) OR latestrxtime < CURRENT_TIMESTAMP - interval '2 hour' then 'X' else 'OK' end end as statusDB, comment from datafeeds_latestrun dlr left outer join datafeeds df on dlr.feedname ilike df.name inner join datafeeds_metadata dm on df.metadata_id = dm.id order by feedname;
Date: 2026-01-26 04:42:15 Duration: 1ms Database: postgres
-
select feedname, to_char(latestrxtime, 'yyyy-mm-dd HH24:MI'), to_char(LatestDBWriteTime, 'yyyy-mm-dd HH24:MI'), to_char(LatestStartupTime, 'yyyy-mm-dd HH24:MI'), StartupTimeInMinutes, dm.source_type, dm.transport_type, case when latestrxtime < (CURRENT_TIMESTAMP - 5 * interval '1 minute') then 'X' else 'OK' end, case when (feedname ilike '%_EOD' OR feedname ilike 'IQFEED_DAILIES' or feedname ilike 'YAHOO%' or feedname ilike 'QUANDL_FUTURES%' or feedname ilike 'BAR_CHART') then case when LatestDBWriteTime < (CURRENT_TIMESTAMP - 24 * interval '1 hour') then 'X' else 'OK' end else case when (LatestDBWriteTime < (CURRENT_TIMESTAMP - 15 * interval '1 minute') and LatestStartupTime < (CURRENT_TIMESTAMP - 30 * interval '1 minute')) OR latestrxtime < CURRENT_TIMESTAMP - interval '2 hour' then 'X' else 'OK' end end as statusDB, comment from datafeeds_latestrun dlr left outer join datafeeds df on dlr.feedname ilike df.name inner join datafeeds_metadata dm on df.metadata_id = dm.id order by feedname;
Date: 2026-01-26 04:47:16 Duration: 1ms Database: postgres
Time consuming bind
Rank Total duration Times executed Min duration Max duration Avg duration Query 1 49s116ms 5,819 0ms 58ms 8ms WITH rar_max as ( ;Times Reported Time consuming bind #1
Day Hour Count Duration Avg duration Jan 26 04 5,819 49s116ms 8ms -
WITH rar_max as ( ;
Date: 2026-01-26 04:21:08 Duration: 58ms Database: postgres parameters: $1 = '667', $2 = '7', $3 = '15', $4 = '30', $5 = '60', $6 = '120', $7 = '240', $8 = '480', $9 = '1440', $10 = '0', $11 = '', $12 = '0', $13 = '', $14 = '700', $15 = '700', $16 = 't', $17 = '10', $18 = '10'
-
WITH rar_max as ( ;
Date: 2026-01-26 04:42:36 Duration: 48ms Database: postgres parameters: $1 = '607567836164958301', $2 = '607567836164958301', $3 = '607567836164958301'
-
WITH rar_max as ( ;
Date: 2026-01-26 04:26:39 Duration: 47ms Database: postgres parameters: $1 = 't', $2 = '667', $3 = '7', $4 = '15', $5 = '30', $6 = '60', $7 = '120', $8 = '240', $9 = '480', $10 = '1440', $11 = '0', $12 = '', $13 = '84', $14 = 'AUDCAD', $15 = 'AUDCHF', $16 = 'AUDJPY', $17 = 'AUDNZD', $18 = 'AUDSGD', $19 = 'CADCHF', $20 = 'CADJPY', $21 = 'CHFJPY', $22 = 'EURAUD', $23 = 'EURCAD', $24 = 'EURCHF', $25 = 'EURCZK', $26 = 'EURGBP', $27 = 'EURHUF', $28 = 'EURJPY', $29 = 'EURNOK', $30 = 'EURNZD', $31 = 'EURPLN', $32 = 'EURSEK', $33 = 'EURSGD', $34 = 'EURTRY', $35 = 'EURZAR', $36 = 'GBPAUD', $37 = 'GBPCAD', $38 = 'GBPCHF', $39 = 'GBPJPY', $40 = 'GBPNZD', $41 = 'GBPPLN', $42 = 'GBPSEK', $43 = 'GBPSGD', $44 = 'NZDCAD', $45 = 'NZDCHF', $46 = 'NZDJPY', $47 = 'NZDSGD', $48 = 'USDCNH', $49 = 'USDCZK', $50 = 'USDHUF', $51 = 'USDNOK', $52 = 'USDPLN', $53 = 'USDSEK', $54 = 'USDSGD', $55 = 'USDTRY', $56 = 'USDZAR', $57 = 'WTI', $58 = 'XBRUSD', $59 = 'XTIUSD', $60 = 'BTCUSD', $61 = 'XAGAUD', $62 = 'XAGUSD', $63 = 'XAUAUD', $64 = 'XAUUSD', $65 = 'XPTUSD', $66 = 'XPDUSD', $67 = 'AUDUSD', $68 = 'EURUSD', $69 = 'GBPUSD', $70 = 'NZDUSD', $71 = 'USDCAD', $72 = 'USDCHF', $73 = 'USDHKD', $74 = 'USDJPY', $75 = 'AUS200', $76 = 'CHINA300', $77 = 'CHINA50', $78 = 'DJ30', $79 = 'ESP35t', $80 = 'EUR50', $81 = 'EURO50', $82 = 'FRA40', $83 = 'GDAXI', $84 = 'GDAXIm', $85 = 'HK50', $86 = 'ITA40', $87 = 'J225', $88 = 'JP225', $89 = 'NAS100', $90 = 'SING30', $91 = 'SPA35', $92 = 'STOXX50', $93 = 'SUI20', $94 = 'UK100', $95 = 'US100', $96 = 'US30', $97 = 'US500', $98 = '0', $99 = '', $100 = '500', $101 = '500', $102 = '0', $103 = '0', $104 = '0', $105 = 't', $106 = '10', $107 = '10'
2 31s368ms 28,966 0ms 29ms 1ms SELECT ;Times Reported Time consuming bind #2
Day Hour Count Duration Avg duration 04 28,966 31s368ms 1ms -
SELECT ;
Date: 2026-01-26 04:26:42 Duration: 29ms Database: postgres parameters: $1 = '958', $2 = '958', $3 = '515840218033333300'
-
SELECT ;
Date: 2026-01-26 04:21:08 Duration: 29ms Database: postgres parameters: $1 = '958', $2 = '958', $3 = '515840216980948300'
-
SELECT ;
Date: 2026-01-26 04:42:36 Duration: 27ms Database: postgres parameters: $1 = '958', $2 = '958', $3 = '515840248627472300'
3 2s618ms 1,104 0ms 6ms 2ms SELECT symbolid, ;Times Reported Time consuming bind #3
Day Hour Count Duration Avg duration 04 1,104 2s618ms 2ms -
SELECT symbolid, ;
Date: 2026-01-26 04:00:03 Duration: 6ms Database: postgres parameters: $1 = 'BDSWISS', $2 = '240', $3 = 'USOIL', $4 = 'USDPLN', $5 = 'USDSGD', $6 = 'XAGUSD', $7 = 'USDZAR', $8 = 'USDSEK'
-
SELECT symbolid, ;
Date: 2026-01-26 04:01:36 Duration: 5ms Database: postgres parameters: $1 = 'BDSWISS', $2 = '240', $3 = '#C', $4 = '#CAT', $5 = '#DIS'
-
SELECT symbolid, ;
Date: 2026-01-26 04:30:48 Duration: 4ms Database: postgres parameters: $1 = 'MILLENNIUMPF', $2 = '15', $3 = 'USDJPY.FX'
4 842ms 31,966 0ms 4ms 0ms select 1;Times Reported Time consuming bind #4
Day Hour Count Duration Avg duration 04 31,966 842ms 0ms -
select 1;
Date: 2026-01-26 04:40:21 Duration: 4ms Database: postgres
-
select 1;
Date: 2026-01-26 04:11:44 Duration: 4ms Database: postgres
-
select 1;
Date: 2026-01-26 04:51:48 Duration: 3ms Database: postgres
5 723ms 437 1ms 4ms 1ms SELECT s.symbolid, dss.downloadfrequency, dss.downloadersymbol;Times Reported Time consuming bind #5
Day Hour Count Duration Avg duration 04 437 723ms 1ms -
SELECT s.symbolid, dss.downloadfrequency, dss.downloadersymbol;
Date: 2026-01-26 04:00:55 Duration: 4ms Database: postgres parameters: $1 = 'MILLENNIUMPF'
-
SELECT s.symbolid, dss.downloadfrequency, dss.downloadersymbol;
Date: 2026-01-26 04:00:22 Duration: 2ms Database: postgres parameters: $1 = 'Alpari'
-
SELECT s.symbolid, dss.downloadfrequency, dss.downloadersymbol;
Date: 2026-01-26 04:15:36 Duration: 2ms Database: postgres parameters: $1 = 'FPMARKETS'
6 449ms 75 4ms 12ms 5ms WITH last_candle AS ( ;Times Reported Time consuming bind #6
Day Hour Count Duration Avg duration 04 75 449ms 5ms -
WITH last_candle AS ( ;
Date: 2026-01-26 04:36:02 Duration: 12ms Database: postgres parameters: $1 = '667', $2 = '667'
-
WITH last_candle AS ( ;
Date: 2026-01-26 04:44:01 Duration: 10ms Database: postgres parameters: $1 = '958', $2 = '958'
-
WITH last_candle AS ( ;
Date: 2026-01-26 04:04:00 Duration: 9ms Database: postgres parameters: $1 = '558', $2 = '558'
7 448ms 22 0ms 40ms 20ms with wh_patitioned as ( ;Times Reported Time consuming bind #7
Day Hour Count Duration Avg duration 04 22 448ms 20ms -
with wh_patitioned as ( ;
Date: 2026-01-26 04:06:24 Duration: 40ms Database: postgres parameters: $1 = '558', $2 = '558', $3 = '558', $4 = '558', $5 = '558', $6 = '558', $7 = '558', $8 = '558', $9 = '558'
-
with wh_patitioned as ( ;
Date: 2026-01-26 04:40:01 Duration: 34ms Database: postgres parameters: $1 = '558', $2 = '558', $3 = '558', $4 = '558', $5 = '558', $6 = '558', $7 = '558', $8 = '558', $9 = '558'
-
with wh_patitioned as ( ;
Date: 2026-01-26 04:05:02 Duration: 30ms Database: postgres parameters: $1 = '558', $2 = '558', $3 = '558', $4 = '558', $5 = '558', $6 = '558', $7 = '558', $8 = '558', $9 = '558'
8 420ms 12 28ms 47ms 35ms with sym_info as ( ;Times Reported Time consuming bind #8
Day Hour Count Duration Avg duration 04 12 420ms 35ms -
with sym_info as ( ;
Date: 2026-01-26 04:06:41 Duration: 47ms Database: postgres parameters: $1 = '620', $2 = 'Forex', $3 = 'Forex', $4 = '620', $5 = 'Forex', $6 = '620', $7 = '620', $8 = 'Forex', $9 = '620'
-
with sym_info as ( ;
Date: 2026-01-26 04:06:43 Duration: 44ms Database: postgres parameters: $1 = '617', $2 = 'Forex', $3 = 'Forex', $4 = '617', $5 = 'Forex', $6 = '617', $7 = '617', $8 = 'Forex', $9 = '617'
-
with sym_info as ( ;
Date: 2026-01-26 04:36:57 Duration: 44ms Database: postgres parameters: $1 = '692', $2 = 'Forex', $3 = 'Forex', $4 = '692', $5 = 'Forex', $6 = '692', $7 = '692', $8 = 'Forex', $9 = '692'
9 372ms 37 0ms 17ms 10ms WITH /*Latest.JapSticks*/ all_results AS ( SELECT ;Times Reported Time consuming bind #9
Day Hour Count Duration Avg duration 04 37 372ms 10ms -
WITH /*Latest.JapSticks*/ all_results AS ( SELECT ;
Date: 2026-01-26 04:11:42 Duration: 17ms Database: postgres parameters: $1 = '689', $2 = '0', $3 = '0', $4 = '0', $5 = '', $6 = '0', $7 = '', $8 = '0', $9 = '', $10 = '0', $11 = '0'
-
WITH /*Latest.JapSticks*/ all_results AS ( SELECT ;
Date: 2026-01-26 04:06:09 Duration: 17ms Database: postgres parameters: $1 = '689', $2 = '0', $3 = '0', $4 = '0', $5 = '', $6 = '0', $7 = '', $8 = '0', $9 = '', $10 = '0', $11 = '0'
-
WITH /*Latest.JapSticks*/ all_results AS ( SELECT ;
Date: 2026-01-26 04:18:09 Duration: 13ms Database: postgres parameters: $1 = '667', $2 = '0', $3 = '0', $4 = '0', $5 = '', $6 = '0', $7 = '', $8 = '0', $9 = '', $10 = '0', $11 = '0'
10 225ms 3,247 0ms 0ms 0ms INSERT INTO T30 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;Times Reported Time consuming bind #10
Day Hour Count Duration Avg duration 04 3,247 225ms 0ms -
INSERT INTO T30 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;
Date: 2026-01-26 04:10:58 Duration: 0ms Database: postgres parameters: $1 = '2026-01-26 03:00:00', $2 = '49025.95', $3 = '49031.35', $4 = '48955.05', $5 = '48958.55', $6 = '8965', $7 = '515840248000726300', $8 = '0', $9 = '2026-01-26 04:10:58.747', $10 = '2026-01-26 04:10:58.661', $11 = '49025.95', $12 = '49031.35', $13 = '48955.05', $14 = '48958.55', $15 = '8965', $16 = '0', $17 = '2026-01-26 04:10:58.747', $18 = '2026-01-26 04:10:58.661'
-
INSERT INTO T30 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;
Date: 2026-01-26 04:31:49 Duration: 0ms Database: postgres parameters: $1 = '2026-01-26 04:00:00', $2 = '154.4275', $3 = '154.4835', $4 = '153.9485', $5 = '154.032', $6 = '5401', $7 = '515840230604866300', $8 = '0', $9 = '2026-01-26 04:31:49.839', $10 = '2026-01-26 04:31:49.838', $11 = '154.4275', $12 = '154.4835', $13 = '153.9485', $14 = '154.032', $15 = '5401', $16 = '0', $17 = '2026-01-26 04:31:49.839', $18 = '2026-01-26 04:31:49.838'
-
INSERT INTO T30 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;
Date: 2026-01-26 04:31:29 Duration: 0ms Database: postgres parameters: $1 = '2026-01-26 04:00:00', $2 = '0.59608', $3 = '0.59697', $4 = '0.59593', $5 = '0.59692', $6 = '2553', $7 = '515840230543033300', $8 = '0', $9 = '2026-01-26 04:31:29.775', $10 = '2026-01-26 04:31:29.775', $11 = '0.59608', $12 = '0.59697', $13 = '0.59593', $14 = '0.59692', $15 = '2553', $16 = '0', $17 = '2026-01-26 04:31:29.775', $18 = '2026-01-26 04:31:29.775'
11 225ms 5,558 0ms 0ms 0ms INSERT INTO T15 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;Times Reported Time consuming bind #11
Day Hour Count Duration Avg duration 04 5,558 225ms 0ms -
INSERT INTO T15 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;
Date: 2026-01-26 04:45:47 Duration: 0ms Database: postgres parameters: $1 = '2026-01-26 04:30:00', $2 = '112.6', $3 = '112.635', $4 = '112.506', $5 = '112.551', $6 = '2119', $7 = '515840247879935300', $8 = '0', $9 = '2026-01-26 04:45:47.141', $10 = '2026-01-26 04:45:46.939', $11 = '112.6', $12 = '112.635', $13 = '112.506', $14 = '112.551', $15 = '2119', $16 = '0', $17 = '2026-01-26 04:45:47.141', $18 = '2026-01-26 04:45:46.939'
-
INSERT INTO T15 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;
Date: 2026-01-26 04:35:12 Duration: 0ms Database: postgres parameters: $1 = '2026-01-26 04:15:00', $2 = '1.159945', $3 = '1.160555', $4 = '1.159925', $5 = '1.16043', $6 = '1520', $7 = '605679104134446300', $8 = '0', $9 = '2026-01-26 04:35:12.634', $10 = '2026-01-26 04:35:12.513', $11 = '1.159945', $12 = '1.160555', $13 = '1.159925', $14 = '1.16043', $15 = '1520', $16 = '0', $17 = '2026-01-26 04:35:12.634', $18 = '2026-01-26 04:35:12.513'
-
INSERT INTO T15 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;
Date: 2026-01-26 04:47:46 Duration: 0ms Database: postgres parameters: $1 = '2026-01-23 21:30:00', $2 = '8871.4', $3 = '8874.9', $4 = '8868.15', $5 = '8871.65', $6 = '1194', $7 = '515840248015086300', $8 = '0', $9 = '2026-01-26 04:47:46.002', $10 = '2026-01-26 04:47:45.939', $11 = '8871.4', $12 = '8874.9', $13 = '8868.15', $14 = '8871.65', $15 = '1194', $16 = '0', $17 = '2026-01-26 04:47:46.002', $18 = '2026-01-26 04:47:45.939'
12 160ms 2,220 0ms 0ms 0ms INSERT INTO T60 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;Times Reported Time consuming bind #12
Day Hour Count Duration Avg duration 04 2,220 160ms 0ms -
INSERT INTO T60 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;
Date: 2026-01-26 04:10:54 Duration: 0ms Database: postgres parameters: $1 = '2026-01-23 18:00:00', $2 = '26715.9', $3 = '26755.8', $4 = '26699.9', $5 = '26732.9', $6 = '8142', $7 = '515840247933961300', $8 = '0', $9 = '2026-01-26 04:10:54.615', $10 = '2026-01-26 04:10:54.573', $11 = '26715.9', $12 = '26755.8', $13 = '26699.9', $14 = '26732.9', $15 = '8142', $16 = '0', $17 = '2026-01-26 04:10:54.615', $18 = '2026-01-26 04:10:54.573'
-
INSERT INTO T60 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;
Date: 2026-01-26 04:11:45 Duration: 0ms Database: postgres parameters: $1 = '2026-01-23 20:00:00', $2 = '8862.4', $3 = '8873.4', $4 = '8859.4', $5 = '8872.35', $6 = '6066', $7 = '515840248015562300', $8 = '0', $9 = '2026-01-26 04:11:45.196', $10 = '2026-01-26 04:11:45.116', $11 = '8862.4', $12 = '8873.4', $13 = '8859.4', $14 = '8872.35', $15 = '6066', $16 = '0', $17 = '2026-01-26 04:11:45.196', $18 = '2026-01-26 04:11:45.116'
-
INSERT INTO T60 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;
Date: 2026-01-26 04:11:59 Duration: 0ms Database: postgres parameters: $1 = '2026-01-26 03:00:00', $2 = '49025.95', $3 = '49031.35', $4 = '48946.55', $5 = '48972.55', $6 = '16506', $7 = '515840248000890300', $8 = '0', $9 = '2026-01-26 04:11:59.418', $10 = '2026-01-26 04:11:59.279', $11 = '49025.95', $12 = '49031.35', $13 = '48946.55', $14 = '48972.55', $15 = '16506', $16 = '0', $17 = '2026-01-26 04:11:59.418', $18 = '2026-01-26 04:11:59.279'
13 105ms 1,244 0ms 0ms 0ms INSERT INTO T240 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;Times Reported Time consuming bind #13
Day Hour Count Duration Avg duration 04 1,244 105ms 0ms -
INSERT INTO T240 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;
Date: 2026-01-26 04:00:53 Duration: 0ms Database: postgres parameters: $1 = '2026-01-26 00:00:00', $2 = '0.91991', $3 = '0.92308', $4 = '0.91885', $5 = '0.9218', $6 = '19837', $7 = '515840233446622300', $8 = '0', $9 = '2026-01-26 04:00:53.977', $10 = '2026-01-26 04:00:53.977', $11 = '0.91991', $12 = '0.92308', $13 = '0.91885', $14 = '0.9218', $15 = '19837', $16 = '0', $17 = '2026-01-26 04:00:53.977', $18 = '2026-01-26 04:00:53.977'
-
INSERT INTO T240 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;
Date: 2026-01-26 04:01:16 Duration: 0ms Database: postgres parameters: $1 = '2026-01-26 00:00:00', $2 = '1.366805', $3 = '1.368325', $4 = '1.36435', $5 = '1.36608', $6 = '15073', $7 = '515840230507104300', $8 = '0', $9 = '2026-01-26 04:01:16.765', $10 = '2026-01-26 04:01:16.765', $11 = '1.366805', $12 = '1.368325', $13 = '1.36435', $14 = '1.36608', $15 = '15073', $16 = '0', $17 = '2026-01-26 04:01:16.765', $18 = '2026-01-26 04:01:16.765'
-
INSERT INTO T240 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;
Date: 2026-01-26 04:10:46 Duration: 0ms Database: postgres parameters: $1 = '2026-01-23 23:00:00', $2 = '65.16', $3 = '65.389', $4 = '65.159', $5 = '65.358', $6 = '1889', $7 = '515840248004879300', $8 = '0', $9 = '2026-01-26 04:10:46.554', $10 = '2026-01-26 04:10:46.511', $11 = '65.16', $12 = '65.389', $13 = '65.159', $14 = '65.358', $15 = '1889', $16 = '0', $17 = '2026-01-26 04:10:46.554', $18 = '2026-01-26 04:10:46.511'
14 65ms 163 0ms 0ms 0ms SELECT NULL AS TABLE_CAT, n.nspname AS TABLE_SCHEM, c.relname AS TABLE_NAME, CASE n.nspname ~ '^pg_' OR n.nspname = 'information_schema' WHEN true THEN CASE WHEN n.nspname = 'pg_catalog' OR n.nspname = 'information_schema' THEN CASE c.relkind WHEN 'r' THEN 'SYSTEM TABLE' WHEN 'v' THEN 'SYSTEM VIEW' WHEN 'i' THEN 'SYSTEM INDEX' ELSE NULL END WHEN n.nspname = 'pg_toast' THEN CASE c.relkind WHEN 'r' THEN 'SYSTEM TOAST TABLE' WHEN 'i' THEN 'SYSTEM TOAST INDEX' ELSE NULL END ELSE CASE c.relkind WHEN 'r' THEN 'TEMPORARY TABLE' WHEN 'p' THEN 'TEMPORARY TABLE' WHEN 'i' THEN 'TEMPORARY INDEX' WHEN 'S' THEN 'TEMPORARY SEQUENCE' WHEN 'v' THEN 'TEMPORARY VIEW' ELSE NULL END END WHEN false THEN CASE c.relkind WHEN 'r' THEN 'TABLE' WHEN 'p' THEN 'PARTITIONED TABLE' WHEN 'i' THEN 'INDEX' WHEN 'S' THEN 'SEQUENCE' WHEN 'v' THEN 'VIEW' WHEN 'c' THEN 'TYPE' WHEN 'f' THEN 'FOREIGN TABLE' WHEN 'm' THEN 'MATERIALIZED VIEW' ELSE NULL END ELSE NULL END AS TABLE_TYPE, d.description AS REMARKS, '' as TYPE_CAT, '' as TYPE_SCHEM, '' as TYPE_NAME, '' AS SELF_REFERENCING_COL_NAME, '' AS REF_GENERATION FROM pg_catalog.pg_namespace n, pg_catalog.pg_class c LEFT JOIN pg_catalog.pg_description d ON (c.oid = d.objoid AND d.objsubid = 0) LEFT JOIN pg_catalog.pg_class dc ON (d.classoid = dc.oid AND dc.relname = 'pg_class') LEFT JOIN pg_catalog.pg_namespace dn ON (dn.oid = dc.relnamespace AND dn.nspname = 'pg_catalog') WHERE c.relnamespace = n.oid AND c.relname LIKE 'PROBABLYNOT' AND (false OR (c.relkind = 'r' AND n.nspname !~ '^pg_' AND n.nspname <> 'information_schema')) ORDER BY TABLE_TYPE, TABLE_SCHEM, TABLE_NAME;Times Reported Time consuming bind #14
Day Hour Count Duration Avg duration 04 163 65ms 0ms -
SELECT NULL AS TABLE_CAT, n.nspname AS TABLE_SCHEM, c.relname AS TABLE_NAME, CASE n.nspname ~ '^pg_' OR n.nspname = 'information_schema' WHEN true THEN CASE WHEN n.nspname = 'pg_catalog' OR n.nspname = 'information_schema' THEN CASE c.relkind WHEN 'r' THEN 'SYSTEM TABLE' WHEN 'v' THEN 'SYSTEM VIEW' WHEN 'i' THEN 'SYSTEM INDEX' ELSE NULL END WHEN n.nspname = 'pg_toast' THEN CASE c.relkind WHEN 'r' THEN 'SYSTEM TOAST TABLE' WHEN 'i' THEN 'SYSTEM TOAST INDEX' ELSE NULL END ELSE CASE c.relkind WHEN 'r' THEN 'TEMPORARY TABLE' WHEN 'p' THEN 'TEMPORARY TABLE' WHEN 'i' THEN 'TEMPORARY INDEX' WHEN 'S' THEN 'TEMPORARY SEQUENCE' WHEN 'v' THEN 'TEMPORARY VIEW' ELSE NULL END END WHEN false THEN CASE c.relkind WHEN 'r' THEN 'TABLE' WHEN 'p' THEN 'PARTITIONED TABLE' WHEN 'i' THEN 'INDEX' WHEN 'S' THEN 'SEQUENCE' WHEN 'v' THEN 'VIEW' WHEN 'c' THEN 'TYPE' WHEN 'f' THEN 'FOREIGN TABLE' WHEN 'm' THEN 'MATERIALIZED VIEW' ELSE NULL END ELSE NULL END AS TABLE_TYPE, d.description AS REMARKS, '' as TYPE_CAT, '' as TYPE_SCHEM, '' as TYPE_NAME, '' AS SELF_REFERENCING_COL_NAME, '' AS REF_GENERATION FROM pg_catalog.pg_namespace n, pg_catalog.pg_class c LEFT JOIN pg_catalog.pg_description d ON (c.oid = d.objoid AND d.objsubid = 0) LEFT JOIN pg_catalog.pg_class dc ON (d.classoid = dc.oid AND dc.relname = 'pg_class') LEFT JOIN pg_catalog.pg_namespace dn ON (dn.oid = dc.relnamespace AND dn.nspname = 'pg_catalog') WHERE c.relnamespace = n.oid AND c.relname LIKE 'PROBABLYNOT' AND (false OR (c.relkind = 'r' AND n.nspname !~ '^pg_' AND n.nspname <> 'information_schema')) ORDER BY TABLE_TYPE, TABLE_SCHEM, TABLE_NAME;
Date: 2026-01-26 04:13:19 Duration: 0ms Database: postgres
-
SELECT NULL AS TABLE_CAT, n.nspname AS TABLE_SCHEM, c.relname AS TABLE_NAME, CASE n.nspname ~ '^pg_' OR n.nspname = 'information_schema' WHEN true THEN CASE WHEN n.nspname = 'pg_catalog' OR n.nspname = 'information_schema' THEN CASE c.relkind WHEN 'r' THEN 'SYSTEM TABLE' WHEN 'v' THEN 'SYSTEM VIEW' WHEN 'i' THEN 'SYSTEM INDEX' ELSE NULL END WHEN n.nspname = 'pg_toast' THEN CASE c.relkind WHEN 'r' THEN 'SYSTEM TOAST TABLE' WHEN 'i' THEN 'SYSTEM TOAST INDEX' ELSE NULL END ELSE CASE c.relkind WHEN 'r' THEN 'TEMPORARY TABLE' WHEN 'p' THEN 'TEMPORARY TABLE' WHEN 'i' THEN 'TEMPORARY INDEX' WHEN 'S' THEN 'TEMPORARY SEQUENCE' WHEN 'v' THEN 'TEMPORARY VIEW' ELSE NULL END END WHEN false THEN CASE c.relkind WHEN 'r' THEN 'TABLE' WHEN 'p' THEN 'PARTITIONED TABLE' WHEN 'i' THEN 'INDEX' WHEN 'S' THEN 'SEQUENCE' WHEN 'v' THEN 'VIEW' WHEN 'c' THEN 'TYPE' WHEN 'f' THEN 'FOREIGN TABLE' WHEN 'm' THEN 'MATERIALIZED VIEW' ELSE NULL END ELSE NULL END AS TABLE_TYPE, d.description AS REMARKS, '' as TYPE_CAT, '' as TYPE_SCHEM, '' as TYPE_NAME, '' AS SELF_REFERENCING_COL_NAME, '' AS REF_GENERATION FROM pg_catalog.pg_namespace n, pg_catalog.pg_class c LEFT JOIN pg_catalog.pg_description d ON (c.oid = d.objoid AND d.objsubid = 0) LEFT JOIN pg_catalog.pg_class dc ON (d.classoid = dc.oid AND dc.relname = 'pg_class') LEFT JOIN pg_catalog.pg_namespace dn ON (dn.oid = dc.relnamespace AND dn.nspname = 'pg_catalog') WHERE c.relnamespace = n.oid AND c.relname LIKE 'PROBABLYNOT' AND (false OR (c.relkind = 'r' AND n.nspname !~ '^pg_' AND n.nspname <> 'information_schema')) ORDER BY TABLE_TYPE, TABLE_SCHEM, TABLE_NAME;
Date: 2026-01-26 04:13:20 Duration: 0ms Database: postgres
-
SELECT NULL AS TABLE_CAT, n.nspname AS TABLE_SCHEM, c.relname AS TABLE_NAME, CASE n.nspname ~ '^pg_' OR n.nspname = 'information_schema' WHEN true THEN CASE WHEN n.nspname = 'pg_catalog' OR n.nspname = 'information_schema' THEN CASE c.relkind WHEN 'r' THEN 'SYSTEM TABLE' WHEN 'v' THEN 'SYSTEM VIEW' WHEN 'i' THEN 'SYSTEM INDEX' ELSE NULL END WHEN n.nspname = 'pg_toast' THEN CASE c.relkind WHEN 'r' THEN 'SYSTEM TOAST TABLE' WHEN 'i' THEN 'SYSTEM TOAST INDEX' ELSE NULL END ELSE CASE c.relkind WHEN 'r' THEN 'TEMPORARY TABLE' WHEN 'p' THEN 'TEMPORARY TABLE' WHEN 'i' THEN 'TEMPORARY INDEX' WHEN 'S' THEN 'TEMPORARY SEQUENCE' WHEN 'v' THEN 'TEMPORARY VIEW' ELSE NULL END END WHEN false THEN CASE c.relkind WHEN 'r' THEN 'TABLE' WHEN 'p' THEN 'PARTITIONED TABLE' WHEN 'i' THEN 'INDEX' WHEN 'S' THEN 'SEQUENCE' WHEN 'v' THEN 'VIEW' WHEN 'c' THEN 'TYPE' WHEN 'f' THEN 'FOREIGN TABLE' WHEN 'm' THEN 'MATERIALIZED VIEW' ELSE NULL END ELSE NULL END AS TABLE_TYPE, d.description AS REMARKS, '' as TYPE_CAT, '' as TYPE_SCHEM, '' as TYPE_NAME, '' AS SELF_REFERENCING_COL_NAME, '' AS REF_GENERATION FROM pg_catalog.pg_namespace n, pg_catalog.pg_class c LEFT JOIN pg_catalog.pg_description d ON (c.oid = d.objoid AND d.objsubid = 0) LEFT JOIN pg_catalog.pg_class dc ON (d.classoid = dc.oid AND dc.relname = 'pg_class') LEFT JOIN pg_catalog.pg_namespace dn ON (dn.oid = dc.relnamespace AND dn.nspname = 'pg_catalog') WHERE c.relnamespace = n.oid AND c.relname LIKE 'PROBABLYNOT' AND (false OR (c.relkind = 'r' AND n.nspname !~ '^pg_' AND n.nspname <> 'information_schema')) ORDER BY TABLE_TYPE, TABLE_SCHEM, TABLE_NAME;
Date: 2026-01-26 04:13:20 Duration: 0ms Database: postgres
15 59ms 70 0ms 1ms 0ms SELECT timegranularity FROM brokersymbollist bsl INNER JOIN symbols s ON bsl.symbolid = s.symbolid INNER JOIN downloadersymbolsettings dss on s.symbolid = dss.symbolid LEFT OUTER JOIN brokerinstrumentmapping bdfi ON bdfi.brokerid = $1 AND dss.datafeedinstrumentid = bdfi.datafeedinstrumentid WHERE s.nonliquid = 0 and s.deleted = 0 and dss.enabled = 1 AND s.symbol ILIKE $2 AND bsl.brokerid = $3 AND timegranularity >= 15 ORDER BY timegranularity LIMIT 1;Times Reported Time consuming bind #15
Day Hour Count Duration Avg duration 04 70 59ms 0ms -
SELECT timegranularity FROM brokersymbollist bsl INNER JOIN symbols s ON bsl.symbolid = s.symbolid INNER JOIN downloadersymbolsettings dss on s.symbolid = dss.symbolid LEFT OUTER JOIN brokerinstrumentmapping bdfi ON bdfi.brokerid = $1 AND dss.datafeedinstrumentid = bdfi.datafeedinstrumentid WHERE s.nonliquid = 0 and s.deleted = 0 and dss.enabled = 1 AND s.symbol ILIKE $2 AND bsl.brokerid = $3 AND timegranularity >= 15 ORDER BY timegranularity LIMIT 1;
Date: 2026-01-26 04:20:07 Duration: 1ms Database: postgres parameters: $1 = '689', $2 = 'XAUUSD', $3 = '689'
-
SELECT timegranularity FROM brokersymbollist bsl INNER JOIN symbols s ON bsl.symbolid = s.symbolid INNER JOIN downloadersymbolsettings dss on s.symbolid = dss.symbolid LEFT OUTER JOIN brokerinstrumentmapping bdfi ON bdfi.brokerid = $1 AND dss.datafeedinstrumentid = bdfi.datafeedinstrumentid WHERE s.nonliquid = 0 and s.deleted = 0 and dss.enabled = 1 AND s.symbol ILIKE $2 AND bsl.brokerid = $3 AND timegranularity >= 15 ORDER BY timegranularity LIMIT 1;
Date: 2026-01-26 04:31:10 Duration: 1ms Database: postgres parameters: $1 = '558', $2 = 'CHFJPY', $3 = '558'
-
SELECT timegranularity FROM brokersymbollist bsl INNER JOIN symbols s ON bsl.symbolid = s.symbolid INNER JOIN downloadersymbolsettings dss on s.symbolid = dss.symbolid LEFT OUTER JOIN brokerinstrumentmapping bdfi ON bdfi.brokerid = $1 AND dss.datafeedinstrumentid = bdfi.datafeedinstrumentid WHERE s.nonliquid = 0 and s.deleted = 0 and dss.enabled = 1 AND s.symbol ILIKE $2 AND bsl.brokerid = $3 AND timegranularity >= 15 ORDER BY timegranularity LIMIT 1;
Date: 2026-01-26 04:47:23 Duration: 1ms Database: postgres parameters: $1 = '558', $2 = 'EURUSD', $3 = '558'
16 55ms 8,876 0ms 12ms 0ms SET extra_float_digits = 3;Times Reported Time consuming bind #16
Day Hour Count Duration Avg duration 04 8,876 55ms 0ms -
SET extra_float_digits = 3;
Date: 2026-01-26 04:31:40 Duration: 12ms Database: postgres
-
SET extra_float_digits = 3;
Date: 2026-01-26 04:11:33 Duration: 1ms Database: postgres
-
SET extra_float_digits = 3;
Date: 2026-01-26 04:38:12 Duration: 1ms Database: postgres
17 50ms 1 50ms 50ms 50ms with maxwhid as ( ;Times Reported Time consuming bind #17
Day Hour Count Duration Avg duration 04 1 50ms 50ms -
with maxwhid as ( ;
Date: 2026-01-26 04:11:42 Duration: 50ms Database: postgres parameters: $1 = '335', $2 = '621', $3 = '637', $4 = '642', $5 = '660', $6 = '666', $7 = '643', $8 = '630', $9 = '680', $10 = '641', $11 = '431', $12 = '622', $13 = '489', $14 = '529', $15 = '576', $16 = '665', $17 = '667', $18 = '558', $19 = '620', $20 = '125', $21 = '488', $22 = '567', $23 = '689', $24 = '700', $25 = '758', $26 = '763', $27 = '765', $28 = '817', $29 = '914', $30 = '972'
18 40ms 8 3ms 6ms 5ms SELECT DISTINCT ON (basegroupname, symbol) ;Times Reported Time consuming bind #18
Day Hour Count Duration Avg duration 04 8 40ms 5ms -
SELECT DISTINCT ON (basegroupname, symbol) ;
Date: 2026-01-26 04:31:24 Duration: 6ms Database: postgres parameters: $1 = '627', $2 = '627'
-
SELECT DISTINCT ON (basegroupname, symbol) ;
Date: 2026-01-26 04:45:43 Duration: 6ms Database: postgres parameters: $1 = '958', $2 = '958'
-
SELECT DISTINCT ON (basegroupname, symbol) ;
Date: 2026-01-26 04:46:00 Duration: 5ms Database: postgres parameters: $1 = '627', $2 = '627'
19 39ms 8 2ms 7ms 4ms WITH pre_symbols AS ( /* find relevant symbols */ ;Times Reported Time consuming bind #19
Day Hour Count Duration Avg duration 04 8 39ms 4ms -
WITH pre_symbols AS ( /* find relevant symbols */ ;
Date: 2026-01-26 04:13:19 Duration: 7ms Database: postgres parameters: $1 = '1018', $2 = 'ICMARKETS-AU-MT5', $3 = 'AAPL.NAS', $4 = 'ABBV.NYSE', $5 = 'AMCR.NYSE', $6 = 'AMP.NYSE', $7 = 'AMZN.NAS', $8 = 'ANZ.ASX', $9 = 'AUDJPY', $10 = 'AUDUSD', $11 = 'AUS200', $12 = 'BABA.NYSE', $13 = 'BIIB.NAS', $14 = 'BXB.ASX', $15 = 'CBA.ASX', $16 = 'CHINA50', $17 = 'CSL.ASX', $18 = 'DE30', $19 = 'ES35', $20 = 'EURCHF', $21 = 'EURGBP', $22 = 'EURUSD', $23 = 'F40', $24 = 'FMG.ASX', $25 = 'GBPJPY', $26 = 'GBPUSD', $27 = 'GOOG.NAS', $28 = 'HK50', $29 = 'IT40', $30 = 'JP225', $31 = 'KO.NYSE', $32 = 'MQG.ASX', $33 = 'MSFT.NAS', $34 = 'NAB.ASX', $35 = 'NFLX.NAS', $36 = 'PYPL.NAS', $37 = 'QBE.ASX', $38 = 'STOXX50', $39 = 'SUN.ASX', $40 = 'TCL.ASX', $41 = 'TLS.ASX', $42 = 'TSLA.NAS', $43 = 'UK100', $44 = 'UNH.NYSE', $45 = 'US2000', $46 = 'US30', $47 = 'US500', $48 = 'USDCAD', $49 = 'USDCHF', $50 = 'USDCNH', $51 = 'USDJPY', $52 = 'USTEC', $53 = 'WBC.ASX', $54 = 'WES.ASX', $55 = 'WOW.ASX', $56 = 'WPL.ASX', $57 = 'XAUEUR', $58 = 'XAUUSD', $59 = 'XBRUSD', $60 = 'XTIUSD', $61 = 'AAPL.NAS', $62 = 'ABBV.NYSE', $63 = 'AMCR.NYSE', $64 = 'AMP.NYSE', $65 = 'AMZN.NAS', $66 = 'ANZ.ASX', $67 = 'AUDJPY', $68 = 'AUDUSD', $69 = 'AUS200', $70 = 'BABA.NYSE', $71 = 'BIIB.NAS', $72 = 'BXB.ASX', $73 = 'CBA.ASX', $74 = 'CHINA50', $75 = 'CSL.ASX', $76 = 'DE30', $77 = 'ES35', $78 = 'EURCHF', $79 = 'EURGBP', $80 = 'EURUSD', $81 = 'F40', $82 = 'FMG.ASX', $83 = 'GBPJPY', $84 = 'GBPUSD', $85 = 'GOOG.NAS', $86 = 'HK50', $87 = 'IT40', $88 = 'JP225', $89 = 'KO.NYSE', $90 = 'MQG.ASX', $91 = 'MSFT.NAS', $92 = 'NAB.ASX', $93 = 'NFLX.NAS', $94 = 'PYPL.NAS', $95 = 'QBE.ASX', $96 = 'STOXX50', $97 = 'SUN.ASX', $98 = 'TCL.ASX', $99 = 'TLS.ASX', $100 = 'TSLA.NAS', $101 = 'UK100', $102 = 'UNH.NYSE', $103 = 'US2000', $104 = 'US30', $105 = 'US500', $106 = 'USDCAD', $107 = 'USDCHF', $108 = 'USDCNH', $109 = 'USDJPY', $110 = 'USTEC', $111 = 'WBC.ASX', $112 = 'WES.ASX', $113 = 'WOW.ASX', $114 = 'WPL.ASX', $115 = 'XAUEUR', $116 = 'XAUUSD', $117 = 'XBRUSD', $118 = 'XTIUSD'
-
WITH pre_symbols AS ( /* find relevant symbols */ ;
Date: 2026-01-26 04:13:18 Duration: 7ms Database: postgres parameters: $1 = '1018', $2 = 'ICMARKETS-AU-MT5', $3 = 'AAPL.NAS', $4 = 'ABBV.NYSE', $5 = 'AMCR.NYSE', $6 = 'AMP.NYSE', $7 = 'AMZN.NAS', $8 = 'ANZ.ASX', $9 = 'AUDJPY', $10 = 'AUDUSD', $11 = 'AUS200', $12 = 'BABA.NYSE', $13 = 'BIIB.NAS', $14 = 'BXB.ASX', $15 = 'CBA.ASX', $16 = 'CHINA50', $17 = 'CSL.ASX', $18 = 'DE30', $19 = 'ES35', $20 = 'EURCHF', $21 = 'EURGBP', $22 = 'EURUSD', $23 = 'F40', $24 = 'FMG.ASX', $25 = 'GBPJPY', $26 = 'GBPUSD', $27 = 'GOOG.NAS', $28 = 'HK50', $29 = 'IT40', $30 = 'JP225', $31 = 'KO.NYSE', $32 = 'MQG.ASX', $33 = 'MSFT.NAS', $34 = 'NAB.ASX', $35 = 'NFLX.NAS', $36 = 'PYPL.NAS', $37 = 'QBE.ASX', $38 = 'STOXX50', $39 = 'SUN.ASX', $40 = 'TCL.ASX', $41 = 'TLS.ASX', $42 = 'TSLA.NAS', $43 = 'UK100', $44 = 'UNH.NYSE', $45 = 'US2000', $46 = 'US30', $47 = 'US500', $48 = 'USDCAD', $49 = 'USDCHF', $50 = 'USDCNH', $51 = 'USDJPY', $52 = 'USTEC', $53 = 'WBC.ASX', $54 = 'WES.ASX', $55 = 'WOW.ASX', $56 = 'WPL.ASX', $57 = 'XAUEUR', $58 = 'XAUUSD', $59 = 'XBRUSD', $60 = 'XTIUSD', $61 = 'AAPL.NAS', $62 = 'ABBV.NYSE', $63 = 'AMCR.NYSE', $64 = 'AMP.NYSE', $65 = 'AMZN.NAS', $66 = 'ANZ.ASX', $67 = 'AUDJPY', $68 = 'AUDUSD', $69 = 'AUS200', $70 = 'BABA.NYSE', $71 = 'BIIB.NAS', $72 = 'BXB.ASX', $73 = 'CBA.ASX', $74 = 'CHINA50', $75 = 'CSL.ASX', $76 = 'DE30', $77 = 'ES35', $78 = 'EURCHF', $79 = 'EURGBP', $80 = 'EURUSD', $81 = 'F40', $82 = 'FMG.ASX', $83 = 'GBPJPY', $84 = 'GBPUSD', $85 = 'GOOG.NAS', $86 = 'HK50', $87 = 'IT40', $88 = 'JP225', $89 = 'KO.NYSE', $90 = 'MQG.ASX', $91 = 'MSFT.NAS', $92 = 'NAB.ASX', $93 = 'NFLX.NAS', $94 = 'PYPL.NAS', $95 = 'QBE.ASX', $96 = 'STOXX50', $97 = 'SUN.ASX', $98 = 'TCL.ASX', $99 = 'TLS.ASX', $100 = 'TSLA.NAS', $101 = 'UK100', $102 = 'UNH.NYSE', $103 = 'US2000', $104 = 'US30', $105 = 'US500', $106 = 'USDCAD', $107 = 'USDCHF', $108 = 'USDCNH', $109 = 'USDJPY', $110 = 'USTEC', $111 = 'WBC.ASX', $112 = 'WES.ASX', $113 = 'WOW.ASX', $114 = 'WPL.ASX', $115 = 'XAUEUR', $116 = 'XAUUSD', $117 = 'XBRUSD', $118 = 'XTIUSD', $119 = '5'
-
WITH pre_symbols AS ( /* find relevant symbols */ ;
Date: 2026-01-26 04:13:18 Duration: 6ms Database: postgres parameters: $1 = '1018', $2 = 'ICMARKETS-AU-MT5', $3 = 'AAPL.NAS', $4 = 'ABBV.NYSE', $5 = 'AMCR.NYSE', $6 = 'AMP.NYSE', $7 = 'AMZN.NAS', $8 = 'ANZ.ASX', $9 = 'AUDJPY', $10 = 'AUDUSD', $11 = 'AUS200', $12 = 'BABA.NYSE', $13 = 'BIIB.NAS', $14 = 'BXB.ASX', $15 = 'CBA.ASX', $16 = 'CHINA50', $17 = 'CSL.ASX', $18 = 'DE30', $19 = 'ES35', $20 = 'EURCHF', $21 = 'EURGBP', $22 = 'EURUSD', $23 = 'F40', $24 = 'FMG.ASX', $25 = 'GBPJPY', $26 = 'GBPUSD', $27 = 'GOOG.NAS', $28 = 'HK50', $29 = 'IT40', $30 = 'JP225', $31 = 'KO.NYSE', $32 = 'MQG.ASX', $33 = 'MSFT.NAS', $34 = 'NAB.ASX', $35 = 'NFLX.NAS', $36 = 'PYPL.NAS', $37 = 'QBE.ASX', $38 = 'STOXX50', $39 = 'SUN.ASX', $40 = 'TCL.ASX', $41 = 'TLS.ASX', $42 = 'TSLA.NAS', $43 = 'UK100', $44 = 'UNH.NYSE', $45 = 'US2000', $46 = 'US30', $47 = 'US500', $48 = 'USDCAD', $49 = 'USDCHF', $50 = 'USDCNH', $51 = 'USDJPY', $52 = 'USTEC', $53 = 'WBC.ASX', $54 = 'WES.ASX', $55 = 'WOW.ASX', $56 = 'WPL.ASX', $57 = 'XAUEUR', $58 = 'XAUUSD', $59 = 'XBRUSD', $60 = 'XTIUSD', $61 = 'AAPL.NAS', $62 = 'ABBV.NYSE', $63 = 'AMCR.NYSE', $64 = 'AMP.NYSE', $65 = 'AMZN.NAS', $66 = 'ANZ.ASX', $67 = 'AUDJPY', $68 = 'AUDUSD', $69 = 'AUS200', $70 = 'BABA.NYSE', $71 = 'BIIB.NAS', $72 = 'BXB.ASX', $73 = 'CBA.ASX', $74 = 'CHINA50', $75 = 'CSL.ASX', $76 = 'DE30', $77 = 'ES35', $78 = 'EURCHF', $79 = 'EURGBP', $80 = 'EURUSD', $81 = 'F40', $82 = 'FMG.ASX', $83 = 'GBPJPY', $84 = 'GBPUSD', $85 = 'GOOG.NAS', $86 = 'HK50', $87 = 'IT40', $88 = 'JP225', $89 = 'KO.NYSE', $90 = 'MQG.ASX', $91 = 'MSFT.NAS', $92 = 'NAB.ASX', $93 = 'NFLX.NAS', $94 = 'PYPL.NAS', $95 = 'QBE.ASX', $96 = 'STOXX50', $97 = 'SUN.ASX', $98 = 'TCL.ASX', $99 = 'TLS.ASX', $100 = 'TSLA.NAS', $101 = 'UK100', $102 = 'UNH.NYSE', $103 = 'US2000', $104 = 'US30', $105 = 'US500', $106 = 'USDCAD', $107 = 'USDCHF', $108 = 'USDCNH', $109 = 'USDJPY', $110 = 'USTEC', $111 = 'WBC.ASX', $112 = 'WES.ASX', $113 = 'WOW.ASX', $114 = 'WPL.ASX', $115 = 'XAUEUR', $116 = 'XAUUSD', $117 = 'XBRUSD', $118 = 'XTIUSD', $119 = '5'
20 37ms 8,850 0ms 1ms 0ms SET application_name = 'PostgreSQL JDBC Driver';Times Reported Time consuming bind #20
Day Hour Count Duration Avg duration 04 8,850 37ms 0ms -
SET application_name = 'PostgreSQL JDBC Driver';
Date: 2026-01-26 04:11:09 Duration: 1ms Database: postgres
-
SET application_name = 'PostgreSQL JDBC Driver';
Date: 2026-01-26 04:31:18 Duration: 1ms Database: postgres
-
SET application_name = 'PostgreSQL JDBC Driver';
Date: 2026-01-26 04:06:12 Duration: 0ms Database: postgres
-
Events
Log levels
Key values
- 546,137 Log entries
Events distribution
Key values
- 0 PANIC entries
- 0 FATAL entries
- 36 ERROR entries
- 0 WARNING entries
Most Frequent Errors/Events
Key values
- 34 Max number of times the same event was reported
- 36 Total events found
Rank Times reported Error 1 34 ERROR: function fixcandlegaps(...) is not unique
Times Reported Most Frequent Error / Event #1
Day Hour Count Jan 26 04 34 - ERROR: function fixcandlegaps(unknown, boolean) is not unique at character 8
Hint: Could not choose a best candidate function. You might need to add explicit type casts.
Statement: select fixcandlegaps('GLOBALFXMT5', false);Date: 2026-01-26 04:06:01
2 1 ERROR: relation "..." does not exist
Times Reported Most Frequent Error / Event #2
Day Hour Count Jan 26 04 1 - ERROR: relation "t0" does not exist at character 83
Statement: SELECT * FROM ( SELECT PriceDateTime, Open, High, Low, Close, Volume, BSF FROM T0 WHERE symbolid = $1 AND (BSF = 0 OR BSF IS NULL) ORDER BY PriceDateTime DESC LIMIT 1050 ) a ORDER BY PriceDateTime ASC
Date: 2026-01-26 04:39:19
3 1 ERROR: duplicate key value violates unique constraint "..."
Times Reported Most Frequent Error / Event #3
Day Hour Count Jan 26 04 1 - ERROR: duplicate key value violates unique constraint "ar_idx_sa_hist_conscandles_symbolid_datetime_direction"
Detail: Key (symbolid, datetime, direction)=(500991627550361200, 2025-10-15 05:00:00, 1) already exists.
Context: SQL statement "with aa as ( select symbolid, direction, "datetime", row_number() over(partition by symbolid,direction order by datetime desc) r from sa_hist_consecutivecandles ), sel AS ( select distinct sa.* from sa_hist_consecutivecandles sa inner join (select * from aa where r = 1) b on sa.symbolid = b.symbolid INNER JOIN downloadersymbolsettings dss on sa.symbolid = dss.symbolid where sa.direction = b.direction and sa.datetime != b.datetime AND (dss.archive is null OR dss.archive is true) LIMIT v_NumRecordsToArchive ), del AS ( DELETE FROM sa_hist_consecutivecandles USING sel WHERE sa_hist_consecutivecandles.symbolid = sel.symbolid and sa_hist_consecutivecandles."datetime" = sel.datetime and sa_hist_consecutivecandles.direction = sel.direction RETURNING sa_hist_consecutivecandles.* ) INSERT INTO archive.sa_hist_consecutivecandles (symbolid, "datetime", image, qty, percentile, direction, lastupdated, height) SELECT symbolid, "datetime", image, qty, percentile, direction, lastupdated, height FROM del WHERE NOT EXISTS (select 1 FROM archive.sa_hist_consecutivecandles WHERE symbolid = del.symbolid and direction = del.direction and "datetime" = del."datetime")" PL/pgSQL function cleanupconsecutivecandlesresults(integer,integer) line 8 at SQL statement
Statement: select cleanupconsecutivecandlesresults()Date: 2026-01-26 04:05:08