-
Global information
- Generated on Wed Jan 21 04:59:37 2026
- Log file: /home/postgres/pg_data/data/pg_log/postgresql-2026-01-21_060000.log
- Parsed 1,370,998 log entries in 35s
- Log start from 2026-01-21 06:00:00 to 2026-01-21 06:59:34
-
Overview
Global Stats
- 1,039 Number of unique normalized queries
- 174,614 Number of queries
- 3h36m32s Total query duration
- 2026-01-21 06:00:00 First query
- 2026-01-21 06:59:34 Last query
- 5,756 queries/s at 2026-01-21 06:19:31 Query peak
- 3h36m32s Total query duration
- 4m25s Prepare/parse total duration
- 55s7ms Bind total duration
- 3h31m11s Execute total duration
- 283 Number of events
- 4 Number of unique normalized events
- 239 Max number of times the same event was reported
- 0 Number of cancellation
- 36 Total number of automatic vacuums
- 51 Total number of automatic analyzes
- 700 Number temporary file
- 167.60 MiB Max size of temporary file
- 7.67 MiB Average size of temporary file
- 2,878 Total number of sessions
- 12 sessions at 2026-01-21 06:57:10 Session peak
- 2d7h2m16s Total duration of sessions
- 1m8s Average duration of sessions
- 60 Average queries per session
- 4s514ms Average queries duration per session
- 1m4s Average idle time per session
- 2,878 Total number of connections
- 34 connections/s at 2026-01-21 06:06:01 Connection peak
- 6 Total number of databases
SQL Traffic
Key values
- 5,756 queries/s Query Peak
- 2026-01-21 06:19:31 Date
SELECT Traffic
Key values
- 2,862 queries/s Query Peak
- 2026-01-21 06:19:31 Date
INSERT/UPDATE/DELETE Traffic
Key values
- 236 queries/s Query Peak
- 2026-01-21 06:30:51 Date
Queries duration
Key values
- 3h36m32s Total query duration
Prepared queries ratio
Key values
- 0.00 Ratio of bind vs prepare
- 0.00 % Ratio between prepared and "usual" statements
General Activity
↑ Back to the top of the General Activity tableDay Hour Count Min duration Max duration Avg duration Latency Percentile(90) Latency Percentile(95) Latency Percentile(99) Jan 21 06 174,614 0ms 14m19s 72ms 5m36s 6m33s 20m36s Day Hour SELECT COPY TO Average Duration Latency Percentile(90) Latency Percentile(95) Latency Percentile(99) Jan 21 06 45,960 601 0ms 0ms 0ms 0ms Day Hour INSERT UPDATE DELETE COPY FROM Average Duration Latency Percentile(90) Latency Percentile(95) Latency Percentile(99) Jan 21 06 28,604 1,820 13 78 0ms 0ms 0ms 0ms Day Hour Prepare Bind Bind/Prepare Percentage of prepare Jan 21 06 18,866 51,107 2.71 22.85% Day Hour Count Average / Second Jan 21 06 2,878 0.80/s Day Hour Count Average Duration Average idle time Jan 21 06 2,878 1m8s 1m4s -
Connections
Established Connections
Key values
- 34 connections Connection Peak
- 2026-01-21 06:06:01 Date
Connections per database
Key values
- acaweb_fx Main Database
- 2,878 connections Total
Connections per user
Key values
- postgres Main User
- 2,878 connections Total
Connections per host
Key values
- 192.168.4.142 Main host with 1105 connections
- 2,878 Total connections
Host Count 127.0.0.1 116 192.168.0.114 2 192.168.0.216 101 192.168.0.74 118 192.168.1.145 188 192.168.1.15 111 192.168.1.20 204 192.168.1.231 20 192.168.1.239 2 192.168.1.90 64 192.168.2.126 62 192.168.2.182 12 192.168.2.82 48 192.168.3.199 36 192.168.4.142 1,105 192.168.4.150 10 192.168.4.221 1 192.168.4.224 4 192.168.4.238 16 192.168.4.33 90 192.168.4.98 330 [local] 238 -
Sessions
Simultaneous sessions
Key values
- 12 sessions Session Peak
- 2026-01-21 06:57:10 Date
Histogram of session times
Key values
- 2,235 0-500ms duration
Sessions per database
Key values
- acaweb_fx Main Database
- 2,878 sessions Total
Sessions per user
Key values
- postgres Main User
- 2,878 sessions Total
Sessions per host
Key values
- 192.168.4.142 Main Host
- 2,878 sessions Total
Host Count Total Duration Average Duration 127.0.0.1 116 36m36s 18s939ms 192.168.0.114 2 10m19s 5m9s 192.168.0.216 101 1m2s 618ms 192.168.0.74 118 4h21m52s 2m13s 192.168.1.145 188 3h22m54s 1m4s 192.168.1.15 111 2h9m49s 1m10s 192.168.1.20 204 13h34m2s 3m59s 192.168.1.231 20 9h52m18s 29m36s 192.168.1.239 2 16ms 8ms 192.168.1.90 64 5m28s 5s132ms 192.168.2.126 62 7s505ms 121ms 192.168.2.182 12 949ms 79ms 192.168.2.82 48 15s144ms 315ms 192.168.3.199 36 1s353ms 37ms 192.168.4.142 1,105 8m29s 460ms 192.168.4.150 10 20h14m3s 2h1m24s 192.168.4.221 1 243ms 243ms 192.168.4.224 4 44s200ms 11s50ms 192.168.4.238 16 20s223ms 1s263ms 192.168.4.33 90 1m50s 1s230ms 192.168.4.98 330 29s808ms 90ms [local] 238 21m29s 5s418ms -
Checkpoints / Restartpoints
Checkpoints Buffers
Key values
- 12,265 buffers Checkpoint Peak
- 2026-01-21 06:06:36 Date
- 209.956 seconds Highest write time
- 0.119 seconds Sync time
Checkpoints Wal files
Key values
- 5 files Wal files usage Peak
- 2026-01-21 06:06:36 Date
Checkpoints distance
Key values
- 173.33 Mo Distance Peak
- 2026-01-21 06:06:36 Date
Checkpoints Activity
↑ Back to the top of the Checkpoint Activity tableDay Hour Written buffers Write time Sync time Total time Jan 21 06 37,233 1,906.336s 0.155s 1,906.825s Day Hour Added Removed Recycled Synced files Longest sync Average sync Jan 21 06 0 0 22 1,903 0.117s 0s Day Hour Count Avg time (sec) Jan 21 06 0 0s Day Hour Mean distance Mean estimate Jan 21 06 30,547.08 kB 63,224.08 kB -
Temporary Files
Size of temporary files
Key values
- 181.97 MiB Temp Files size Peak
- 2026-01-21 06:50:10 Date
Number of temporary files
Key values
- 30 per second Temp Files Peak
- 2026-01-21 06:32:31 Date
Temporary Files Activity
↑ Back to the top of the Temporary Files Activity tableDay Hour Count Total size Average size Jan 21 06 700 5.24 GiB 7.67 MiB Queries generating the most temporary files (N)
Rank Count Total size Min size Max size Avg size Query 1 76 262.80 MiB 3.21 MiB 3.84 MiB 3.46 MiB select resultuid from relevance_consecutivecandles_results order by resultuid desc limit ?), all_results as ( select ccr.resultuid as resultuid, ccr.direction as direction, s.exchange as exchange, s.symbolid as symbolid, coalesce(bim.code, s.symbol) as symbol_code, s.longname as symbol_name, s.timegranularity as interval, ccr.patternendtime as identified, dtt.timezone as timezone, ccr.qtyconsecutivecandles as length, g.basegroupname, case when rcr.age is not null then rcr.age when ccr.resultuid <= rm.resultuid then ? else ? end as age, case when rcr.relevant is not null then rcr.relevant when ccr.resultuid <= rm.resultuid then ? else ? end as relevant, cps.pip, newlevels.filtered from consecutivecandles_results ccr inner join brokersymbollist bsl on bsl.brokerid = ? and bsl.symbolid = ccr.symbolid inner join symbols s on ccr.symbolid = s.symbolid and s.nonliquid = ? inner join downloadersymbolsettings dss on ccr.symbolid = dss.symbolid inner join datafeedstimetable dtt on dss.classname = dtt.classname and dtt.dayofweek = ? inner join symbolgroup sg on ccr.symbolid = sg.symbolid inner join groups g on sg.groupid = g.groupid inner join brokergroups bg on g.groupid = bg.groupid and bsl.brokerid = bg.brokerid inner join rar_max rm on ? = ? left outer join relevance_consecutivecandles_results rcr on rcr.resultuid = ccr.resultuid left join currencypips cps on cps.symbol = s.symbol left outer join brokerinstrumentmap bim on dss.datafeedinstrumentid = bim.datafeedinstrumentid and bim.brokerid = bsl.brokerid and bim.type = ? left join lateral calc_cc_signal_filter (ccr.resultuid) newlevels on true where ccr.gmttimefound > now() - interval ? and s.deleted = ? and (ccr.simulation = ? or ccr.simulation is null) and (? = ? or s.timegranularity in (...)) and (? = ? or s.exchange in (...)) and (? = ? or coalesce(bim.code, s.symbol) in (...)) and (? = ? or ccr.patternlengthbars <= ?)), results as ( select distinct on (symbolid) * from all_results where (false = ? or relevant = ?) and (? = ? or age <= ?) order by symbolid, resultuid ) select * from results order by identified desc, length desc;-
SELECT resultuid FROM relevance_consecutivecandles_results ORDER BY resultuid DESC LIMIT 1), all_results AS ( SELECT ccr.resultuid AS resultuid, ccr.direction AS direction, s.exchange AS exchange, s.symbolid AS symbolid, coalesce(bim.code, s.symbol) AS symbol_code, s.longname AS symbol_name, s.timegranularity AS interval, ccr.patternendtime AS identified, dtt.timezone AS timezone, ccr.qtyconsecutivecandles AS length, g.basegroupname, CASE WHEN rcr.age IS NOT NULL THEN rcr.age WHEN ccr.resultuid <= rm.resultuid THEN 1 ELSE 0 END as age, CASE WHEN rcr.relevant IS NOT NULL THEN rcr.relevant WHEN ccr.resultuid <= rm.resultuid THEN 0 ELSE 1 END as relevant, cps.pip, newLevels.filtered FROM consecutivecandles_results ccr INNER JOIN brokersymbollist bsl ON bsl.brokerid = $1 AND bsl.symbolid = ccr.symbolid INNER JOIN symbols s ON ccr.symbolid = s.symbolid AND s.nonliquid = 0 INNER JOIN downloadersymbolsettings dss ON ccr.symbolid = dss.symbolid INNER JOIN datafeedstimetable dtt ON dss.classname = dtt.classname AND dtt.dayofweek = 3 INNER JOIN symbolgroup sg on ccr.symbolid = sg.symbolid INNER JOIN groups g ON sg.groupid = g.groupid INNER JOIN brokergroups bg on g.groupid = bg.groupid AND bsl.brokerid = bg.brokerid INNER JOIN rar_max rm ON 1 = 1 LEFT OUTER JOIN relevance_consecutivecandles_results rcr ON rcr.resultuid = ccr.resultuid LEFT JOIN currencypips cps on cps.symbol = s.symbol LEFT OUTER JOIN brokerinstrumentmap bim ON dss.datafeedinstrumentid = bim.datafeedinstrumentid AND bim.brokerid = bsl.brokerid AND bim.TYPE = 'OUTBOUND' LEFT JOIN LATERAL calc_cc_signal_filter (ccr.resultuid) newLevels on true WHERE ccr.gmttimefound > now() - INTERVAL '7 DAYS' AND s.deleted = 0 AND (ccr.simulation = 0 OR ccr.simulation IS NULL) AND ($2 = 0 OR s.timegranularity in ($3, $4, $5, $6, $7, $8, $9)) AND ($10 = 0 OR s.exchange in ($11)) AND ($12 = 0 OR coalesce(bim.code, s.symbol) in ($13, $14, $15, $16, $17, $18, $19, $20, $21, $22, $23, $24, $25, $26, $27, $28, $29, $30, $31, $32, $33, $34, $35, $36, $37, $38, $39, $40, $41, $42, $43, $44, $45, $46, $47, $48, $49, $50, $51, $52, $53, $54, $55, $56, $57, $58, $59, $60, $61, $62, $63, $64, $65, $66, $67, $68, $69, $70, $71, $72, $73, $74, $75, $76, $77, $78, $79, $80, $81, $82, $83, $84, $85, $86, $87, $88, $89, $90, $91, $92, $93, $94, $95, $96, $97, $98, $99, $100, $101, $102, $103, $104, $105, $106, $107, $108, $109, $110, $111, $112, $113, $114, $115, $116, $117, $118, $119, $120, $121, $122, $123, $124, $125, $126, $127, $128, $129, $130, $131, $132, $133, $134, $135, $136, $137, $138, $139, $140, $141, $142, $143, $144, $145, $146, $147, $148, $149, $150, $151, $152, $153, $154, $155, $156, $157, $158, $159, $160, $161, $162, $163, $164, $165, $166, $167, $168, $169, $170, $171, $172, $173, $174, $175, $176, $177, $178, $179, $180, $181, $182, $183, $184, $185, $186, $187, $188, $189, $190, $191, $192, $193, $194, $195, $196, $197, $198, $199, $200, $201, $202, $203, $204, $205, $206, $207, $208, $209, $210, $211, $212, $213, $214, $215, $216, $217, $218, $219, $220, $221, $222, $223, $224, $225)) AND ($226 = 0 OR ccr.patternlengthbars <= $227)), results AS ( SELECT DISTINCT ON (symbolid) * FROM all_results WHERE (FALSE = $228 OR relevant = 1) AND ($229 = 0 OR age <= $230) ORDER BY symbolid, resultuid ) SELECT * from results ORDER BY identified DESC, length DESC;
Date: 2026-01-21 06:00:00 Duration: 0ms
2 47 289.20 MiB 3.24 MiB 8.97 MiB 6.15 MiB jr.resultuid as resultuid, jr.direction as direction, jr.patternendtime as identified, jr.patternlengthbars as length, jr.patternstarttime as patternstarttime, case when jr.trendchangeid = ? then ? else ? end as trendchange, s.exchange as exchange, s.symbolid as symbolid, coalesce(bim.code, s.symbol) as symbol_code, s.longname as symbol_name, s.timegranularity as interval, jp.patternname as pattern_name, dtt.timezone as timezone, ? as age, cps.pip, g.basegroupname from japsticks_results jr inner join brokersymbollist bsl on bsl.brokerid = ? and bsl.symbolid = jr.symbolid inner join relevance_japsticks_results rar on rar.resultuid = jr.resultuid inner join symbols s on jr.symbolid = s.symbolid and s.nonliquid = ? inner join japsticks_patterns jp on jr.patternid = jp.id inner join downloadersymbolsettings dss on jr.symbolid = dss.symbolid inner join datafeedstimetable dtt on dss.classname = dtt.classname and dtt.dayofweek = ? inner join symbolgroup sg on s.symbolid = sg.symbolid inner join groups g on sg.groupid = g.groupid inner join brokergroups bg on g.groupid = bg.groupid and bsl.brokerid = bg.brokerid left join currencypips cps on cps.symbol = s.symbol left outer join brokerinstrumentmap bim on dss.datafeedinstrumentid = bim.datafeedinstrumentid and bim.brokerid = bsl.brokerid and bim.type = ? where jr.gmttimefound > now() - interval ? and s.deleted = ? and (jr.simulation = ? or jr.simulation is null) and (rar.relevant = ?) --and (semicolon_age = ? or rar.age <= semicolon_age) and (? = ? or s.timegranularity in (...)) and (? = ? or s.exchange in (...)) and (? = ? or coalesce(bim.code, s.symbol) in (...)) and (? = ? or jp.patternname in (...)) and (? = ? or jr.patternlengthbars <= ?) ), results as ( select distinct on (symbolid) * from all_results order by symbolid, resultuid ) select * from results order by identified desc, length desc ;-
jr.resultuid AS resultuid, jr.direction AS direction, jr.patternendtime AS identified, jr.patternlengthbars AS length, jr.patternstarttime AS patternstarttime, case when jr.trendchangeid = 1 then 'Continuation' else 'Reversal' end AS trendchange, s.exchange AS exchange, s.symbolid AS symbolid, coalesce(bim.code, s.symbol) AS symbol_code, s.longname AS symbol_name, s.timegranularity AS interval, jp.patternname AS pattern_name, dtt.timezone AS timezone, 0 AS age, cps.pip, g.basegroupname FROM japsticks_results jr INNER JOIN brokersymbollist bsl ON bsl.brokerid = $1 AND bsl.symbolid = jr.symbolid INNER JOIN relevance_japsticks_results rar ON rar.resultuid = jr.resultuid INNER JOIN symbols s ON jr.symbolid = s.symbolid AND s.nonliquid = 0 INNER JOIN japsticks_patterns jp ON jr.patternid = jp.id INNER JOIN downloadersymbolsettings dss ON jr.symbolid = dss.symbolid INNER JOIN datafeedstimetable dtt ON dss.classname = dtt.classname AND dtt.dayofweek = 3 INNER JOIN symbolgroup sg on s.symbolid = sg.symbolid INNER JOIN groups g ON sg.groupid = g.groupid INNER JOIN brokergroups bg on g.groupid = bg.groupid AND bsl.brokerid = bg.brokerid LEFT JOIN currencypips cps on cps.symbol = s.symbol LEFT OUTER JOIN brokerinstrumentmap bim ON dss.datafeedinstrumentid = bim.datafeedinstrumentid AND bim.brokerid = bsl.brokerid AND bim.TYPE = 'OUTBOUND' WHERE jr.gmttimefound > now() - INTERVAL '7 DAYS' AND s.deleted = 0 AND (jr.simulation = 0 OR jr.simulation IS NULL) AND (rar.relevant = 1) --AND (semicolon_age = 0 OR rar.age <= semicolon_age) AND ($2 = 0 OR s.timegranularity in ($3)) AND ($4 = 0 OR s.exchange in ($5)) AND ($6 = 0 OR coalesce(bim.code, s.symbol) in ($7)) AND ($8 = 0 OR jp.patternname in ($9)) AND ($10 = 0 OR jr.patternlengthbars <= $11)), results AS ( SELECT DISTINCT ON (symbolid) * FROM all_results ORDER BY symbolid, resultuid ) SELECT * from results ORDER BY identified DESC, length DESC;
Date: 2026-01-21 06:01:53 Duration: 0ms
3 29 1.64 GiB 2.97 MiB 167.60 MiB 57.99 MiB with rankedmt4 as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors ), last_feed_entry as ( select * from rankedmt4 where r = ? ), ok_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where status = ? ), earliest_entry_after_ok as ( select m.datafeedname, min(m.eventtimestamp) as eventtimestamp from mt4datafeederrors m left outer join ( select datafeedname, eventtimestamp from ok_entries where r = ?) oo on m.datafeedname = oo.datafeedname where m.eventtimestamp > coalesce(oo.eventtimestamp, ?::timestamp without time zone) group by m.datafeedname ), notified_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where notified is not null and notified <> ? ), broker as ( select *, row_number() over (partition by feedname order by brokerid) r from ( select distinct b.brokerid, b.name as brokername, dss.classname as feedname from downloadersymbolsettings dss inner join brokersymbollist bsl on dss.symbolid = bsl.symbolid inner join broker b on bsl.brokerid = b.brokerid where dss.enabled = ?) a ) select last.id, last.datafeedname, last.eventtimestamp, last.status, last.errordescription, last.serveraddress, last.username, note.notified, note.eventtimestamp, broker.brokername from last_feed_entry last inner join earliest_entry_after_ok after_ok on last.datafeedname = after_ok.datafeedname inner join broker on last.datafeedname = broker.feedname left outer join ok_entries ok on ok.datafeedname = last.datafeedname left outer join notified_entries note on note.datafeedname = last.datafeedname and note.r = ? where (ok.r is null or ok.r = ?) and last.datafeedname not in ( select distinct datafeedname from last_feed_entry where status = ?) and extract(epoch from (last.eventtimestamp - after_ok.eventtimestamp)) > ? * ? and last.eventtimestamp > current_timestamp - interval ? and (note.eventtimestamp is null or note.eventtimestamp < current_timestamp - interval ?) and last.eventtimestamp > current_timestamp - interval ? and broker.r = ?;-
with rankedmt4 as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors ), last_feed_entry as ( select * from rankedmt4 where r = 1 ), ok_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where status = 'OK' ), earliest_entry_after_ok as ( select m.datafeedname, min(m.eventtimestamp) as eventtimestamp from mt4datafeederrors m left outer join ( select datafeedname, eventtimestamp from ok_entries where r = 1) oo on m.datafeedname = oo.datafeedname where m.eventtimestamp > coalesce(oo.eventtimestamp, '1900-01-01'::timestamp without time zone) group by m.datafeedname ), notified_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where notified is not null and notified <> '' ), broker as ( select *, row_number() over (partition by feedname order by brokerid) r from ( select distinct b.brokerid, b.name as brokername, dss.classname as feedname from downloadersymbolsettings dss inner join brokersymbollist bsl on dss.symbolid = bsl.symbolid inner join broker b on bsl.brokerid = b.brokerid where dss.enabled = 1) a ) select last.id, last.datafeedname, last.eventtimestamp, last.status, last.errordescription, last.serveraddress, last.username, note.notified, note.eventtimestamp, broker.brokername from last_feed_entry last inner join earliest_entry_after_ok after_ok on last.datafeedname = after_ok.datafeedname inner join broker on last.datafeedname = broker.feedname left outer join ok_entries ok on ok.datafeedname = last.datafeedname left outer join notified_entries note on note.datafeedname = last.datafeedname and note.r = 1 where (ok.r is null or ok.r = 1) and last.datafeedname not in ( select distinct datafeedname from last_feed_entry where status = 'OK') and extract(epoch from (last.eventtimestamp - after_ok.eventtimestamp)) > 60 * 60 and last.eventtimestamp > current_timestamp - interval '1 day' and (note.eventtimestamp is null or note.eventtimestamp < current_timestamp - interval '10 hours') and last.eventtimestamp > current_timestamp - interval '1 hour' and broker.r = 1;
Date: 2026-01-21 06:00:07 Duration: 0ms
4 13 598.90 MiB 46.06 MiB 46.07 MiB 46.07 MiB update solr_relevance_old set new_hod_correct = sub.hod_correct, new_hod_percent = sub.hod_percent, new_hod_total = sub.hod_total, new_pattern_correct = sub.pattern_correct, new_pattern_percent = sub.pattern_percent, new_pattern_total = sub.pattern_total, new_percent = sub.percent, new_symbol_correct = sub.symbol_correct, new_symbol_percent = sub.symbol_percent, new_symbol_total = sub.symbol_total from ( select distinct resultuid, hod_correct, hod_percent, hod_total, hod, pattern_correct, pattern_percent, pattern_total, percent, symbol_correct, symbol_percent, symbol_total from whatshot_probability where type = ?) sub where result_uid = sub.resultuid;-
UPDATE solr_relevance_old SET new_hod_correct = sub.hod_correct, new_hod_percent = sub.hod_percent, new_hod_total = sub.hod_total, new_pattern_correct = sub.pattern_correct, new_pattern_percent = sub.pattern_percent, new_pattern_total = sub.pattern_total, new_percent = sub.percent, new_symbol_correct = sub.symbol_correct, new_symbol_percent = sub.symbol_percent, new_symbol_total = sub.symbol_total FROM ( select distinct resultuid, hod_correct, hod_percent, hod_total, hod, pattern_correct, pattern_percent, pattern_total, percent, symbol_correct, symbol_percent, symbol_total FROM whatshot_probability WHERE type = 'cp') sub WHERE result_uid = sub.resultuid;
Date: 2026-01-21 06:01:13 Duration: 0ms
5 13 1016.23 MiB 78.17 MiB 78.17 MiB 78.17 MiB with max_ra as ( select resultuid from relevance_keylevels_results order by resultuid desc limit ?) update solr_relevance_old set newrelevant = sub.relevant, newage = sub.age from ( select so.uuid, case when ra.relevant is not null then ra.relevant when so.result_uid < max_ra.resultuid then ? else ? end as relevant, case when ra.age is not null then ra.age when so.result_uid < max_ra.resultuid then ? else ? end as age, so.result_uid from max_ra, solr_relevance_old so inner join keylevels_results k on so.result_uid = k.resultuid and so.uuid ilike ? inner join downloadersymbolsettings dss on k.symbolid = dss.symbolid left outer join relevance_keylevels_results ra on so.result_uid = ra.resultuid and so.uuid ilike ?) sub where solr_relevance_old.result_uid = sub.result_uid and solr_relevance_old.uuid ilike ?; update solr_relevance_old set newrelevant = ? where result_uid in ( select result_uid from solr_relevance_old s left outer join keylevels_results a on a.resultuid = s.result_uid where s.uuid ilike ? and a.resultuid is null); update solr_relevance_old set new_hod_correct = sub.hod_correct, new_hod_percent = sub.hod_percent, new_hod_total = sub.hod_total, new_pattern_correct = sub.pattern_correct, new_pattern_percent = sub.pattern_percent, new_pattern_total = sub.pattern_total, new_percent = sub.percent, new_symbol_correct = sub.symbol_correct, new_symbol_percent = sub.symbol_percent, new_symbol_total = sub.symbol_total from ( select distinct resultuid, hod_correct, hod_percent, hod_total, hod, pattern_correct, pattern_percent, pattern_total, percent, symbol_correct, symbol_percent, symbol_total from whatshot_probability where type in (...)) sub where result_uid = sub.resultuid;-
with max_ra as ( select resultuid from relevance_keylevels_results order by resultuid desc limit 1) update solr_relevance_old set newrelevant = sub.relevant, newage = sub.age from ( select so.uuid, case when ra.relevant is not null then ra.relevant when so.result_uid < max_ra.resultuid then 0 else 1 end as relevant, case when ra.age is not null then ra.age when so.result_uid < max_ra.resultuid then 11 else 0 end as age, so.result_uid from max_ra, solr_relevance_old so inner join keylevels_results k on so.result_uid = k.resultuid and so.uuid ilike 'kl_%' inner join downloadersymbolsettings dss on k.symbolid = dss.symbolid left outer join relevance_keylevels_results ra on so.result_uid = ra.resultuid and so.uuid ilike 'kl_%') sub where solr_relevance_old.result_uid = sub.result_uid and solr_relevance_old.uuid ilike 'kl_%'; update solr_relevance_old set newrelevant = 0 where result_uid in ( select result_uid from solr_relevance_old s left outer join keylevels_results a on a.resultuid = s.result_uid where s.uuid ilike 'kl_%' and a.resultuid is null); UPDATE solr_relevance_old SET new_hod_correct = sub.hod_correct, new_hod_percent = sub.hod_percent, new_hod_total = sub.hod_total, new_pattern_correct = sub.pattern_correct, new_pattern_percent = sub.pattern_percent, new_pattern_total = sub.pattern_total, new_percent = sub.percent, new_symbol_correct = sub.symbol_correct, new_symbol_percent = sub.symbol_percent, new_symbol_total = sub.symbol_total FROM ( select distinct resultuid, hod_correct, hod_percent, hod_total, hod, pattern_correct, pattern_percent, pattern_total, percent, symbol_correct, symbol_percent, symbol_total FROM whatshot_probability WHERE type in ('kl', 'ekl')) sub WHERE result_uid = sub.resultuid;
Date: 2026-01-21 06:01:16 Duration: 0ms
6 11 36.58 MiB 3.32 MiB 3.33 MiB 3.33 MiB select resultuid from relevance_autochartist_results order by resultuid desc limit ?), ar as ( select a.*, rr.age, rr.relevant from autochartist_results a left outer join relevance_autochartist_results rr on a.resultuid = rr.resultuid where case when false = ? then true else a.resultuid > ( select min(resultuid) from relevance_autochartist_results) end), all_results as ( select ar.resultuid as resultuid, ar.direction as direction, ar.predictiontimeto as predictiontimeto, ar.predictionpricefrom as predictionpricefrom, ar.predictionpriceto as predictionpriceto, cp.pip as pip, s.exchange as exchange, s.symbolid as symbolid, coalesce(bim.code, s.symbol) as symbol_code, s.longname as symbol_name, s.timegranularity as interval, ar.pattern as pattern_name, ar.breakout as breakout, ar.patternendtime as identified, dtt.timezone as timezone, ar.patternlengthbars as length, g.basegroupname, newlevels.profit, newlevels.stop, newlevels.filtered, case when ar.age is not null then ar.age when ar.resultuid <= rm.resultuid then ? else ? end as age, case when ar.relevant is not null then ar.relevant when ar.resultuid <= rm.resultuid then ? else ? end as relevant from ar inner join symbols s on ar.symbolid = s.symbolid and s.nonliquid = ? inner join brokersymbollist bsl on bsl.brokerid = ? and bsl.symbolid = s.symbolid inner join symbolgroup sg on bsl.symbolid = sg.symbolid inner join groups g on sg.groupid = g.groupid inner join brokergroups bg on g.groupid = bg.groupid and bsl.brokerid = bg.brokerid inner join downloadersymbolsettings dss on sg.symbolid = dss.symbolid inner join datafeedstimetable dtt on dss.classname = dtt.classname and dtt.dayofweek = ? inner join rar_max rm on ? = ? left outer join autochartist_symbolupdates au on dss.symbolid = au.symbolid left outer join currencypips cp on s.symbol = cp.symbol left join lateral calc_cp_signal (ar.resultuid) newlevels on true left outer join brokerinstrumentmap bim on dss.datafeedinstrumentid = bim.datafeedinstrumentid and bim.brokerid = bsl.brokerid and bim.type = ? where ar.gmttimefound > now() - interval ? and dss.enabled = ? and s.deleted = ? and (ar.simulation = ? or ar.simulation is null) and (? = ? or s.timegranularity in (...)) and (? = ? or s.exchange in (...)) and (? = ? or coalesce(bim.code, s.symbol) in (...)) and (? = ? or ar.pattern in (...)) and (? = ? or (? = ? and ar.breakout >= ?) or (? = ? and ar.breakout < ?)) and (? = ? or ar.patternlengthbars <= ?) and newlevels.filtered = false and ar.patternstarttime >= coalesce(au.earliestpricedatetime, ?::timestamp without time zone) -- to make sure patternstarttime is in our t-tables ), results as ( select distinct on (symbolid) * from all_results where (false = ? or relevant = ?) and (? = ? or age <= ?) order by symbolid, resultuid ) select * from results order by identified desc, length desc ;-
SELECT resultuid FROM relevance_autochartist_results ORDER BY resultuid DESC LIMIT 1), ar AS ( SELECT a.*, rr.age, rr.relevant from autochartist_results a LEFT OUTER JOIN relevance_autochartist_results rr on a.resultuid = rr.resultuid WHERE CASE WHEN FALSE = $1 THEN true ELSE a.resultuid > ( select min(resultuid) from relevance_autochartist_results) END), all_results AS ( SELECT ar.resultuid AS resultuid, ar.direction AS direction, ar.predictiontimeto AS predictiontimeto, ar.predictionpricefrom AS predictionpricefrom, ar.predictionpriceto AS predictionpriceto, cp.pip AS pip, s.exchange AS exchange, s.symbolid AS symbolid, coalesce(bim.code, s.symbol) AS symbol_code, s.longname AS symbol_name, s.timegranularity AS interval, ar.pattern AS pattern_name, ar.breakout AS breakout, ar.patternendtime AS identified, dtt.timezone AS timezone, ar.patternlengthbars AS length, g.basegroupname, newLevels.profit, newLevels.stop, newLevels.filtered, CASE WHEN ar.age IS NOT NULL THEN ar.age WHEN ar.resultuid <= rm.resultuid THEN 11 ELSE 0 END as age, CASE WHEN ar.relevant IS NOT NULL THEN ar.relevant WHEN ar.resultuid <= rm.resultuid THEN 0 ELSE 1 END as relevant FROM ar INNER JOIN symbols s ON ar.symbolid = s.symbolid AND s.nonliquid = 0 INNER JOIN brokersymbollist bsl ON bsl.brokerid = $2 AND bsl.symbolid = s.symbolid INNER JOIN symbolgroup sg on bsl.symbolid = sg.symbolid INNER JOIN groups g ON sg.groupid = g.groupid INNER JOIN brokergroups bg on g.groupid = bg.groupid AND bsl.brokerid = bg.brokerid INNER JOIN downloadersymbolsettings dss ON sg.symbolid = dss.symbolid INNER JOIN datafeedstimetable dtt ON dss.classname = dtt.classname AND dtt.dayofweek = 3 INNER JOIN rar_max rm ON 1 = 1 LEFT OUTER JOIN autochartist_symbolupdates au on dss.symbolid = au.symbolid LEFT OUTER JOIN currencypips cp ON s.symbol = cp.symbol LEFT JOIN LATERAL calc_cp_signal (ar.resultuid) newLevels on true LEFT OUTER JOIN brokerinstrumentmap bim ON dss.datafeedinstrumentid = bim.datafeedinstrumentid AND bim.brokerid = bsl.brokerid AND bim.TYPE = 'OUTBOUND' WHERE ar.gmttimefound > now() - INTERVAL '7 DAYS' AND dss.enabled = 1 AND s.deleted = 0 AND (ar.simulation = 0 OR ar.simulation IS NULL) AND ($3 = 0 OR s.timegranularity in ($4)) AND ($5 = 0 OR s.exchange in ($6)) AND ($7 = 0 OR coalesce(bim.code, s.symbol) in ($8)) AND ($9 = 0 OR ar.pattern in ($10)) AND ($11 = 0 OR ($12 = 1 AND ar.breakout >= 0) OR ($13 = 2 AND ar.breakout < 0)) AND ($14 = 0 OR ar.patternlengthbars <= $15) and newLevels.filtered = false AND ar.patternstarttime >= coalesce(au.earliestpricedatetime, '1900-01-01'::timestamp without time zone) -- To make sure patternstarttime is in our t-tables ), results AS ( SELECT DISTINCT ON (symbolid) * FROM all_results WHERE (FALSE = $16 OR relevant = 1) AND ($17 = 0 OR age <= $18) ORDER BY symbolid, resultuid ) SELECT * from results ORDER BY identified DESC, length DESC;
Date: 2026-01-21 06:02:43 Duration: 0ms
7 8 1010.43 MiB 126.28 MiB 126.33 MiB 126.30 MiB select updateresultsmaterializedview ();-
select updateresultsmaterializedview ();
Date: 2026-01-21 06:02:21 Duration: 0ms
8 4 345.99 MiB 86.44 MiB 86.56 MiB 86.50 MiB select updateageforrelevantresults ();-
select updateageforrelevantresults ();
Date: 2026-01-21 06:02:07 Duration: 0ms
Queries generating the largest temporary files
Rank Size Query 1 167.60 MiB with rankedmt4 as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors ), last_feed_entry as ( select * from rankedmt4 where r = 1 ), ok_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where status = 'OK' ), earliest_entry_after_ok as ( select m.datafeedname, min(m.eventtimestamp) as eventtimestamp from mt4datafeederrors m left outer join ( select datafeedname, eventtimestamp from ok_entries where r = 1) oo on m.datafeedname = oo.datafeedname where m.eventtimestamp > coalesce(oo.eventtimestamp, '1900-01-01'::timestamp without time zone) group by m.datafeedname ), notified_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where notified is not null and notified <> '' ), broker as ( select *, row_number() over (partition by feedname order by brokerid) r from ( select distinct b.brokerid, b.name as brokername, dss.classname as feedname from downloadersymbolsettings dss inner join brokersymbollist bsl on dss.symbolid = bsl.symbolid inner join broker b on bsl.brokerid = b.brokerid where dss.enabled = 1) a ) select last.id, last.datafeedname, last.eventtimestamp, last.status, last.errordescription, last.serveraddress, last.username, note.notified, note.eventtimestamp, broker.brokername from last_feed_entry last inner join earliest_entry_after_ok after_ok on last.datafeedname = after_ok.datafeedname inner join broker on last.datafeedname = broker.feedname left outer join ok_entries ok on ok.datafeedname = last.datafeedname left outer join notified_entries note on note.datafeedname = last.datafeedname and note.r = 1 where (ok.r is null or ok.r = 1) and last.datafeedname not in ( select distinct datafeedname from last_feed_entry where status = 'OK') and extract(epoch from (last.eventtimestamp - after_ok.eventtimestamp)) > 60 * 60 and last.eventtimestamp > current_timestamp - interval '1 day' and (note.eventtimestamp is null or note.eventtimestamp < current_timestamp - interval '10 hours') and last.eventtimestamp > current_timestamp - interval '1 hour' and broker.r = 1;[ Date: 2026-01-21 06:10:06 ]
2 140.61 MiB with rankedmt4 as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors ), last_feed_entry as ( select * from rankedmt4 where r = 1 ), ok_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where status = 'OK' ), earliest_entry_after_ok as ( select m.datafeedname, min(m.eventtimestamp) as eventtimestamp from mt4datafeederrors m left outer join ( select datafeedname, eventtimestamp from ok_entries where r = 1) oo on m.datafeedname = oo.datafeedname where m.eventtimestamp > coalesce(oo.eventtimestamp, '1900-01-01'::timestamp without time zone) group by m.datafeedname ), notified_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where notified is not null and notified <> '' ), broker as ( select *, row_number() over (partition by feedname order by brokerid) r from ( select distinct b.brokerid, b.name as brokername, dss.classname as feedname from downloadersymbolsettings dss inner join brokersymbollist bsl on dss.symbolid = bsl.symbolid inner join broker b on bsl.brokerid = b.brokerid where dss.enabled = 1) a ) select last.id, last.datafeedname, last.eventtimestamp, last.status, last.errordescription, last.serveraddress, last.username, note.notified, note.eventtimestamp, broker.brokername from last_feed_entry last inner join earliest_entry_after_ok after_ok on last.datafeedname = after_ok.datafeedname inner join broker on last.datafeedname = broker.feedname left outer join ok_entries ok on ok.datafeedname = last.datafeedname left outer join notified_entries note on note.datafeedname = last.datafeedname and note.r = 1 where (ok.r is null or ok.r = 1) and last.datafeedname not in ( select distinct datafeedname from last_feed_entry where status = 'OK') and extract(epoch from (last.eventtimestamp - after_ok.eventtimestamp)) > 60 * 60 and last.eventtimestamp > current_timestamp - interval '1 day' and (note.eventtimestamp is null or note.eventtimestamp < current_timestamp - interval '10 hours') and last.eventtimestamp > current_timestamp - interval '1 hour' and broker.r = 1;[ Date: 2026-01-21 06:40:25 ]
3 126.33 MiB select updateresultsmaterializedview ();[ Date: 2026-01-21 06:47:34 ]
4 126.32 MiB select updateresultsmaterializedview ();[ Date: 2026-01-21 06:02:21 ]
5 126.31 MiB select updateresultsmaterializedview ();[ Date: 2026-01-21 06:32:40 ]
6 126.30 MiB select updateresultsmaterializedview ();[ Date: 2026-01-21 06:17:36 ]
7 126.29 MiB select updateresultsmaterializedview ();[ Date: 2026-01-21 06:20:39 ]
8 126.29 MiB select updateresultsmaterializedview ();[ Date: 2026-01-21 06:50:36 ]
9 126.29 MiB select updateresultsmaterializedview ();[ Date: 2026-01-21 06:05:32 ]
10 126.28 MiB select updateresultsmaterializedview ();[ Date: 2026-01-21 06:35:41 ]
11 110.20 MiB with rankedmt4 as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors ), last_feed_entry as ( select * from rankedmt4 where r = 1 ), ok_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where status = 'OK' ), earliest_entry_after_ok as ( select m.datafeedname, min(m.eventtimestamp) as eventtimestamp from mt4datafeederrors m left outer join ( select datafeedname, eventtimestamp from ok_entries where r = 1) oo on m.datafeedname = oo.datafeedname where m.eventtimestamp > coalesce(oo.eventtimestamp, '1900-01-01'::timestamp without time zone) group by m.datafeedname ), notified_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where notified is not null and notified <> '' ), broker as ( select *, row_number() over (partition by feedname order by brokerid) r from ( select distinct b.brokerid, b.name as brokername, dss.classname as feedname from downloadersymbolsettings dss inner join brokersymbollist bsl on dss.symbolid = bsl.symbolid inner join broker b on bsl.brokerid = b.brokerid where dss.enabled = 1) a ) select last.id, last.datafeedname, last.eventtimestamp, last.status, last.errordescription, last.serveraddress, last.username, note.notified, note.eventtimestamp, broker.brokername from last_feed_entry last inner join earliest_entry_after_ok after_ok on last.datafeedname = after_ok.datafeedname inner join broker on last.datafeedname = broker.feedname left outer join ok_entries ok on ok.datafeedname = last.datafeedname left outer join notified_entries note on note.datafeedname = last.datafeedname and note.r = 1 where (ok.r is null or ok.r = 1) and last.datafeedname not in ( select distinct datafeedname from last_feed_entry where status = 'OK') and extract(epoch from (last.eventtimestamp - after_ok.eventtimestamp)) > 60 * 60 and last.eventtimestamp > current_timestamp - interval '1 day' and (note.eventtimestamp is null or note.eventtimestamp < current_timestamp - interval '10 hours') and last.eventtimestamp > current_timestamp - interval '1 hour' and broker.r = 1;[ Date: 2026-01-21 06:00:05 ]
12 101.27 MiB with rankedmt4 as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors ), last_feed_entry as ( select * from rankedmt4 where r = 1 ), ok_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where status = 'OK' ), earliest_entry_after_ok as ( select m.datafeedname, min(m.eventtimestamp) as eventtimestamp from mt4datafeederrors m left outer join ( select datafeedname, eventtimestamp from ok_entries where r = 1) oo on m.datafeedname = oo.datafeedname where m.eventtimestamp > coalesce(oo.eventtimestamp, '1900-01-01'::timestamp without time zone) group by m.datafeedname ), notified_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where notified is not null and notified <> '' ), broker as ( select *, row_number() over (partition by feedname order by brokerid) r from ( select distinct b.brokerid, b.name as brokername, dss.classname as feedname from downloadersymbolsettings dss inner join brokersymbollist bsl on dss.symbolid = bsl.symbolid inner join broker b on bsl.brokerid = b.brokerid where dss.enabled = 1) a ) select last.id, last.datafeedname, last.eventtimestamp, last.status, last.errordescription, last.serveraddress, last.username, note.notified, note.eventtimestamp, broker.brokername from last_feed_entry last inner join earliest_entry_after_ok after_ok on last.datafeedname = after_ok.datafeedname inner join broker on last.datafeedname = broker.feedname left outer join ok_entries ok on ok.datafeedname = last.datafeedname left outer join notified_entries note on note.datafeedname = last.datafeedname and note.r = 1 where (ok.r is null or ok.r = 1) and last.datafeedname not in ( select distinct datafeedname from last_feed_entry where status = 'OK') and extract(epoch from (last.eventtimestamp - after_ok.eventtimestamp)) > 60 * 60 and last.eventtimestamp > current_timestamp - interval '1 day' and (note.eventtimestamp is null or note.eventtimestamp < current_timestamp - interval '10 hours') and last.eventtimestamp > current_timestamp - interval '1 hour' and broker.r = 1;[ Date: 2026-01-21 06:20:09 ]
13 93.64 MiB with rankedmt4 as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors ), last_feed_entry as ( select * from rankedmt4 where r = 1 ), ok_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where status = 'OK' ), earliest_entry_after_ok as ( select m.datafeedname, min(m.eventtimestamp) as eventtimestamp from mt4datafeederrors m left outer join ( select datafeedname, eventtimestamp from ok_entries where r = 1) oo on m.datafeedname = oo.datafeedname where m.eventtimestamp > coalesce(oo.eventtimestamp, '1900-01-01'::timestamp without time zone) group by m.datafeedname ), notified_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where notified is not null and notified <> '' ), broker as ( select *, row_number() over (partition by feedname order by brokerid) r from ( select distinct b.brokerid, b.name as brokername, dss.classname as feedname from downloadersymbolsettings dss inner join brokersymbollist bsl on dss.symbolid = bsl.symbolid inner join broker b on bsl.brokerid = b.brokerid where dss.enabled = 1) a ) select last.id, last.datafeedname, last.eventtimestamp, last.status, last.errordescription, last.serveraddress, last.username, note.notified, note.eventtimestamp, broker.brokername from last_feed_entry last inner join earliest_entry_after_ok after_ok on last.datafeedname = after_ok.datafeedname inner join broker on last.datafeedname = broker.feedname left outer join ok_entries ok on ok.datafeedname = last.datafeedname left outer join notified_entries note on note.datafeedname = last.datafeedname and note.r = 1 where (ok.r is null or ok.r = 1) and last.datafeedname not in ( select distinct datafeedname from last_feed_entry where status = 'OK') and extract(epoch from (last.eventtimestamp - after_ok.eventtimestamp)) > 60 * 60 and last.eventtimestamp > current_timestamp - interval '1 day' and (note.eventtimestamp is null or note.eventtimestamp < current_timestamp - interval '10 hours') and last.eventtimestamp > current_timestamp - interval '1 hour' and broker.r = 1;[ Date: 2026-01-21 06:50:07 ]
14 92.62 MiB with rankedmt4 as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors ), last_feed_entry as ( select * from rankedmt4 where r = 1 ), ok_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where status = 'OK' ), earliest_entry_after_ok as ( select m.datafeedname, min(m.eventtimestamp) as eventtimestamp from mt4datafeederrors m left outer join ( select datafeedname, eventtimestamp from ok_entries where r = 1) oo on m.datafeedname = oo.datafeedname where m.eventtimestamp > coalesce(oo.eventtimestamp, '1900-01-01'::timestamp without time zone) group by m.datafeedname ), notified_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where notified is not null and notified <> '' ), broker as ( select *, row_number() over (partition by feedname order by brokerid) r from ( select distinct b.brokerid, b.name as brokername, dss.classname as feedname from downloadersymbolsettings dss inner join brokersymbollist bsl on dss.symbolid = bsl.symbolid inner join broker b on bsl.brokerid = b.brokerid where dss.enabled = 1) a ) select last.id, last.datafeedname, last.eventtimestamp, last.status, last.errordescription, last.serveraddress, last.username, note.notified, note.eventtimestamp, broker.brokername from last_feed_entry last inner join earliest_entry_after_ok after_ok on last.datafeedname = after_ok.datafeedname inner join broker on last.datafeedname = broker.feedname left outer join ok_entries ok on ok.datafeedname = last.datafeedname left outer join notified_entries note on note.datafeedname = last.datafeedname and note.r = 1 where (ok.r is null or ok.r = 1) and last.datafeedname not in ( select distinct datafeedname from last_feed_entry where status = 'OK') and extract(epoch from (last.eventtimestamp - after_ok.eventtimestamp)) > 60 * 60 and last.eventtimestamp > current_timestamp - interval '1 day' and (note.eventtimestamp is null or note.eventtimestamp < current_timestamp - interval '10 hours') and last.eventtimestamp > current_timestamp - interval '1 hour' and broker.r = 1;[ Date: 2026-01-21 06:30:07 ]
15 88.02 MiB with rankedmt4 as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors ), last_feed_entry as ( select * from rankedmt4 where r = 1 ), ok_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where status = 'OK' ), earliest_entry_after_ok as ( select m.datafeedname, min(m.eventtimestamp) as eventtimestamp from mt4datafeederrors m left outer join ( select datafeedname, eventtimestamp from ok_entries where r = 1) oo on m.datafeedname = oo.datafeedname where m.eventtimestamp > coalesce(oo.eventtimestamp, '1900-01-01'::timestamp without time zone) group by m.datafeedname ), notified_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where notified is not null and notified <> '' ), broker as ( select *, row_number() over (partition by feedname order by brokerid) r from ( select distinct b.brokerid, b.name as brokername, dss.classname as feedname from downloadersymbolsettings dss inner join brokersymbollist bsl on dss.symbolid = bsl.symbolid inner join broker b on bsl.brokerid = b.brokerid where dss.enabled = 1) a ) select last.id, last.datafeedname, last.eventtimestamp, last.status, last.errordescription, last.serveraddress, last.username, note.notified, note.eventtimestamp, broker.brokername from last_feed_entry last inner join earliest_entry_after_ok after_ok on last.datafeedname = after_ok.datafeedname inner join broker on last.datafeedname = broker.feedname left outer join ok_entries ok on ok.datafeedname = last.datafeedname left outer join notified_entries note on note.datafeedname = last.datafeedname and note.r = 1 where (ok.r is null or ok.r = 1) and last.datafeedname not in ( select distinct datafeedname from last_feed_entry where status = 'OK') and extract(epoch from (last.eventtimestamp - after_ok.eventtimestamp)) > 60 * 60 and last.eventtimestamp > current_timestamp - interval '1 day' and (note.eventtimestamp is null or note.eventtimestamp < current_timestamp - interval '10 hours') and last.eventtimestamp > current_timestamp - interval '1 hour' and broker.r = 1;[ Date: 2026-01-21 06:00:07 ]
16 86.56 MiB select updateageforrelevantresults ();[ Date: 2026-01-21 06:02:07 ]
17 86.52 MiB select updateageforrelevantresults ();[ Date: 2026-01-21 06:32:21 ]
18 86.48 MiB select updateageforrelevantresults ();[ Date: 2026-01-21 06:47:20 ]
19 86.44 MiB select updateageforrelevantresults ();[ Date: 2026-01-21 06:17:21 ]
20 86.40 MiB with rankedmt4 as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors ), last_feed_entry as ( select * from rankedmt4 where r = 1 ), ok_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where status = 'OK' ), earliest_entry_after_ok as ( select m.datafeedname, min(m.eventtimestamp) as eventtimestamp from mt4datafeederrors m left outer join ( select datafeedname, eventtimestamp from ok_entries where r = 1) oo on m.datafeedname = oo.datafeedname where m.eventtimestamp > coalesce(oo.eventtimestamp, '1900-01-01'::timestamp without time zone) group by m.datafeedname ), notified_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where notified is not null and notified <> '' ), broker as ( select *, row_number() over (partition by feedname order by brokerid) r from ( select distinct b.brokerid, b.name as brokername, dss.classname as feedname from downloadersymbolsettings dss inner join brokersymbollist bsl on dss.symbolid = bsl.symbolid inner join broker b on bsl.brokerid = b.brokerid where dss.enabled = 1) a ) select last.id, last.datafeedname, last.eventtimestamp, last.status, last.errordescription, last.serveraddress, last.username, note.notified, note.eventtimestamp, broker.brokername from last_feed_entry last inner join earliest_entry_after_ok after_ok on last.datafeedname = after_ok.datafeedname inner join broker on last.datafeedname = broker.feedname left outer join ok_entries ok on ok.datafeedname = last.datafeedname left outer join notified_entries note on note.datafeedname = last.datafeedname and note.r = 1 where (ok.r is null or ok.r = 1) and last.datafeedname not in ( select distinct datafeedname from last_feed_entry where status = 'OK') and extract(epoch from (last.eventtimestamp - after_ok.eventtimestamp)) > 60 * 60 and last.eventtimestamp > current_timestamp - interval '1 day' and (note.eventtimestamp is null or note.eventtimestamp < current_timestamp - interval '10 hours') and last.eventtimestamp > current_timestamp - interval '1 hour' and broker.r = 1;[ Date: 2026-01-21 06:40:28 ]
-
Vacuums
Vacuums / Analyzes Distribution
Key values
- 0 sec Highest CPU-cost vacuum
Table
Database - Date
- 0 sec Highest CPU-cost analyze
Table
Database - Date
Analyzes per table
Key values
- public.solr_relevance_old (13) Main table analyzed (database acaweb_fx)
- 51 analyzes Total
Table Number of analyzes acaweb_fx.public.solr_relevance_old 13 acaweb_fx.pg_catalog.pg_attribute 5 acaweb_fx.public.datafeeds_latestrun 4 acaweb_fx.public.relevance_keylevels_results 4 acaweb_fx.pg_catalog.pg_class 4 acaweb_fx.public.relevance_autochartist_results 4 acaweb_fx.public.relevance_fibonacci_results 4 acaweb_fx.pg_catalog.pg_type 3 acaweb_fx.public.latest_t15_candle_view 3 acaweb_fx.pg_catalog.pg_index 2 acaweb_fx.public.autochartist_symbolupdates 1 acaweb_fx.public.solr_imports 1 acaweb_fx.public.symbollatestupdatetime 1 acaweb_fx.pg_catalog.pg_depend 1 acaweb_fx.public.latest_candle_datetime_per_receng 1 Total 51 Vacuums per table
Key values
- public.solr_relevance_old (13) Main table vacuumed on database acaweb_fx
- 36 vacuums Total
Index Buffer usage Skipped WAL usage Table Vacuums scans hits misses dirtied pins frozen records full page bytes acaweb_fx.public.solr_relevance_old 13 13 10,884 0 44 0 0 8,226 13 1,472,233 acaweb_fx.public.datafeeds_latestrun 5 0 590 0 14 0 0 48 11 64,813 acaweb_fx.pg_toast.pg_toast_2619 2 2 264 0 64 0 0 162 54 215,743 acaweb_fx.pg_catalog.pg_statistic 2 2 1,908 0 399 0 1,176 835 344 1,303,255 acaweb_fx.pg_catalog.pg_attribute 2 2 1,565 0 351 0 134 760 275 1,645,597 acaweb_fx.public.latest_t15_candle_view 2 2 186 0 8 0 0 12 2 17,143 acaweb_fx.public.relevance_keylevels_results 2 2 7,730 0 316 2 178 1,974 458 1,713,152 acaweb_fx.public.relevance_autochartist_results 2 2 6,806 0 175 0 500 1,464 161 464,557 acaweb_fx.public.relevance_fibonacci_results 2 2 2,497 0 49 0 110 448 35 106,756 acaweb_fx.pg_catalog.pg_type 1 1 133 0 21 0 0 53 15 108,220 acaweb_fx.public.autochartist_symbolupdates 1 1 24,128 0 747 2 38,306 6,133 602 584,475 acaweb_fx.public.solr_imports 1 1 43 0 1 0 0 6 1 7,710 acaweb_fx.pg_catalog.pg_class 1 1 470 0 47 0 0 164 41 238,975 Total 36 31 57,204 45,075 2,236 4 40,404 20,285 2,012 7,942,629 Tuples removed per table
Key values
- public.solr_relevance_old (66423) Main table with removed tuples on database acaweb_fx
- 83597 tuples Total removed
Index Tuples Pages Table Vacuums scans removed remain not yet removable removed remain acaweb_fx.public.solr_relevance_old 13 13 66,423 75,445 0 0 2,657 acaweb_fx.public.autochartist_symbolupdates 1 1 4,642 48,720 779 0 40,691 acaweb_fx.public.relevance_keylevels_results 2 2 3,773 23,966 0 0 558 acaweb_fx.pg_catalog.pg_attribute 2 2 3,404 21,697 137 17 505 acaweb_fx.public.relevance_autochartist_results 2 2 2,942 16,637 0 0 760 acaweb_fx.pg_catalog.pg_statistic 2 2 936 7,642 179 0 2,388 acaweb_fx.public.relevance_fibonacci_results 2 2 586 2,923 0 0 204 acaweb_fx.pg_catalog.pg_class 1 1 308 1,655 6 0 150 acaweb_fx.public.datafeeds_latestrun 5 0 211 159 89 0 80 acaweb_fx.pg_toast.pg_toast_2619 2 2 116 379 39 0 100 acaweb_fx.pg_catalog.pg_type 1 1 104 1,446 0 0 39 acaweb_fx.public.latest_t15_candle_view 2 2 102 52 24 0 2 acaweb_fx.public.solr_imports 1 1 50 3 2 0 2 Total 36 31 83,597 200,724 1,255 17 48,136 Pages removed per table
Key values
- pg_catalog.pg_attribute (17) Main table with removed pages on database acaweb_fx
- 17 pages Total removed
Table Number of vacuums Index scans Tuples removed Pages removed acaweb_fx.pg_catalog.pg_attribute 2 2 3404 17 acaweb_fx.pg_toast.pg_toast_2619 2 2 116 0 acaweb_fx.pg_catalog.pg_type 1 1 104 0 acaweb_fx.public.datafeeds_latestrun 5 0 211 0 acaweb_fx.public.autochartist_symbolupdates 1 1 4642 0 acaweb_fx.public.solr_imports 1 1 50 0 acaweb_fx.pg_catalog.pg_statistic 2 2 936 0 acaweb_fx.public.latest_t15_candle_view 2 2 102 0 acaweb_fx.public.relevance_keylevels_results 2 2 3773 0 acaweb_fx.pg_catalog.pg_class 1 1 308 0 acaweb_fx.public.solr_relevance_old 13 13 66423 0 acaweb_fx.public.relevance_autochartist_results 2 2 2942 0 acaweb_fx.public.relevance_fibonacci_results 2 2 586 0 Total 36 31 83,597 17 Autovacuum Activity
↑ Back to the top of the Autovacuum Activity tableDay Hour VACUUMs ANALYZEs Jan 21 06 36 51 - 0 sec Highest CPU-cost vacuum
-
Locks
Locks by types
Key values
- AccessShareLock Main Lock Type
- 5 locks Total
Most frequent waiting queries (N)
Rank Count Total time Min time Max time Avg duration Query 1 1 14m19s 14m19s 14m19s 14m19s truncate table solr_relevance_old;-
TRUNCATE TABLE solr_relevance_old;
Date: 2026-01-21 06:19:30
2 2 4m17s 1m51s 2m26s 2m8s select recognitionengine, to_char(datetimeupdate, ?) from latest_candle_datetime_per_receng;-
select recognitionengine, to_char(datetimeupdate, 'yyyy-mm-dd HH24:MI') from latest_candle_datetime_per_receng;
Date: 2026-01-21 06:19:31
3 1 2m29s 2m29s 2m29s 2m29s refresh materialized view latest_candle_datetime_per_receng;-
refresh materialized view latest_candle_datetime_per_receng;
Date: 2026-01-21 06:19:30
4 1 2m11s 2m11s 2m11s 2m11s select datetimeupdate from latest_candle_datetime_per_receng where recognitionengine ilike ?;-
SELECT datetimeupdate FROM latest_candle_datetime_per_receng WHERE recognitionengine ILIKE 'ICMARKETS-AU-MT5 - 1';
Date: 2026-01-21 06:19:31
Queries that waited the most
Rank Wait time Query 1 14m19s TRUNCATE TABLE solr_relevance_old;[ Date: 2026-01-21 06:19:30 ]
2 2m29s refresh materialized view latest_candle_datetime_per_receng;[ Date: 2026-01-21 06:19:30 ]
3 2m26s select recognitionengine, to_char(datetimeupdate, 'yyyy-mm-dd HH24:MI') from latest_candle_datetime_per_receng;[ Date: 2026-01-21 06:19:31 ]
4 2m11s SELECT datetimeupdate FROM latest_candle_datetime_per_receng WHERE recognitionengine ILIKE 'ICMARKETS-AU-MT5 - 1';[ Date: 2026-01-21 06:19:31 ]
5 1m51s select recognitionengine, to_char(datetimeupdate, 'yyyy-mm-dd HH24:MI') from latest_candle_datetime_per_receng;[ Date: 2026-01-21 06:19:31 ]
-
Queries
Queries by type
Key values
- 45,960 Total read queries
- 36,590 Total write queries
Queries by database
Key values
- unknown Main database
- 173,639 Requests
- 3h31m11s (unknown)
- Main time consuming database
Database Request type Count Duration acaweb_fx Total 878 0ms copy from 65 0ms copy to 26 0ms cte 85 0ms ddl 13 0ms delete 13 0ms others 203 0ms select 103 0ms tcl 331 0ms update 39 0ms acaweb_fx_integer Total 1 0ms select 1 0ms postgres Total 3 0ms select 3 0ms socialmedia Total 92 0ms select 91 0ms tcl 1 0ms translations Total 1 0ms select 1 0ms unknown Total 173,639 3h31m11s copy from 13 0ms copy to 575 0ms cte 4,677 0ms insert 28,604 0ms others 4,690 0ms select 45,761 0ms tcl 367 0ms update 1,781 0ms Queries by user
Key values
- unknown Main user
- 173,639 Requests
User Request type Count Duration postgres Total 975 0ms copy from 65 0ms copy to 26 0ms cte 85 0ms ddl 13 0ms delete 13 0ms others 203 0ms select 199 0ms tcl 332 0ms update 39 0ms unknown Total 173,639 3h31m11s copy from 13 0ms copy to 575 0ms cte 4,677 0ms insert 28,604 0ms others 4,690 0ms select 45,761 0ms tcl 367 0ms update 1,781 0ms Duration by user
Key values
- 3h31m11s (unknown) Main time consuming user
User Request type Count Duration postgres Total 975 0ms copy from 65 0ms copy to 26 0ms cte 85 0ms ddl 13 0ms delete 13 0ms others 203 0ms select 199 0ms tcl 332 0ms update 39 0ms unknown Total 173,639 3h31m11s copy from 13 0ms copy to 575 0ms cte 4,677 0ms insert 28,604 0ms others 4,690 0ms select 45,761 0ms tcl 367 0ms update 1,781 0ms Queries by host
Key values
- unknown Main host
- 174,614 Requests
- 3h31m11s (unknown)
- Main time consuming host
Queries by application
Key values
- unknown Main application
- 174,260 Requests
- 3h31m11s (unknown)
- Main time consuming application
Application Request type Count Duration pg_dump Total 5 0ms select 5 0ms psql Total 349 0ms copy from 65 0ms copy to 26 0ms cte 85 0ms ddl 13 0ms delete 13 0ms others 4 0ms select 104 0ms update 39 0ms unknown Total 174,260 3h31m11s copy from 13 0ms copy to 575 0ms cte 4,677 0ms insert 28,604 0ms others 4,889 0ms select 45,851 0ms tcl 699 0ms update 1,781 0ms Number of cancelled queries
Key values
- 0 per second Cancelled query Peak
- 2026-01-21 06:57:10 Date
Number of cancelled queries (5 minutes period)
NO DATASET
-
Top Queries
Histogram of query times
Key values
- 54,006 0-1ms duration
Slowest individual queries
Rank Duration Query NO DATASET
Time consuming queries
Rank Total duration Times executed Min duration Max duration Avg duration Query 1 0ms 1 0ms 0ms 0ms lock table public.stats_hrs_summary_by_groups in access share mode;Times Reported Time consuming queries #1
Day Hour Count Duration Avg duration Jan 21 06 1 0ms 0ms 2 0ms 2 0ms 0ms 0ms lock table public.timezones_korean in access share mode;Times Reported Time consuming queries #2
Day Hour Count Duration Avg duration Jan 21 06 2 0ms 0ms 3 0ms 32 0ms 0ms 0ms select key, value from datasources ds inner join datasourceparams dsp on ds.id = dsp.datasourceid where ds.name = ?;Times Reported Time consuming queries #3
Day Hour Count Duration Avg duration Jan 21 06 32 0ms 0ms 4 0ms 2 0ms 0ms 0ms copy public.t30 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived, pricesadjusted, relevanceprocessed, recengprocessedadjusted, historical) to stdout;Times Reported Time consuming queries #4
Day Hour Count Duration Avg duration Jan 21 06 2 0ms 0ms 5 0ms 1 0ms 0ms 0ms copy public.processresults (id, processid, resultdate, shorttext, longtext, title, test) to stdout;Times Reported Time consuming queries #5
Day Hour Count Duration Avg duration Jan 21 06 1 0ms 0ms 6 0ms 2,114 0ms 0ms 0ms insert into t60 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) values (?, ?, ?, ?, ?, ?, ?, ?, ?, ?) on conflict (pricedatetime, symbolid) do update set open = ?, high = ?, low = ?, close = ?, volume = ?, bsf = ?, sastdatetimewritten = ?, sastdatetimereceived = ?;Times Reported Time consuming queries #6
Day Hour Count Duration Avg duration Jan 21 06 2,114 0ms 0ms 7 0ms 975 0ms 0ms 0ms select oid, tableoid, pol.polname, pol.polcmd, pol.polpermissive, case when pol.polroles = ? then null else pg_catalog.array_to_string(array ( select pg_catalog.quote_ident(rolname) from pg_catalog.pg_roles where oid = any (pol.polroles)), ?) end as polroles, pg_catalog.pg_get_expr(pol.polqual, pol.polrelid) as polqual, pg_catalog.pg_get_expr(pol.polwithcheck, pol.polrelid) as polwithcheck from pg_catalog.pg_policy pol where polrelid = ?;Times Reported Time consuming queries #7
Day Hour Count Duration Avg duration Jan 21 06 975 0ms 0ms 8 0ms 2 0ms 0ms 0ms copy public.timezones_greek (timezoneid, timezone) to stdout;Times Reported Time consuming queries #8
Day Hour Count Duration Avg duration Jan 21 06 2 0ms 0ms 9 0ms 2 0ms 0ms 0ms lock table public.futures_symbols in access share mode;Times Reported Time consuming queries #9
Day Hour Count Duration Avg duration Jan 21 06 2 0ms 0ms 10 0ms 2 0ms 0ms 0ms copy archive.autochartist_results (resultid, bandwidth, pattern, gmttimefound, direction, initialtrend, breakout, volumeincrease, noise, symmetry, patternstarttime, patternendtime, patternstartprice, patternendprice, resx0, resx1, supportx0, supportx1, resy0, resy1, supporty0, supporty1, supportgradient, resgradient, riskreward, qtytp, predictionpricefrom, predictionpriceto, predictiontimefrom, predictiontimeto, patternquality, trendchange, maxmovementafterbreakout, latestbaratbreakouttime, latestbaratbreakoutprice, patternlengthbars, temporarypattern, symbolid, resultuid, relevancestartdistance, simulation, writtendatetime, old_resultuid) to stdout;Times Reported Time consuming queries #10
Day Hour Count Duration Avg duration Jan 21 06 2 0ms 0ms 11 0ms 2 0ms 0ms 0ms copy public.satelliteauthentication (id, satelliteserverid, brokerid, authenticationstatus, maketradeurl, notifications, validusername, validpassword, enabled) to stdout;Times Reported Time consuming queries #11
Day Hour Count Duration Avg duration Jan 21 06 2 0ms 0ms 12 0ms 18 0ms 0ms 0ms select cast(count(*) / cast(setting as numeric) * ? as int) from pg_stat_activity, pg_settings where name = ? group by setting;Times Reported Time consuming queries #12
Day Hour Count Duration Avg duration Jan 21 06 18 0ms 0ms 13 0ms 5 0ms 0ms 0ms select s.tableoid, s.oid, s.subname, ( select rolname from pg_catalog.pg_roles where oid = s.subowner) as rolname, s.subconninfo, s.subslotname, s.subsynccommit, s.subpublications from pg_subscription s where s.subdbid = ( select oid from pg_database where datname = current_database());Times Reported Time consuming queries #13
Day Hour Count Duration Avg duration Jan 21 06 5 0ms 0ms 14 0ms 1 0ms 0ms 0ms select usename, array_to_string(array ( select quote_ident(option_name) || ? || quote_literal(option_value) from pg_options_to_table(umoptions) order by option_name), e ?) as umoptions from pg_user_mappings where srvid = ? order by usename;Times Reported Time consuming queries #14
Day Hour Count Duration Avg duration Jan 21 06 1 0ms 0ms 15 0ms 2 0ms 0ms 0ms copy archive.sa_hist_consecutivecandles (id, symbolid, datetime, image, qty, percentile, direction, lastupdated, height) to stdout;Times Reported Time consuming queries #15
Day Hour Count Duration Avg duration Jan 21 06 2 0ms 0ms 16 0ms 1 0ms 0ms 0ms copy public.commoncontenttypeparams (id, name, required, description, "default", type) to stdout;Times Reported Time consuming queries #16
Day Hour Count Duration Avg duration Jan 21 06 1 0ms 0ms 17 0ms 451 0ms 0ms 0ms with rar_max as ( select resultuid from relevance_keylevels_results order by resultuid desc limit ? ), kr as ( select a.*, rr.age, rr.relevant from keylevels_results a left outer join relevance_keylevels_results rr on a.resultuid = rr.resultuid where case when false = ? then true else a.resultuid > ( select min(resultuid) from relevance_keylevels_results) end ), all_results as ( select kr.resultuid as resultuid, kr.direction as direction, s.exchange as exchange, s.symbolid as symbolid, coalesce(bim.code, s.symbol) as symbol_code, s.longname as symbol_name, s.timegranularity as interval, p.patternname as pattern_name, kr.breakout as breakout, kr.atbaridentified as identified, dtt.timezone as timezone, kr.patternlengthbars as length, g.basegroupname, newlevels.filtered, case when kr.age is not null then kr.age when kr.resultuid <= rm.resultuid then ? else ? end as age, case when kr.relevant is not null then kr.relevant when kr.resultuid <= rm.resultuid then ? else ? end as relevant, cps.pip from kr inner join brokersymbollist bsl on bsl.brokerid = ? and bsl.symbolid = kr.symbolid inner join symbols s on bsl.symbolid = s.symbolid and s.nonliquid = ? inner join symbolgroup sg on s.symbolid = sg.symbolid inner join groups g on sg.groupid = g.groupid inner join brokergroups bg on g.groupid = bg.groupid and bsl.brokerid = bg.brokerid inner join hrspatterns p on kr.patternid = p.patternid inner join downloadersymbolsettings dss on s.symbolid = dss.symbolid inner join datafeedstimetable dtt on dss.classname = dtt.classname and dtt.dayofweek = ? inner join rar_max rm on ? = ? left outer join autochartist_symbolupdates au on dss.symbolid = au.symbolid left outer join relevance_keylevels_results rar on rar.resultuid = kr.resultuid left join lateral calc_kl_signal_filter (kr.resultuid) newlevels on true left join currencypips cps on cps.symbol = s.symbol left outer join brokerinstrumentmap bim on dss.datafeedinstrumentid = bim.datafeedinstrumentid and bim.brokerid = bsl.brokerid and bim.type = ? where kr.gmttimefound > now() - interval ? and dss.enabled = ? and s.deleted = ? and (kr.simulation = ? or kr.simulation is null) and (? = ? or s.timegranularity in (...)) and (? = ? or s.exchange in (...)) and (? = ? or coalesce(bim.code, s.symbol) in (...)) and (? = ? or p.patternname in (...)) and (? = ? or kr.patternclassid in (...)) and (? = ? or kr.patternlengthbars <= ?) and kr.patternstarttime::timestamp without time zone >= coalesce(au.earliestpricedatetime, ?::timestamp without time zone) -- to make sure patternstarttime is in our t-tables ), results as ( select distinct on (symbolid) * from all_results where (false = ? or relevant = ?) and (? = ? or age <= ?) order by symbolid, resultuid ) select * from results order by identified desc, length desc limit ?;Times Reported Time consuming queries #17
Day Hour Count Duration Avg duration Jan 21 06 451 0ms 0ms 18 0ms 2 0ms 0ms 0ms lock table public.powerstats in access share mode;Times Reported Time consuming queries #18
Day Hour Count Duration Avg duration Jan 21 06 2 0ms 0ms 19 0ms 4 0ms 0ms 0ms select oid, enumlabel from pg_catalog.pg_enum where enumtypid = ? order by enumsortorder;Times Reported Time consuming queries #19
Day Hour Count Duration Avg duration Jan 21 06 4 0ms 0ms 20 0ms 2 0ms 0ms 0ms lock table public.market_report_results in access share mode;Times Reported Time consuming queries #20
Day Hour Count Duration Avg duration Jan 21 06 2 0ms 0ms Most frequent queries (N)
Rank Times executed Total duration Min duration Max duration Avg duration Query 1 13,825 0ms 0ms 0ms 0ms select ?;Times Reported Time consuming queries #1
Day Hour Count Duration Avg duration Jan 21 06 13,825 0ms 0ms 2 7,732 0ms 0ms 0ms 0ms insert into executionlogs (executionid, status, message, details, detailtype) values (null, ?, ?, null, null);Times Reported Time consuming queries #2
Day Hour Count Duration Avg duration Jan 21 06 7,732 0ms 0ms 3 7,064 0ms 0ms 0ms 0ms select distinct on (coalesce(bim.code, s.symbol) , s.exchange, s.timegranularity, df.timezone) s.symbolid as id, coalesce(bim.code, s.symbol) as name, s.symbol as symbol, dss.downloadersymbol as ticker, s.exchange as exchange, s.timegranularity as interval, df.timezone as timezone from symbols s inner join downloadersymbolsettings dss on dss.symbolid = s.symbolid inner join datafeedstimetable df on df.classname ilike dss.classname left join brokersymbollist bsl on bsl.brokerid = ? and bsl.symbolid = s.symbolid left outer join brokerinstrumentmap bim on dss.datafeedinstrumentid = bim.datafeedinstrumentid and bim.brokerid = ? and bim.type = ? where s.symbolid = ?;Times Reported Time consuming queries #3
Day Hour Count Duration Avg duration Jan 21 06 7,064 0ms 0ms 4 5,640 0ms 0ms 0ms 0ms insert into t15 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) values (?, ?, ?, ?, ?, ?, ?, ?, ?, ?) on conflict (pricedatetime, symbolid) do update set open = ?, high = ?, low = ?, close = ?, volume = ?, bsf = ?, sastdatetimewritten = ?, sastdatetimereceived = ?;Times Reported Time consuming queries #4
Day Hour Count Duration Avg duration Jan 21 06 5,640 0ms 0ms 5 5,388 0ms 0ms 0ms 0ms select s.symbolid as id, s.symbol as name, s.exchange as exchange, s.timegranularity as interval, dtt.timezone as timezone from symbols s inner join downloadersymbolsettings dss on dss.symbolid = s.symbolid inner join datafeedstimetable dtt on dss.classname = dtt.classname and dtt.dayofweek = ? inner join brokersymbollist bsl on bsl.symbolid = s.symbolid where bsl.brokerid = ? and (? = ? or s.timegranularity = ?) and (s.symbol = ? or dss.downloadersymbol = ?) and dss.enabled = ?;Times Reported Time consuming queries #5
Day Hour Count Duration Avg duration Jan 21 06 5,388 0ms 0ms 6 5,070 0ms 0ms 0ms 0ms insert into autochartist_results (resultid, symbolid, bandwidth, pattern, qtytp, gmttimefound, direction, initialtrend, breakout, volumeincrease, noise, symmetry, predictionpricefrom, predictionpriceto, predictiontimefrom, predictiontimeto, patternstarttime, patternendtime, patternstartprice, patternendprice, resx0, resx1, supportx0, supportx1, resy0, resy1, supporty0, supporty1, supportgradient, resgradient, riskreward, patternquality, trendchange, maxmovementafterbreakout, latestbaratbreakouttime, latestbaratbreakoutprice, patternlengthbars, temporarypattern, relevancestartdistance, simulation, writtendatetime) values (?, ?, ?.?, ?, ?, ?::timestamp without time zone, ?, ?.?, ?.?, ?.?, ?.?, ?.?, ?.?, ?.?, ?::timestamp without time zone, ?::timestamp without time zone, ?::timestamp without time zone, ?::timestamp without time zone, ?.?, ?.?, ?::timestamp without time zone, ?::timestamp without time zone, ?::timestamp without time zone, ?::timestamp without time zone, ?.?, ?.?, ?.?, ?.?, ?.?, ?.?, ?.?, ?.?, ?, ?.?, ?::timestamp without time zone, ?.?, ?, ?, ?.?, ?, current_timestamp::timestamp without time zone) on conflict do nothing;Times Reported Time consuming queries #6
Day Hour Count Duration Avg duration Jan 21 06 5,070 0ms 0ms 7 3,226 0ms 0ms 0ms 0ms insert into t30 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) values (?, ?, ?, ?, ?, ?, ?, ?, ?, ?) on conflict (pricedatetime, symbolid) do update set open = ?, high = ?, low = ?, close = ?, volume = ?, bsf = ?, sastdatetimewritten = ?, sastdatetimereceived = ?;Times Reported Time consuming queries #7
Day Hour Count Duration Avg duration Jan 21 06 3,226 0ms 0ms 8 2,505 0ms 0ms 0ms 0ms insert into fibonacci_results (bandwidth, pattern, gmttimefound, direction, patternstarttime, patternendtime, patternstartprice, patternendprice, qtytp, pricex, timex, pricea, timea, priceb, timeb, pricec, timec, priced, timed, averagequality, timequality, errormargin, patternlengthbars, target10, target06, target16, target07, target12, target05, target03, symbolid, noise, ratiosfound, temporarypattern, uniqueindex, completed, simulation, writtendatetime) values (?.?, ?, ?::timestamp without time zone, ?, ?::timestamp without time zone, ?::timestamp without time zone, ?.?, ?.?, ?, ?.?, ?::timestamp without time zone, ?.?, ?::timestamp without time zone, ?.?, ?::timestamp without time zone, ?.?, ?::timestamp without time zone, ?.?, ?::timestamp without time zone, ?.?, ?.?, ?.?, ?, ?.?, ?.?, ?.?, ?.?, ?.?, ?.?, ?.?, ?, ?.?, ?, ?, ?, ?, ?, current_timestamp::timestamp without time zone) on conflict do nothing;Times Reported Time consuming queries #8
Day Hour Count Duration Avg duration Jan 21 06 2,505 0ms 0ms 9 2,114 0ms 0ms 0ms 0ms insert into t60 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) values (?, ?, ?, ?, ?, ?, ?, ?, ?, ?) on conflict (pricedatetime, symbolid) do update set open = ?, high = ?, low = ?, close = ?, volume = ?, bsf = ?, sastdatetimewritten = ?, sastdatetimereceived = ?;Times Reported Time consuming queries #9
Day Hour Count Duration Avg duration Jan 21 06 2,114 0ms 0ms 10 1,815 0ms 0ms 0ms 0ms set extra_float_digits = ?;Times Reported Time consuming queries #10
Day Hour Count Duration Avg duration Jan 21 06 1,815 0ms 0ms 11 1,789 0ms 0ms 0ms 0ms set application_name = ?;Times Reported Time consuming queries #11
Day Hour Count Duration Avg duration Jan 21 06 1,789 0ms 0ms 12 1,656 0ms 0ms 0ms 0ms select pg_catalog.format_type(?::pg_catalog.oid, null);Times Reported Time consuming queries #12
Day Hour Count Duration Avg duration Jan 21 06 1,656 0ms 0ms 13 1,234 0ms 0ms 0ms 0ms update patternresultsrelevance set relevant = ?, saxo_relevant = ?, notrelevantpricedatetime = ?, reason = ? where uniqueindex = ? and relevant = ?;Times Reported Time consuming queries #13
Day Hour Count Duration Avg duration Jan 21 06 1,234 0ms 0ms 14 1,075 0ms 0ms 0ms 0ms insert into keylevels_results (bandwidth, breakout, patternid, gmttimefound, approachingtimestamp, approachingregion, qtytp, patternlengthbars, patternprice, x0, x1, x2, x3, x4, x5, x6, x7, x8, x9, breakoutbars, breakoutprice, patternendtime, atbaridentified, atpriceidentified, errormargin, direction, symbolid, predictionpricefrom, predictionpriceto, predictiontimefrom, predictiontimebars, uniquepointsvalue, furthestprice, relevancestartdistance, patternclassid, patternstarttime, stoplosslevel, simulation, writtendatetime) values (?.?, ?, ?, ?::timestamp without time zone, ?, ?.?, ?, ?, ?.?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?.?, ?::timestamp without time zone, ?, ?.?, ?.?, ?, ?, ?.?, ?.?, ?::timestamp without time zone, ?, ?, ?.?, ?.?, ?, ?, ?.?, ?, current_timestamp::timestamp without time zone) on conflict do nothing;Times Reported Time consuming queries #14
Day Hour Count Duration Avg duration Jan 21 06 1,075 0ms 0ms 15 980 0ms 0ms 0ms 0ms select symbolid, pricedatetime, classname, downloadfrequency, downloadersymbol, open, high, low, close, volume, bsf, sastdatetimereceived from ( select pricedatetime, dss.classname, dss.downloadfrequency, dss.symbolid, dss.downloadersymbol, t.open, t.high, t.low, t.close, t.volume, t.bsf, t.sastdatetimereceived, row_number() over (partition by t.symbolid order by t.pricedatetime desc) as rn from t15 t, downloadersymbolsettings dss, symbols s where dss.classname = ? and dss.downloadfrequency = ? and dss.symbolid = t.symbolid and s.symbolid = dss.symbolid and dss.enabled = ? and s.deleted = ? and dss.downloadersymbol in (...) and t.pricedatetime > now() - interval ?) as ranked_candles_table where rn = ?;Times Reported Time consuming queries #15
Day Hour Count Duration Avg duration Jan 21 06 980 0ms 0ms 16 975 0ms 0ms 0ms 0ms select oid, tableoid, pol.polname, pol.polcmd, pol.polpermissive, case when pol.polroles = ? then null else pg_catalog.array_to_string(array ( select pg_catalog.quote_ident(rolname) from pg_catalog.pg_roles where oid = any (pol.polroles)), ?) end as polroles, pg_catalog.pg_get_expr(pol.polqual, pol.polrelid) as polqual, pg_catalog.pg_get_expr(pol.polwithcheck, pol.polrelid) as polwithcheck from pg_catalog.pg_policy pol where polrelid = ?;Times Reported Time consuming queries #16
Day Hour Count Duration Avg duration Jan 21 06 975 0ms 0ms 17 902 0ms 0ms 0ms 0ms with rar_max as ( select resultuid from relevance_autochartist_results order by resultuid desc limit ? ) select a.symbolid, pattern, patternid, resy0, resy1, resx0, resx1, supporty0, supporty1, supportx0, supportx1, predictiontimeto, patternstarttime, timegranularity, patternendtime, direction, trendchange, patternlengthbars, patternquality, case when a.old_resultuid = ? then a.old_resultuid else a.resultuid end as uid, breakout, initialtrend, volumeincrease, symmetry as uniformity, predictionpricefrom, predictionpriceto, noise, s.exchange, s.symbol, s.longname, s.shortname, breakout, dtt.timezone, patternstartprice, patternendprice, qtytp, newlevels.profit, newlevels.stop, newlevels.filtered, case when rar.age is not null then rar.age when a.resultuid <= rm.resultuid then ? else ? end as age, case when rar.relevant is not null then rar.relevant when a.resultuid <= rm.resultuid then ? else ? end as relevant, cps.pip from autochartist_results a inner join downloadersymbolsettings dss on a.symbolid = dss.symbolid inner join datafeedstimetable dtt on dss.classname = dtt.classname inner join symbols s on a.symbolid = s.symbolid inner join patterns p on p.patternname = a.pattern inner join rar_max rm on ? = ? left outer join relevance_autochartist_results rar on rar.resultuid = a.resultuid left join lateral calc_cp_signal (a.resultuid) newlevels on true left join currencypips cps on cps.symbol = s.symbol where (a.old_resultuid = ? or a.resultuid = ?) and dtt.dayofweek = ?;Times Reported Time consuming queries #17
Day Hour Count Duration Avg duration Jan 21 06 902 0ms 0ms 18 785 0ms 0ms 0ms 0ms select a.attnum, a.attname, a.atttypmod, a.attstattarget, a.attstorage, t.typstorage, a.attnotnull, a.atthasdef, a.attisdropped, a.attlen, a.attalign, a.attislocal, pg_catalog.format_type(t.oid, a.atttypmod) as atttypname, a.attgenerated, case when a.atthasmissing and not a.attisdropped then a.attmissingval else null end as attmissingval, a.attidentity, pg_catalog.array_to_string(array ( select pg_catalog.quote_ident(option_name) || ? || pg_catalog.quote_literal(option_value) from pg_catalog.pg_options_to_table(attfdwoptions) order by option_name), e ?) as attfdwoptions, case when a.attcollation <> t.typcollation then a.attcollation else ? end as attcollation, array_to_string(a.attoptions, ?) as attoptions from pg_catalog.pg_attribute a left join pg_catalog.pg_type t on a.atttypid = t.oid where a.attrelid = ?::pg_catalog.oid and a.attnum > ?::pg_catalog.int2 order by a.attnum;Times Reported Time consuming queries #18
Day Hour Count Duration Avg duration Jan 21 06 785 0ms 0ms 19 650 0ms 0ms 0ms 0ms select proretset, prosrc, probin, pg_catalog.pg_get_function_arguments(oid) as funcargs, pg_catalog.pg_get_function_identity_arguments(oid) as funciargs, pg_catalog.pg_get_function_result(oid) as funcresult, array_to_string(protrftypes, ?) as protrftypes, prokind, provolatile, proisstrict, prosecdef, proleakproof, proconfig, procost, prorows, prosupport, proparallel, ( select lanname from pg_catalog.pg_language where oid = prolang) as lanname from pg_catalog.pg_proc where oid = ?::pg_catalog.oid;Times Reported Time consuming queries #19
Day Hour Count Duration Avg duration Jan 21 06 650 0ms 0ms 20 627 0ms 0ms 0ms 0ms select at.attname, ( select pg_catalog.array_agg(acl order by row_n) from ( select acl, row_n from pg_catalog.unnest(coalesce(at.attacl, pg_catalog.acldefault(?, c.relowner))) with ordinality as perm (acl, row_n) where not exists ( select ? from pg_catalog.unnest(coalesce(pip.initprivs, pg_catalog.acldefault(?, c.relowner))) as init (init_acl) where acl = init_acl)) as foo) as attacl, ( select pg_catalog.array_agg(acl order by row_n) from ( select acl, row_n from pg_catalog.unnest(coalesce(pip.initprivs, pg_catalog.acldefault(?, c.relowner))) with ordinality as initp (acl, row_n) where not exists ( select ? from pg_catalog.unnest(coalesce(at.attacl, pg_catalog.acldefault(?, c.relowner))) as permp (orig_acl) where acl = orig_acl)) as foo) as rattacl, null as initattacl, null as initrattacl from pg_catalog.pg_attribute at join pg_catalog.pg_class c on (at.attrelid = c.oid) left join pg_catalog.pg_init_privs pip on (at.attrelid = pip.objoid and pip.classoid = ?::pg_catalog.regclass and at.attnum = pip.objsubid) where at.attrelid = ?::pg_catalog.oid and not at.attisdropped and (( select pg_catalog.array_agg(acl order by row_n) from ( select acl, row_n from pg_catalog.unnest(coalesce(at.attacl, pg_catalog.acldefault(?, c.relowner))) with ordinality as perm (acl, row_n) where not exists ( select ? from pg_catalog.unnest(coalesce(pip.initprivs, pg_catalog.acldefault(?, c.relowner))) as init (init_acl) where acl = init_acl)) as foo) is not null or ( select pg_catalog.array_agg(acl order by row_n) from ( select acl, row_n from pg_catalog.unnest(coalesce(pip.initprivs, pg_catalog.acldefault(?, c.relowner))) with ordinality as initp (acl, row_n) where not exists ( select ? from pg_catalog.unnest(coalesce(at.attacl, pg_catalog.acldefault(?, c.relowner))) as permp (orig_acl) where acl = orig_acl)) as foo) is not null or null is not null or null is not null) order by at.attnum;Times Reported Time consuming queries #20
Day Hour Count Duration Avg duration Jan 21 06 627 0ms 0ms Normalized slowest queries (N)
Rank Min duration Max duration Avg duration Times executed Total duration Query 1 0ms 0ms 0ms 1 0ms lock table public.stats_hrs_summary_by_groups in access share mode;Times Reported Time consuming queries #1
Day Hour Count Duration Avg duration Jan 21 06 1 0ms 0ms 2 0ms 0ms 0ms 2 0ms lock table public.timezones_korean in access share mode;Times Reported Time consuming queries #2
Day Hour Count Duration Avg duration Jan 21 06 2 0ms 0ms 3 0ms 0ms 0ms 32 0ms select key, value from datasources ds inner join datasourceparams dsp on ds.id = dsp.datasourceid where ds.name = ?;Times Reported Time consuming queries #3
Day Hour Count Duration Avg duration Jan 21 06 32 0ms 0ms 4 0ms 0ms 0ms 2 0ms copy public.t30 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived, pricesadjusted, relevanceprocessed, recengprocessedadjusted, historical) to stdout;Times Reported Time consuming queries #4
Day Hour Count Duration Avg duration Jan 21 06 2 0ms 0ms 5 0ms 0ms 0ms 1 0ms copy public.processresults (id, processid, resultdate, shorttext, longtext, title, test) to stdout;Times Reported Time consuming queries #5
Day Hour Count Duration Avg duration Jan 21 06 1 0ms 0ms 6 0ms 0ms 0ms 2,114 0ms insert into t60 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) values (?, ?, ?, ?, ?, ?, ?, ?, ?, ?) on conflict (pricedatetime, symbolid) do update set open = ?, high = ?, low = ?, close = ?, volume = ?, bsf = ?, sastdatetimewritten = ?, sastdatetimereceived = ?;Times Reported Time consuming queries #6
Day Hour Count Duration Avg duration Jan 21 06 2,114 0ms 0ms 7 0ms 0ms 0ms 975 0ms select oid, tableoid, pol.polname, pol.polcmd, pol.polpermissive, case when pol.polroles = ? then null else pg_catalog.array_to_string(array ( select pg_catalog.quote_ident(rolname) from pg_catalog.pg_roles where oid = any (pol.polroles)), ?) end as polroles, pg_catalog.pg_get_expr(pol.polqual, pol.polrelid) as polqual, pg_catalog.pg_get_expr(pol.polwithcheck, pol.polrelid) as polwithcheck from pg_catalog.pg_policy pol where polrelid = ?;Times Reported Time consuming queries #7
Day Hour Count Duration Avg duration Jan 21 06 975 0ms 0ms 8 0ms 0ms 0ms 2 0ms copy public.timezones_greek (timezoneid, timezone) to stdout;Times Reported Time consuming queries #8
Day Hour Count Duration Avg duration Jan 21 06 2 0ms 0ms 9 0ms 0ms 0ms 2 0ms lock table public.futures_symbols in access share mode;Times Reported Time consuming queries #9
Day Hour Count Duration Avg duration Jan 21 06 2 0ms 0ms 10 0ms 0ms 0ms 2 0ms copy archive.autochartist_results (resultid, bandwidth, pattern, gmttimefound, direction, initialtrend, breakout, volumeincrease, noise, symmetry, patternstarttime, patternendtime, patternstartprice, patternendprice, resx0, resx1, supportx0, supportx1, resy0, resy1, supporty0, supporty1, supportgradient, resgradient, riskreward, qtytp, predictionpricefrom, predictionpriceto, predictiontimefrom, predictiontimeto, patternquality, trendchange, maxmovementafterbreakout, latestbaratbreakouttime, latestbaratbreakoutprice, patternlengthbars, temporarypattern, symbolid, resultuid, relevancestartdistance, simulation, writtendatetime, old_resultuid) to stdout;Times Reported Time consuming queries #10
Day Hour Count Duration Avg duration Jan 21 06 2 0ms 0ms 11 0ms 0ms 0ms 2 0ms copy public.satelliteauthentication (id, satelliteserverid, brokerid, authenticationstatus, maketradeurl, notifications, validusername, validpassword, enabled) to stdout;Times Reported Time consuming queries #11
Day Hour Count Duration Avg duration Jan 21 06 2 0ms 0ms 12 0ms 0ms 0ms 18 0ms select cast(count(*) / cast(setting as numeric) * ? as int) from pg_stat_activity, pg_settings where name = ? group by setting;Times Reported Time consuming queries #12
Day Hour Count Duration Avg duration Jan 21 06 18 0ms 0ms 13 0ms 0ms 0ms 5 0ms select s.tableoid, s.oid, s.subname, ( select rolname from pg_catalog.pg_roles where oid = s.subowner) as rolname, s.subconninfo, s.subslotname, s.subsynccommit, s.subpublications from pg_subscription s where s.subdbid = ( select oid from pg_database where datname = current_database());Times Reported Time consuming queries #13
Day Hour Count Duration Avg duration Jan 21 06 5 0ms 0ms 14 0ms 0ms 0ms 1 0ms select usename, array_to_string(array ( select quote_ident(option_name) || ? || quote_literal(option_value) from pg_options_to_table(umoptions) order by option_name), e ?) as umoptions from pg_user_mappings where srvid = ? order by usename;Times Reported Time consuming queries #14
Day Hour Count Duration Avg duration Jan 21 06 1 0ms 0ms 15 0ms 0ms 0ms 2 0ms copy archive.sa_hist_consecutivecandles (id, symbolid, datetime, image, qty, percentile, direction, lastupdated, height) to stdout;Times Reported Time consuming queries #15
Day Hour Count Duration Avg duration Jan 21 06 2 0ms 0ms 16 0ms 0ms 0ms 1 0ms copy public.commoncontenttypeparams (id, name, required, description, "default", type) to stdout;Times Reported Time consuming queries #16
Day Hour Count Duration Avg duration Jan 21 06 1 0ms 0ms 17 0ms 0ms 0ms 451 0ms with rar_max as ( select resultuid from relevance_keylevels_results order by resultuid desc limit ? ), kr as ( select a.*, rr.age, rr.relevant from keylevels_results a left outer join relevance_keylevels_results rr on a.resultuid = rr.resultuid where case when false = ? then true else a.resultuid > ( select min(resultuid) from relevance_keylevels_results) end ), all_results as ( select kr.resultuid as resultuid, kr.direction as direction, s.exchange as exchange, s.symbolid as symbolid, coalesce(bim.code, s.symbol) as symbol_code, s.longname as symbol_name, s.timegranularity as interval, p.patternname as pattern_name, kr.breakout as breakout, kr.atbaridentified as identified, dtt.timezone as timezone, kr.patternlengthbars as length, g.basegroupname, newlevels.filtered, case when kr.age is not null then kr.age when kr.resultuid <= rm.resultuid then ? else ? end as age, case when kr.relevant is not null then kr.relevant when kr.resultuid <= rm.resultuid then ? else ? end as relevant, cps.pip from kr inner join brokersymbollist bsl on bsl.brokerid = ? and bsl.symbolid = kr.symbolid inner join symbols s on bsl.symbolid = s.symbolid and s.nonliquid = ? inner join symbolgroup sg on s.symbolid = sg.symbolid inner join groups g on sg.groupid = g.groupid inner join brokergroups bg on g.groupid = bg.groupid and bsl.brokerid = bg.brokerid inner join hrspatterns p on kr.patternid = p.patternid inner join downloadersymbolsettings dss on s.symbolid = dss.symbolid inner join datafeedstimetable dtt on dss.classname = dtt.classname and dtt.dayofweek = ? inner join rar_max rm on ? = ? left outer join autochartist_symbolupdates au on dss.symbolid = au.symbolid left outer join relevance_keylevels_results rar on rar.resultuid = kr.resultuid left join lateral calc_kl_signal_filter (kr.resultuid) newlevels on true left join currencypips cps on cps.symbol = s.symbol left outer join brokerinstrumentmap bim on dss.datafeedinstrumentid = bim.datafeedinstrumentid and bim.brokerid = bsl.brokerid and bim.type = ? where kr.gmttimefound > now() - interval ? and dss.enabled = ? and s.deleted = ? and (kr.simulation = ? or kr.simulation is null) and (? = ? or s.timegranularity in (...)) and (? = ? or s.exchange in (...)) and (? = ? or coalesce(bim.code, s.symbol) in (...)) and (? = ? or p.patternname in (...)) and (? = ? or kr.patternclassid in (...)) and (? = ? or kr.patternlengthbars <= ?) and kr.patternstarttime::timestamp without time zone >= coalesce(au.earliestpricedatetime, ?::timestamp without time zone) -- to make sure patternstarttime is in our t-tables ), results as ( select distinct on (symbolid) * from all_results where (false = ? or relevant = ?) and (? = ? or age <= ?) order by symbolid, resultuid ) select * from results order by identified desc, length desc limit ?;Times Reported Time consuming queries #17
Day Hour Count Duration Avg duration Jan 21 06 451 0ms 0ms 18 0ms 0ms 0ms 2 0ms lock table public.powerstats in access share mode;Times Reported Time consuming queries #18
Day Hour Count Duration Avg duration Jan 21 06 2 0ms 0ms 19 0ms 0ms 0ms 4 0ms select oid, enumlabel from pg_catalog.pg_enum where enumtypid = ? order by enumsortorder;Times Reported Time consuming queries #19
Day Hour Count Duration Avg duration Jan 21 06 4 0ms 0ms 20 0ms 0ms 0ms 2 0ms lock table public.market_report_results in access share mode;Times Reported Time consuming queries #20
Day Hour Count Duration Avg duration Jan 21 06 2 0ms 0ms Time consuming prepare
Rank Total duration Times executed Min duration Max duration Avg duration Query 1 4m17s 12 0ms 2m26s 21s487ms select recognitionengine, to_char(datetimeupdate, 'yyyy-mm-dd HH24:MI') from latest_candle_datetime_per_receng;Times Reported Time consuming prepare #1
Day Hour Count Duration Avg duration Jan 21 06 12 4m17s 21s487ms -
select recognitionengine, to_char(datetimeupdate, 'yyyy-mm-dd HH24:MI') from latest_candle_datetime_per_receng;
Date: 2026-01-21 06:19:31 Duration: 2m26s Database: postgres
-
select recognitionengine, to_char(datetimeupdate, 'yyyy-mm-dd HH24:MI') from latest_candle_datetime_per_receng;
Date: 2026-01-21 06:19:31 Duration: 1m51s Database: postgres
-
select recognitionengine, to_char(datetimeupdate, 'yyyy-mm-dd HH24:MI') from latest_candle_datetime_per_receng;
Date: 2026-01-21 06:02:38 Duration: 0ms Database: postgres
2 3s91ms 2,680 0ms 18ms 1ms WITH rar_max as ( ;Times Reported Time consuming prepare #2
Day Hour Count Duration Avg duration 06 2,680 3s91ms 1ms -
WITH rar_max as ( ;
Date: 2026-01-21 06:26:03 Duration: 18ms Database: postgres
-
WITH rar_max as ( ;
Date: 2026-01-21 06:38:31 Duration: 15ms Database: postgres
-
WITH rar_max as ( ;
Date: 2026-01-21 06:16:38 Duration: 14ms Database: postgres
3 1s385ms 1,085 0ms 3ms 1ms SELECT symbolid, ;Times Reported Time consuming prepare #3
Day Hour Count Duration Avg duration 06 1,085 1s385ms 1ms -
SELECT symbolid, ;
Date: 2026-01-21 06:17:02 Duration: 3ms Database: postgres
-
SELECT symbolid, ;
Date: 2026-01-21 06:30:58 Duration: 2ms Database: postgres
-
SELECT symbolid, ;
Date: 2026-01-21 06:46:12 Duration: 2ms Database: postgres
4 888ms 2,663 0ms 8ms 0ms SELECT ;Times Reported Time consuming prepare #4
Day Hour Count Duration Avg duration 06 2,663 888ms 0ms -
SELECT ;
Date: 2026-01-21 06:47:41 Duration: 8ms Database: postgres
-
SELECT ;
Date: 2026-01-21 06:46:12 Duration: 5ms Database: postgres
-
SELECT ;
Date: 2026-01-21 06:16:04 Duration: 5ms Database: postgres
5 668ms 575 0ms 2ms 1ms SELECT s.symbolid, dss.downloadfrequency, dss.downloadersymbol;Times Reported Time consuming prepare #5
Day Hour Count Duration Avg duration 06 575 668ms 1ms -
SELECT s.symbolid, dss.downloadfrequency, dss.downloadersymbol;
Date: 2026-01-21 06:31:02 Duration: 2ms Database: postgres
-
SELECT s.symbolid, dss.downloadfrequency, dss.downloadersymbol;
Date: 2026-01-21 06:15:52 Duration: 1ms Database: postgres
-
SELECT s.symbolid, dss.downloadfrequency, dss.downloadersymbol;
Date: 2026-01-21 06:16:41 Duration: 1ms Database: postgres
6 274ms 2,917 0ms 0ms 0ms INSERT INTO T30 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;Times Reported Time consuming prepare #6
Day Hour Count Duration Avg duration 06 2,917 274ms 0ms -
INSERT INTO T30 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;
Date: 2026-01-21 06:45:23 Duration: 0ms Database: postgres
-
INSERT INTO T30 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;
Date: 2026-01-21 06:40:53 Duration: 0ms Database: postgres
-
INSERT INTO T30 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;
Date: 2026-01-21 06:11:53 Duration: 0ms Database: postgres
7 262ms 1,815 0ms 6ms 0ms SET extra_float_digits = 3;Times Reported Time consuming prepare #7
Day Hour Count Duration Avg duration 06 1,815 262ms 0ms -
SET extra_float_digits = 3;
Date: 2026-01-21 06:58:02 Duration: 6ms Database: postgres
-
SET extra_float_digits = 3;
Date: 2026-01-21 06:31:44 Duration: 2ms Database: postgres
-
SET extra_float_digits = 3;
Date: 2026-01-21 06:46:09 Duration: 2ms Database: postgres
8 211ms 1,959 0ms 1ms 0ms INSERT INTO T60 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;Times Reported Time consuming prepare #8
Day Hour Count Duration Avg duration 06 1,959 211ms 0ms -
INSERT INTO T60 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;
Date: 2026-01-21 06:00:54 Duration: 1ms Database: postgres
-
INSERT INTO T60 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;
Date: 2026-01-21 06:02:39 Duration: 0ms Database: postgres
-
INSERT INTO T60 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;
Date: 2026-01-21 06:00:50 Duration: 0ms Database: postgres
9 159ms 993 0ms 0ms 0ms INSERT INTO T15 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;Times Reported Time consuming prepare #9
Day Hour Count Duration Avg duration 06 993 159ms 0ms -
INSERT INTO T15 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;
Date: 2026-01-21 06:47:41 Duration: 0ms Database: postgres
-
INSERT INTO T15 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;
Date: 2026-01-21 06:11:53 Duration: 0ms Database: postgres
-
INSERT INTO T15 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;
Date: 2026-01-21 06:40:51 Duration: 0ms Database: postgres
10 123ms 16 4ms 23ms 7ms with sym_info as ( ;Times Reported Time consuming prepare #10
Day Hour Count Duration Avg duration 06 16 123ms 7ms -
with sym_info as ( ;
Date: 2026-01-21 06:36:43 Duration: 23ms Database: postgres
-
with sym_info as ( ;
Date: 2026-01-21 06:21:45 Duration: 11ms Database: postgres
-
with sym_info as ( ;
Date: 2026-01-21 06:51:43 Duration: 9ms Database: postgres
11 78ms 1,716 0ms 2ms 0ms select 1;Times Reported Time consuming prepare #11
Day Hour Count Duration Avg duration 06 1,716 78ms 0ms -
select 1;
Date: 2026-01-21 06:21:21 Duration: 2ms Database: postgres
-
select 1;
Date: 2026-01-21 06:46:37 Duration: 1ms Database: postgres
-
select 1;
Date: 2026-01-21 06:47:41 Duration: 0ms Database: postgres
12 74ms 18 1ms 28ms 4ms select cast(count(*) / cast(setting as numeric) * 100 as int) from pg_stat_activity, pg_settings WHERE name = 'max_connections' group by setting;Times Reported Time consuming prepare #12
Day Hour Count Duration Avg duration 06 18 74ms 4ms -
select cast(count(*) / cast(setting as numeric) * 100 as int) from pg_stat_activity, pg_settings WHERE name = 'max_connections' group by setting;
Date: 2026-01-21 06:40:02 Duration: 28ms Database: postgres
-
select cast(count(*) / cast(setting as numeric) * 100 as int) from pg_stat_activity, pg_settings WHERE name = 'max_connections' group by setting;
Date: 2026-01-21 06:41:11 Duration: 3ms Database: postgres
-
select cast(count(*) / cast(setting as numeric) * 100 as int) from pg_stat_activity, pg_settings WHERE name = 'max_connections' group by setting;
Date: 2026-01-21 06:31:01 Duration: 2ms Database: postgres
13 63ms 278 0ms 0ms 0ms SELECT NULL AS TABLE_CAT, n.nspname AS TABLE_SCHEM, c.relname AS TABLE_NAME, CASE n.nspname ~ '^pg_' OR n.nspname = 'information_schema' WHEN true THEN CASE WHEN n.nspname = 'pg_catalog' OR n.nspname = 'information_schema' THEN CASE c.relkind WHEN 'r' THEN 'SYSTEM TABLE' WHEN 'v' THEN 'SYSTEM VIEW' WHEN 'i' THEN 'SYSTEM INDEX' ELSE NULL END WHEN n.nspname = 'pg_toast' THEN CASE c.relkind WHEN 'r' THEN 'SYSTEM TOAST TABLE' WHEN 'i' THEN 'SYSTEM TOAST INDEX' ELSE NULL END ELSE CASE c.relkind WHEN 'r' THEN 'TEMPORARY TABLE' WHEN 'p' THEN 'TEMPORARY TABLE' WHEN 'i' THEN 'TEMPORARY INDEX' WHEN 'S' THEN 'TEMPORARY SEQUENCE' WHEN 'v' THEN 'TEMPORARY VIEW' ELSE NULL END END WHEN false THEN CASE c.relkind WHEN 'r' THEN 'TABLE' WHEN 'p' THEN 'PARTITIONED TABLE' WHEN 'i' THEN 'INDEX' WHEN 'S' THEN 'SEQUENCE' WHEN 'v' THEN 'VIEW' WHEN 'c' THEN 'TYPE' WHEN 'f' THEN 'FOREIGN TABLE' WHEN 'm' THEN 'MATERIALIZED VIEW' ELSE NULL END ELSE NULL END AS TABLE_TYPE, d.description AS REMARKS, '' as TYPE_CAT, '' as TYPE_SCHEM, '' as TYPE_NAME, '' AS SELF_REFERENCING_COL_NAME, '' AS REF_GENERATION FROM pg_catalog.pg_namespace n, pg_catalog.pg_class c LEFT JOIN pg_catalog.pg_description d ON (c.oid = d.objoid AND d.objsubid = 0) LEFT JOIN pg_catalog.pg_class dc ON (d.classoid = dc.oid AND dc.relname = 'pg_class') LEFT JOIN pg_catalog.pg_namespace dn ON (dn.oid = dc.relnamespace AND dn.nspname = 'pg_catalog') WHERE c.relnamespace = n.oid AND c.relname LIKE 'PROBABLYNOT' AND (false OR (c.relkind = 'r' AND n.nspname !~ '^pg_' AND n.nspname <> 'information_schema')) ORDER BY TABLE_TYPE, TABLE_SCHEM, TABLE_NAME;Times Reported Time consuming prepare #13
Day Hour Count Duration Avg duration 06 278 63ms 0ms -
SELECT NULL AS TABLE_CAT, n.nspname AS TABLE_SCHEM, c.relname AS TABLE_NAME, CASE n.nspname ~ '^pg_' OR n.nspname = 'information_schema' WHEN true THEN CASE WHEN n.nspname = 'pg_catalog' OR n.nspname = 'information_schema' THEN CASE c.relkind WHEN 'r' THEN 'SYSTEM TABLE' WHEN 'v' THEN 'SYSTEM VIEW' WHEN 'i' THEN 'SYSTEM INDEX' ELSE NULL END WHEN n.nspname = 'pg_toast' THEN CASE c.relkind WHEN 'r' THEN 'SYSTEM TOAST TABLE' WHEN 'i' THEN 'SYSTEM TOAST INDEX' ELSE NULL END ELSE CASE c.relkind WHEN 'r' THEN 'TEMPORARY TABLE' WHEN 'p' THEN 'TEMPORARY TABLE' WHEN 'i' THEN 'TEMPORARY INDEX' WHEN 'S' THEN 'TEMPORARY SEQUENCE' WHEN 'v' THEN 'TEMPORARY VIEW' ELSE NULL END END WHEN false THEN CASE c.relkind WHEN 'r' THEN 'TABLE' WHEN 'p' THEN 'PARTITIONED TABLE' WHEN 'i' THEN 'INDEX' WHEN 'S' THEN 'SEQUENCE' WHEN 'v' THEN 'VIEW' WHEN 'c' THEN 'TYPE' WHEN 'f' THEN 'FOREIGN TABLE' WHEN 'm' THEN 'MATERIALIZED VIEW' ELSE NULL END ELSE NULL END AS TABLE_TYPE, d.description AS REMARKS, '' as TYPE_CAT, '' as TYPE_SCHEM, '' as TYPE_NAME, '' AS SELF_REFERENCING_COL_NAME, '' AS REF_GENERATION FROM pg_catalog.pg_namespace n, pg_catalog.pg_class c LEFT JOIN pg_catalog.pg_description d ON (c.oid = d.objoid AND d.objsubid = 0) LEFT JOIN pg_catalog.pg_class dc ON (d.classoid = dc.oid AND dc.relname = 'pg_class') LEFT JOIN pg_catalog.pg_namespace dn ON (dn.oid = dc.relnamespace AND dn.nspname = 'pg_catalog') WHERE c.relnamespace = n.oid AND c.relname LIKE 'PROBABLYNOT' AND (false OR (c.relkind = 'r' AND n.nspname !~ '^pg_' AND n.nspname <> 'information_schema')) ORDER BY TABLE_TYPE, TABLE_SCHEM, TABLE_NAME;
Date: 2026-01-21 06:11:49 Duration: 0ms Database: postgres
-
SELECT NULL AS TABLE_CAT, n.nspname AS TABLE_SCHEM, c.relname AS TABLE_NAME, CASE n.nspname ~ '^pg_' OR n.nspname = 'information_schema' WHEN true THEN CASE WHEN n.nspname = 'pg_catalog' OR n.nspname = 'information_schema' THEN CASE c.relkind WHEN 'r' THEN 'SYSTEM TABLE' WHEN 'v' THEN 'SYSTEM VIEW' WHEN 'i' THEN 'SYSTEM INDEX' ELSE NULL END WHEN n.nspname = 'pg_toast' THEN CASE c.relkind WHEN 'r' THEN 'SYSTEM TOAST TABLE' WHEN 'i' THEN 'SYSTEM TOAST INDEX' ELSE NULL END ELSE CASE c.relkind WHEN 'r' THEN 'TEMPORARY TABLE' WHEN 'p' THEN 'TEMPORARY TABLE' WHEN 'i' THEN 'TEMPORARY INDEX' WHEN 'S' THEN 'TEMPORARY SEQUENCE' WHEN 'v' THEN 'TEMPORARY VIEW' ELSE NULL END END WHEN false THEN CASE c.relkind WHEN 'r' THEN 'TABLE' WHEN 'p' THEN 'PARTITIONED TABLE' WHEN 'i' THEN 'INDEX' WHEN 'S' THEN 'SEQUENCE' WHEN 'v' THEN 'VIEW' WHEN 'c' THEN 'TYPE' WHEN 'f' THEN 'FOREIGN TABLE' WHEN 'm' THEN 'MATERIALIZED VIEW' ELSE NULL END ELSE NULL END AS TABLE_TYPE, d.description AS REMARKS, '' as TYPE_CAT, '' as TYPE_SCHEM, '' as TYPE_NAME, '' AS SELF_REFERENCING_COL_NAME, '' AS REF_GENERATION FROM pg_catalog.pg_namespace n, pg_catalog.pg_class c LEFT JOIN pg_catalog.pg_description d ON (c.oid = d.objoid AND d.objsubid = 0) LEFT JOIN pg_catalog.pg_class dc ON (d.classoid = dc.oid AND dc.relname = 'pg_class') LEFT JOIN pg_catalog.pg_namespace dn ON (dn.oid = dc.relnamespace AND dn.nspname = 'pg_catalog') WHERE c.relnamespace = n.oid AND c.relname LIKE 'PROBABLYNOT' AND (false OR (c.relkind = 'r' AND n.nspname !~ '^pg_' AND n.nspname <> 'information_schema')) ORDER BY TABLE_TYPE, TABLE_SCHEM, TABLE_NAME;
Date: 2026-01-21 06:11:50 Duration: 0ms Database: postgres
-
SELECT NULL AS TABLE_CAT, n.nspname AS TABLE_SCHEM, c.relname AS TABLE_NAME, CASE n.nspname ~ '^pg_' OR n.nspname = 'information_schema' WHEN true THEN CASE WHEN n.nspname = 'pg_catalog' OR n.nspname = 'information_schema' THEN CASE c.relkind WHEN 'r' THEN 'SYSTEM TABLE' WHEN 'v' THEN 'SYSTEM VIEW' WHEN 'i' THEN 'SYSTEM INDEX' ELSE NULL END WHEN n.nspname = 'pg_toast' THEN CASE c.relkind WHEN 'r' THEN 'SYSTEM TOAST TABLE' WHEN 'i' THEN 'SYSTEM TOAST INDEX' ELSE NULL END ELSE CASE c.relkind WHEN 'r' THEN 'TEMPORARY TABLE' WHEN 'p' THEN 'TEMPORARY TABLE' WHEN 'i' THEN 'TEMPORARY INDEX' WHEN 'S' THEN 'TEMPORARY SEQUENCE' WHEN 'v' THEN 'TEMPORARY VIEW' ELSE NULL END END WHEN false THEN CASE c.relkind WHEN 'r' THEN 'TABLE' WHEN 'p' THEN 'PARTITIONED TABLE' WHEN 'i' THEN 'INDEX' WHEN 'S' THEN 'SEQUENCE' WHEN 'v' THEN 'VIEW' WHEN 'c' THEN 'TYPE' WHEN 'f' THEN 'FOREIGN TABLE' WHEN 'm' THEN 'MATERIALIZED VIEW' ELSE NULL END ELSE NULL END AS TABLE_TYPE, d.description AS REMARKS, '' as TYPE_CAT, '' as TYPE_SCHEM, '' as TYPE_NAME, '' AS SELF_REFERENCING_COL_NAME, '' AS REF_GENERATION FROM pg_catalog.pg_namespace n, pg_catalog.pg_class c LEFT JOIN pg_catalog.pg_description d ON (c.oid = d.objoid AND d.objsubid = 0) LEFT JOIN pg_catalog.pg_class dc ON (d.classoid = dc.oid AND dc.relname = 'pg_class') LEFT JOIN pg_catalog.pg_namespace dn ON (dn.oid = dc.relnamespace AND dn.nspname = 'pg_catalog') WHERE c.relnamespace = n.oid AND c.relname LIKE 'PROBABLYNOT' AND (false OR (c.relkind = 'r' AND n.nspname !~ '^pg_' AND n.nspname <> 'information_schema')) ORDER BY TABLE_TYPE, TABLE_SCHEM, TABLE_NAME;
Date: 2026-01-21 06:11:50 Duration: 0ms Database: postgres
14 53ms 4 1ms 49ms 13ms select distinct replace(recognitionengine, ' ', '_') || '%' || replace(datetimeupdate::text, ' ', '_') as t from latest_candle_datetime_per_receng order by t;Times Reported Time consuming prepare #14
Day Hour Count Duration Avg duration 06 4 53ms 13ms -
select distinct replace(recognitionengine, ' ', '_') || '%' || replace(datetimeupdate::text, ' ', '_') as t from latest_candle_datetime_per_receng order by t;
Date: 2026-01-21 06:40:10 Duration: 49ms Database: postgres
-
select distinct replace(recognitionengine, ' ', '_') || '%' || replace(datetimeupdate::text, ' ', '_') as t from latest_candle_datetime_per_receng order by t;
Date: 2026-01-21 06:40:10 Duration: 1ms Database: postgres
-
select distinct replace(recognitionengine, ' ', '_') || '%' || replace(datetimeupdate::text, ' ', '_') as t from latest_candle_datetime_per_receng order by t;
Date: 2026-01-21 06:10:12 Duration: 1ms Database: postgres
15 44ms 23 0ms 10ms 1ms WITH last_candle AS ( ;Times Reported Time consuming prepare #15
Day Hour Count Duration Avg duration 06 23 44ms 1ms -
WITH last_candle AS ( ;
Date: 2026-01-21 06:16:00 Duration: 10ms Database: postgres
-
WITH last_candle AS ( ;
Date: 2026-01-21 06:28:00 Duration: 8ms Database: postgres
-
WITH last_candle AS ( ;
Date: 2026-01-21 06:32:00 Duration: 8ms Database: postgres
16 23ms 1,789 0ms 0ms 0ms SET application_name = 'PostgreSQL JDBC Driver';Times Reported Time consuming prepare #16
Day Hour Count Duration Avg duration 06 1,789 23ms 0ms -
SET application_name = 'PostgreSQL JDBC Driver';
Date: 2026-01-21 06:31:38 Duration: 0ms Database: postgres
-
SET application_name = 'PostgreSQL JDBC Driver';
Date: 2026-01-21 06:16:30 Duration: 0ms Database: postgres
-
SET application_name = 'PostgreSQL JDBC Driver';
Date: 2026-01-21 06:21:21 Duration: 0ms Database: postgres
17 22ms 8 0ms 8ms 2ms WITH pre_symbols AS ( /* find relevant symbols */ ;Times Reported Time consuming prepare #17
Day Hour Count Duration Avg duration 06 8 22ms 2ms -
WITH pre_symbols AS ( /* find relevant symbols */ ;
Date: 2026-01-21 06:11:48 Duration: 8ms Database: postgres
-
WITH pre_symbols AS ( /* find relevant symbols */ ;
Date: 2026-01-21 06:11:48 Duration: 5ms Database: postgres
-
WITH pre_symbols AS ( /* find relevant symbols */ ;
Date: 2026-01-21 06:11:49 Duration: 3ms Database: postgres
18 16ms 12 0ms 2ms 1ms select distinct classname, to_char(created_datetime, 'yyyy-mm-dd HH24:MI'), to_char(cleared_datetime, 'yyyy-mm-dd HH24:MI'), action_to_take, description, created_datetime from datafeed_restarter_events where (is_current_entry = 1 OR cleared_datetime > current_timestamp - interval '17 hour') order by created_datetime desc;Times Reported Time consuming prepare #18
Day Hour Count Duration Avg duration 06 12 16ms 1ms -
select distinct classname, to_char(created_datetime, 'yyyy-mm-dd HH24:MI'), to_char(cleared_datetime, 'yyyy-mm-dd HH24:MI'), action_to_take, description, created_datetime from datafeed_restarter_events where (is_current_entry = 1 OR cleared_datetime > current_timestamp - interval '17 hour') order by created_datetime desc;
Date: 2026-01-21 06:50:15 Duration: 2ms Database: postgres
-
select distinct classname, to_char(created_datetime, 'yyyy-mm-dd HH24:MI'), to_char(cleared_datetime, 'yyyy-mm-dd HH24:MI'), action_to_take, description, created_datetime from datafeed_restarter_events where (is_current_entry = 1 OR cleared_datetime > current_timestamp - interval '17 hour') order by created_datetime desc;
Date: 2026-01-21 06:12:03 Duration: 2ms Database: postgres
-
select distinct classname, to_char(created_datetime, 'yyyy-mm-dd HH24:MI'), to_char(cleared_datetime, 'yyyy-mm-dd HH24:MI'), action_to_take, description, created_datetime from datafeed_restarter_events where (is_current_entry = 1 OR cleared_datetime > current_timestamp - interval '17 hour') order by created_datetime desc;
Date: 2026-01-21 06:17:04 Duration: 1ms Database: postgres
19 16ms 6 2ms 3ms 2ms select client_addr, count(1) from pg_stat_activity, pg_settings WHERE name = 'max_connections' group by client_addr, setting having (client_addr is not null OR (client_addr is null and count(1) > (cast(setting as numeric) / 3 * 2))) order by count desc;Times Reported Time consuming prepare #19
Day Hour Count Duration Avg duration 06 6 16ms 2ms -
select client_addr, count(1) from pg_stat_activity, pg_settings WHERE name = 'max_connections' group by client_addr, setting having (client_addr is not null OR (client_addr is null and count(1) > (cast(setting as numeric) / 3 * 2))) order by count desc;
Date: 2026-01-21 06:10:05 Duration: 3ms Database: postgres
-
select client_addr, count(1) from pg_stat_activity, pg_settings WHERE name = 'max_connections' group by client_addr, setting having (client_addr is not null OR (client_addr is null and count(1) > (cast(setting as numeric) / 3 * 2))) order by count desc;
Date: 2026-01-21 06:20:05 Duration: 3ms Database: postgres
-
select client_addr, count(1) from pg_stat_activity, pg_settings WHERE name = 'max_connections' group by client_addr, setting having (client_addr is not null OR (client_addr is null and count(1) > (cast(setting as numeric) / 3 * 2))) order by count desc;
Date: 2026-01-21 06:00:05 Duration: 2ms Database: postgres
20 16ms 6 2ms 3ms 2ms with rankedmt4 as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors ), last_feed_entry as ( select * from rankedmt4 where r = 1 ), ok_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where status = 'OK' ), earliest_entry_after_ok as ( select m.datafeedname, min(m.eventtimestamp) as eventtimestamp from mt4datafeederrors m left outer join ( select datafeedname, eventtimestamp from ok_entries where r = 1) oo on m.datafeedname = oo.datafeedname where m.eventtimestamp > coalesce(oo.eventtimestamp, '1900-01-01'::timestamp without time zone) group by m.datafeedname ), notified_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where notified is not null and notified <> '' ), broker as ( select *, row_number() over (partition by feedname order by brokerid) r from ( select distinct b.brokerid, b.name as brokername, dss.classname as feedname from downloadersymbolsettings dss inner join brokersymbollist bsl on dss.symbolid = bsl.symbolid inner join broker b on bsl.brokerid = b.brokerid where dss.enabled = 1) a ) select last.id, last.datafeedname, last.eventtimestamp, last.status, last.errordescription, last.serveraddress, last.username, note.notified, note.eventtimestamp, broker.brokername from last_feed_entry last inner join earliest_entry_after_ok after_ok on last.datafeedname = after_ok.datafeedname inner join broker on last.datafeedname = broker.feedname left outer join ok_entries ok on ok.datafeedname = last.datafeedname left outer join notified_entries note on note.datafeedname = last.datafeedname and note.r = 1 where (ok.r is null or ok.r = 1) and last.datafeedname not in ( select distinct datafeedname from last_feed_entry where status = 'OK') and extract(epoch from (last.eventtimestamp - after_ok.eventtimestamp)) > 60 * 60 and last.eventtimestamp > current_timestamp - interval '1 day' and (note.eventtimestamp is null or note.eventtimestamp < current_timestamp - interval '10 hours') and last.eventtimestamp > current_timestamp - interval '1 hour' and broker.r = 1;Times Reported Time consuming prepare #20
Day Hour Count Duration Avg duration 06 6 16ms 2ms -
with rankedmt4 as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors ), last_feed_entry as ( select * from rankedmt4 where r = 1 ), ok_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where status = 'OK' ), earliest_entry_after_ok as ( select m.datafeedname, min(m.eventtimestamp) as eventtimestamp from mt4datafeederrors m left outer join ( select datafeedname, eventtimestamp from ok_entries where r = 1) oo on m.datafeedname = oo.datafeedname where m.eventtimestamp > coalesce(oo.eventtimestamp, '1900-01-01'::timestamp without time zone) group by m.datafeedname ), notified_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where notified is not null and notified <> '' ), broker as ( select *, row_number() over (partition by feedname order by brokerid) r from ( select distinct b.brokerid, b.name as brokername, dss.classname as feedname from downloadersymbolsettings dss inner join brokersymbollist bsl on dss.symbolid = bsl.symbolid inner join broker b on bsl.brokerid = b.brokerid where dss.enabled = 1) a ) select last.id, last.datafeedname, last.eventtimestamp, last.status, last.errordescription, last.serveraddress, last.username, note.notified, note.eventtimestamp, broker.brokername from last_feed_entry last inner join earliest_entry_after_ok after_ok on last.datafeedname = after_ok.datafeedname inner join broker on last.datafeedname = broker.feedname left outer join ok_entries ok on ok.datafeedname = last.datafeedname left outer join notified_entries note on note.datafeedname = last.datafeedname and note.r = 1 where (ok.r is null or ok.r = 1) and last.datafeedname not in ( select distinct datafeedname from last_feed_entry where status = 'OK') and extract(epoch from (last.eventtimestamp - after_ok.eventtimestamp)) > 60 * 60 and last.eventtimestamp > current_timestamp - interval '1 day' and (note.eventtimestamp is null or note.eventtimestamp < current_timestamp - interval '10 hours') and last.eventtimestamp > current_timestamp - interval '1 hour' and broker.r = 1;
Date: 2026-01-21 06:20:02 Duration: 3ms Database: postgres
-
with rankedmt4 as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors ), last_feed_entry as ( select * from rankedmt4 where r = 1 ), ok_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where status = 'OK' ), earliest_entry_after_ok as ( select m.datafeedname, min(m.eventtimestamp) as eventtimestamp from mt4datafeederrors m left outer join ( select datafeedname, eventtimestamp from ok_entries where r = 1) oo on m.datafeedname = oo.datafeedname where m.eventtimestamp > coalesce(oo.eventtimestamp, '1900-01-01'::timestamp without time zone) group by m.datafeedname ), notified_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where notified is not null and notified <> '' ), broker as ( select *, row_number() over (partition by feedname order by brokerid) r from ( select distinct b.brokerid, b.name as brokername, dss.classname as feedname from downloadersymbolsettings dss inner join brokersymbollist bsl on dss.symbolid = bsl.symbolid inner join broker b on bsl.brokerid = b.brokerid where dss.enabled = 1) a ) select last.id, last.datafeedname, last.eventtimestamp, last.status, last.errordescription, last.serveraddress, last.username, note.notified, note.eventtimestamp, broker.brokername from last_feed_entry last inner join earliest_entry_after_ok after_ok on last.datafeedname = after_ok.datafeedname inner join broker on last.datafeedname = broker.feedname left outer join ok_entries ok on ok.datafeedname = last.datafeedname left outer join notified_entries note on note.datafeedname = last.datafeedname and note.r = 1 where (ok.r is null or ok.r = 1) and last.datafeedname not in ( select distinct datafeedname from last_feed_entry where status = 'OK') and extract(epoch from (last.eventtimestamp - after_ok.eventtimestamp)) > 60 * 60 and last.eventtimestamp > current_timestamp - interval '1 day' and (note.eventtimestamp is null or note.eventtimestamp < current_timestamp - interval '10 hours') and last.eventtimestamp > current_timestamp - interval '1 hour' and broker.r = 1;
Date: 2026-01-21 06:10:02 Duration: 3ms Database: postgres
-
with rankedmt4 as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors ), last_feed_entry as ( select * from rankedmt4 where r = 1 ), ok_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where status = 'OK' ), earliest_entry_after_ok as ( select m.datafeedname, min(m.eventtimestamp) as eventtimestamp from mt4datafeederrors m left outer join ( select datafeedname, eventtimestamp from ok_entries where r = 1) oo on m.datafeedname = oo.datafeedname where m.eventtimestamp > coalesce(oo.eventtimestamp, '1900-01-01'::timestamp without time zone) group by m.datafeedname ), notified_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where notified is not null and notified <> '' ), broker as ( select *, row_number() over (partition by feedname order by brokerid) r from ( select distinct b.brokerid, b.name as brokername, dss.classname as feedname from downloadersymbolsettings dss inner join brokersymbollist bsl on dss.symbolid = bsl.symbolid inner join broker b on bsl.brokerid = b.brokerid where dss.enabled = 1) a ) select last.id, last.datafeedname, last.eventtimestamp, last.status, last.errordescription, last.serveraddress, last.username, note.notified, note.eventtimestamp, broker.brokername from last_feed_entry last inner join earliest_entry_after_ok after_ok on last.datafeedname = after_ok.datafeedname inner join broker on last.datafeedname = broker.feedname left outer join ok_entries ok on ok.datafeedname = last.datafeedname left outer join notified_entries note on note.datafeedname = last.datafeedname and note.r = 1 where (ok.r is null or ok.r = 1) and last.datafeedname not in ( select distinct datafeedname from last_feed_entry where status = 'OK') and extract(epoch from (last.eventtimestamp - after_ok.eventtimestamp)) > 60 * 60 and last.eventtimestamp > current_timestamp - interval '1 day' and (note.eventtimestamp is null or note.eventtimestamp < current_timestamp - interval '10 hours') and last.eventtimestamp > current_timestamp - interval '1 hour' and broker.r = 1;
Date: 2026-01-21 06:40:03 Duration: 2ms Database: postgres
Time consuming bind
Rank Total duration Times executed Min duration Max duration Avg duration Query 1 40s654ms 3,695 0ms 68ms 11ms WITH rar_max as ( ;Times Reported Time consuming bind #1
Day Hour Count Duration Avg duration Jan 21 06 3,695 40s654ms 11ms -
WITH rar_max as ( ;
Date: 2026-01-21 06:21:12 Duration: 68ms Database: postgres parameters: $1 = '627', $2 = '7', $3 = '15', $4 = '30', $5 = '60', $6 = '120', $7 = '240', $8 = '480', $9 = '1440', $10 = '0', $11 = '', $12 = '213', $13 = '#ADBE', $14 = '#ALVG', $15 = '#AMZN', $16 = '#APPL', $17 = '#BA', $18 = '#BABA', $19 = '#BAYGn', $20 = '#BMWG', $21 = '#BNPP', $22 = '#CAT', $23 = '#CBKG', $24 = '#DAIGn', $25 = '#DIS', $26 = '#EA', $27 = '#FB', $28 = '#FDX', $29 = '#GE', $30 = '#GM', $31 = '#GOOGL', $32 = '#GS', $33 = '#INTC', $34 = '#JPM', $35 = '#KO', $36 = '#META', $37 = '#MSFT', $38 = '#NFLX', $39 = '#TSLA', $40 = '#VOWG', $41 = '#WMT', $42 = '#XOM', $43 = 'AUDCAD', $44 = 'AUDCHF', $45 = 'AUDJPY', $46 = 'AUDNZD', $47 = 'AUDUSD', $48 = 'AUS_200', $49 = 'BTCEUR', $50 = 'BTCGBP', $51 = 'BTCUSD', $52 = 'CADCHF', $53 = 'CADJPY', $54 = 'CHFJPY', $55 = 'CL_BRENT', $56 = 'DASHUSD', $57 = 'EOSUSD', $58 = 'ESP_35', $59 = 'ETHEUR', $60 = 'ETHGBP', $61 = 'ETHUSD', $62 = 'EURAUD', $63 = 'EURCAD', $64 = 'EURCHF', $65 = 'EURGBP', $66 = 'EURJPY', $67 = 'EURMXN', $68 = 'EURNOK', $69 = 'EURNZD', $70 = 'EURPLN', $71 = 'EURSEK', $72 = 'EURTRY', $73 = 'EURUSD', $74 = 'EUR_50', $75 = 'FRA_40', $76 = 'GBPAUD', $77 = 'GBPCAD', $78 = 'GBPCHF', $79 = 'GBPJPY', $80 = 'GBPNZD', $81 = 'GBPUSD', $82 = 'GBPZAR', $83 = 'GBR_100', $84 = 'HKDJPY', $85 = 'HKG_50', $86 = 'IOTAUSD', $87 = 'LTCEUR', $88 = 'LTCUSD', $89 = 'NAS100', $90 = 'NEOUSD', $91 = 'NOKJPY', $92 = 'NZDCAD', $93 = 'NZDCHF', $94 = 'NZDJPY', $95 = 'NZDUSD', $96 = 'OMGUSD', $97 = 'SPX500', $98 = 'TRXUSD', $99 = 'US30', $100 = 'USDCAD', $101 = 'USDCHF', $102 = 'USDCNH', $103 = 'USDDKK', $104 = 'USDJPY', $105 = 'USDMXN', $106 = 'USDNOK', $107 = 'USDPLN', $108 = 'USDSEK', $109 = 'USDSGD', $110 = 'USDZAR', $111 = 'USOIL', $112 = 'XAGUSD', $113 = 'XAUEUR', $114 = 'XAUUSD', $115 = 'XMRUSD', $116 = 'XPTUSD', $117 = 'XRPUSD', $118 = 'ZARJPY', $119 = 'ZECUSD', $120 = 'AUDCAD', $121 = 'AUDCHF', $122 = 'AUDJPY', $123 = 'AUDNZD', $124 = 'AUDUSD', $125 = 'CADCHF', $126 = 'CADJPY', $127 = 'CHFJPY', $128 = 'EURAUD', $129 = 'EURCAD', $130 = 'EURCHF', $131 = 'EURGBP', $132 = 'EURJPY', $133 = 'EURMXN', $134 = 'EURNOK', $135 = 'EURNZD', $136 = 'EURPLN', $137 = 'EURSEK', $138 = 'EURTRY', $139 = 'EURUSD', $140 = 'GBPAUD', $141 = 'GBPCAD', $142 = 'GBPCHF', $143 = 'GBPJPY', $144 = 'GBPNZD', $145 = 'GBPUSD', $146 = 'GBPZAR', $147 = 'HKDJPY', $148 = 'NOKJPY', $149 = 'NZDCAD', $150 = 'NZDCHF', $151 = 'NZDJPY', $152 = 'NZDUSD', $153 = 'USDCAD', $154 = 'USDCHF', $155 = 'USDCNH', $156 = 'USDDKK', $157 = 'USDJPY', $158 = 'USDMXN', $159 = 'USDNOK', $160 = 'USDPLN', $161 = 'USDSEK', $162 = 'USDSGD', $163 = 'USDZAR', $164 = 'ZARJPY', $165 = 'BTCEUR', $166 = 'BTCGBP', $167 = 'BTCUSD', $168 = 'DASHUSD', $169 = 'EOSUSD', $170 = 'ETHEUR', $171 = 'ETHGBP', $172 = 'ETHUSD', $173 = 'IOTAUSD', $174 = 'LTCEUR', $175 = 'LTCUSD', $176 = 'NEOUSD', $177 = 'OMGUSD', $178 = 'TRXUSD', $179 = 'XMRUSD', $180 = 'XRPUSD', $181 = 'ZECUSD', $182 = 'XAGUSD', $183 = 'XAUEUR', $184 = 'XAUUSD', $185 = 'XPTUSD', $186 = 'CL_BRENT', $187 = 'USOIL', $188 = '#ALVG', $189 = '#BAYGn', $190 = '#BMWG', $191 = '#BNPP', $192 = '#CBKG', $193 = '#DAIGn', $194 = '#VOWG', $195 = 'AUS_200', $196 = 'ESP_35', $197 = 'EUR_50', $198 = 'FRA_40', $199 = 'GBR_100', $200 = 'HKG_50', $201 = 'NAS100', $202 = 'SPX500', $203 = 'US30', $204 = '#ADBE', $205 = '#AMZN', $206 = '#APPL', $207 = '#BA', $208 = '#BABA', $209 = '#CAT', $210 = '#DIS', $211 = '#EA', $212 = '#FB', $213 = '#FDX', $214 = '#GE', $215 = '#GM', $216 = '#GOOGL', $217 = '#GS', $218 = '#INTC', $219 = '#JPM', $220 = '#KO', $221 = '#MSFT', $222 = '#NFLX', $223 = '#TSLA', $224 = '#WMT', $225 = '#XOM', $226 = '700', $227 = '700', $228 = 't', $229 = '10', $230 = '10'
-
WITH rar_max as ( ;
Date: 2026-01-21 06:06:36 Duration: 59ms Database: postgres parameters: $1 = '538', $2 = '7', $3 = '15', $4 = '30', $5 = '60', $6 = '120', $7 = '240', $8 = '480', $9 = '1440', $10 = '0', $11 = '', $12 = '125', $13 = 'ANGLO', $14 = 'BARC', $15 = 'BAY', $16 = 'BPLON', $17 = 'HSBCL', $18 = 'LLOY', $19 = 'RIO', $20 = 'RollsRoyce', $21 = 'TESCO', $22 = 'VOD', $23 = 'AUDCAD', $24 = 'AUDCHF', $25 = 'AUDJPY', $26 = 'AUDNZD', $27 = 'AUDUSD', $28 = 'CADCHF', $29 = 'CADJPY', $30 = 'CHFJPY', $31 = 'EURAUD', $32 = 'EURCAD', $33 = 'EURCHF', $34 = 'EURDKK', $35 = 'EURGBP', $36 = 'EURHUF', $37 = 'EURJPY', $38 = 'EURNOK', $39 = 'EURNZD', $40 = 'EURPLN', $41 = 'EURUSD', $42 = 'GBPAUD', $43 = 'GBPCAD', $44 = 'GBPCHF', $45 = 'GBPJPY', $46 = 'GBPNZD', $47 = 'GBPUSD', $48 = 'GBPZAR', $49 = 'NZDCAD', $50 = 'NZDCHF', $51 = 'NZDJPY', $52 = 'NZDUSD', $53 = 'USDCAD', $54 = 'USDCHF', $55 = 'USDCNH', $56 = 'USDCZK', $57 = 'USDDKK', $58 = 'USDHKD', $59 = 'USDHUF', $60 = 'USDJPY', $61 = 'USDMXN', $62 = 'USDNOK', $63 = 'USDPLN', $64 = 'USDSEK', $65 = 'USDSGD', $66 = 'USDTRY', $67 = 'USDZAR', $68 = 'XAGEUR', $69 = 'XAGUSD', $70 = 'XAUEUR', $71 = 'XAUUSD', $72 = 'ZARJPY', $73 = 'Cocoa', $74 = 'Coffee', $75 = 'Copper', $76 = 'Palladium', $77 = 'Platinum', $78 = 'Sugar', $79 = 'UKOIL', $80 = 'USOIL', $81 = 'AUS200', $82 = 'FRA40', $83 = 'JPN225', $84 = 'NETH25', $85 = 'SPA35', $86 = 'SUI20', $87 = 'UK100', $88 = 'USA100', $89 = 'USA30', $90 = 'USDIndex', $91 = 'ALCOA', $92 = 'ALIBABA', $93 = 'AMAZON', $94 = 'AMEX', $95 = 'APPLE', $96 = 'BBVA', $97 = 'BOA', $98 = 'BOEING', $99 = 'CHEVRON', $100 = 'CISCO', $101 = 'COKE', $102 = 'EBAY', $103 = 'GE', $104 = 'GOOGLE', $105 = 'GS', $106 = 'HLT', $107 = 'IBM', $108 = 'ILMN', $109 = 'INTEL', $110 = 'Iberdrola', $111 = 'MCARD', $112 = 'MCDON', $113 = 'META', $114 = 'MSFT', $115 = 'Mapfre', $116 = 'Netflix', $117 = 'PFIZER', $118 = 'QCOM', $119 = 'Santander', $120 = 'TEVA', $121 = 'Telefonica', $122 = 'Tesla', $123 = 'AUDUSD', $124 = 'EURGBP', $125 = 'EURUSD', $126 = 'GBPUSD', $127 = 'NZDUSD', $128 = 'USDCAD', $129 = 'USDCHF', $130 = 'USDJPY', $131 = 'Adidas', $132 = 'Bayer', $133 = 'Daimler', $134 = 'Danone', $135 = 'LVMH', $136 = 'Lufthansa', $137 = 'Volksw', $138 = '700', $139 = '700', $140 = 't', $141 = '10', $142 = '10'
-
WITH rar_max as ( ;
Date: 2026-01-21 06:40:46 Duration: 56ms Database: postgres parameters: $1 = '489', $2 = '7', $3 = '15', $4 = '30', $5 = '60', $6 = '120', $7 = '240', $8 = '480', $9 = '1440', $10 = '0', $11 = '', $12 = '37', $13 = 'AUDCAD', $14 = 'AUDCHF', $15 = 'AUDJPY', $16 = 'AUDNZD', $17 = 'AUDUSD', $18 = 'CADJPY', $19 = 'CHFJPY', $20 = 'EURAUD', $21 = 'EURCAD', $22 = 'EURCHF', $23 = 'EURGBP', $24 = 'EURJPY', $25 = 'EURNZD', $26 = 'EURUSD', $27 = 'GBPAUD', $28 = 'GBPCAD', $29 = 'GBPCHF', $30 = 'GBPJPY', $31 = 'GBPNZD', $32 = 'GBPUSD', $33 = 'NZDCAD', $34 = 'NZDJPY', $35 = 'NZDUSD', $36 = 'USDCAD', $37 = 'USDCHF', $38 = 'USDJPY', $39 = 'DOW30', $40 = 'HK-HSI.', $41 = 'NASDAQ100', $42 = 'RK_HSI', $43 = 'RK_SSI', $44 = 'SP500', $45 = 'XAGUSD', $46 = 'XAUUSD', $47 = 'OIL', $48 = 'R_HSI', $49 = 'R_SSI', $50 = '700', $51 = '700', $52 = 't', $53 = '10', $54 = '10'
2 6s521ms 14,837 0ms 22ms 0ms SELECT ;Times Reported Time consuming bind #2
Day Hour Count Duration Avg duration 06 14,837 6s521ms 0ms -
SELECT ;
Date: 2026-01-21 06:15:03 Duration: 22ms Database: postgres parameters: $1 = '515840233911395300'
-
SELECT ;
Date: 2026-01-21 06:58:02 Duration: 20ms Database: postgres parameters: $1 = '558', $2 = '558', $3 = '515840243211281300'
-
SELECT ;
Date: 2026-01-21 06:58:02 Duration: 13ms Database: postgres parameters: $1 = '558', $2 = '558', $3 = '515840243280578300'
3 2s619ms 1,085 1ms 19ms 2ms SELECT symbolid, ;Times Reported Time consuming bind #3
Day Hour Count Duration Avg duration 06 1,085 2s619ms 2ms -
SELECT symbolid, ;
Date: 2026-01-21 06:02:02 Duration: 19ms Database: postgres parameters: $1 = 'GLOBALGTMT5', $2 = '15', $3 = 'GBPJPY', $4 = 'LTCUSD', $5 = 'LTCJPY', $6 = 'JP225', $7 = 'GTi12', $8 = 'GBPUSD'
-
SELECT symbolid, ;
Date: 2026-01-21 06:16:52 Duration: 13ms Database: postgres parameters: $1 = 'MILLENNIUMPF', $2 = '15', $3 = 'GBPUSD.ID', $4 = 'GBPNZD.FX', $5 = 'GBPUSD.FX', $6 = 'GBPUSD'
-
SELECT symbolid, ;
Date: 2026-01-21 06:31:38 Duration: 7ms Database: postgres parameters: $1 = 'ATFX', $2 = '15', $3 = 'SPX500', $4 = 'US30'
4 1s55ms 575 1ms 3ms 1ms SELECT s.symbolid, dss.downloadfrequency, dss.downloadersymbol;Times Reported Time consuming bind #4
Day Hour Count Duration Avg duration 06 575 1s55ms 1ms -
SELECT s.symbolid, dss.downloadfrequency, dss.downloadersymbol;
Date: 2026-01-21 06:31:02 Duration: 3ms Database: postgres parameters: $1 = 'AXIORY'
-
SELECT s.symbolid, dss.downloadfrequency, dss.downloadersymbol;
Date: 2026-01-21 06:31:21 Duration: 2ms Database: postgres parameters: $1 = 'Alpari'
-
SELECT s.symbolid, dss.downloadfrequency, dss.downloadersymbol;
Date: 2026-01-21 06:16:42 Duration: 2ms Database: postgres parameters: $1 = 'AXIORY'
5 828ms 16 28ms 110ms 51ms with sym_info as ( ;Times Reported Time consuming bind #5
Day Hour Count Duration Avg duration 06 16 828ms 51ms -
with sym_info as ( ;
Date: 2026-01-21 06:06:48 Duration: 110ms Database: postgres parameters: $1 = '627', $2 = 'Forex', $3 = 'Forex', $4 = '627', $5 = 'Forex', $6 = '627', $7 = '627', $8 = 'Forex', $9 = '627'
-
with sym_info as ( ;
Date: 2026-01-21 06:06:40 Duration: 90ms Database: postgres parameters: $1 = '620', $2 = 'Forex', $3 = 'Forex', $4 = '620', $5 = 'Forex', $6 = '620', $7 = '620', $8 = 'Forex', $9 = '620'
-
with sym_info as ( ;
Date: 2026-01-21 06:06:43 Duration: 84ms Database: postgres parameters: $1 = '617', $2 = 'Forex', $3 = 'Forex', $4 = '617', $5 = 'Forex', $6 = '617', $7 = '617', $8 = 'Forex', $9 = '617'
6 709ms 90 4ms 22ms 7ms WITH last_candle AS ( ;Times Reported Time consuming bind #6
Day Hour Count Duration Avg duration 06 90 709ms 7ms -
WITH last_candle AS ( ;
Date: 2026-01-21 06:36:00 Duration: 22ms Database: postgres parameters: $1 = '558', $2 = '558'
-
WITH last_candle AS ( ;
Date: 2026-01-21 06:20:01 Duration: 17ms Database: postgres parameters: $1 = '958', $2 = '958'
-
WITH last_candle AS ( ;
Date: 2026-01-21 06:28:01 Duration: 16ms Database: postgres parameters: $1 = '958', $2 = '958'
7 382ms 19 0ms 39ms 20ms with wh_patitioned as ( ;Times Reported Time consuming bind #7
Day Hour Count Duration Avg duration 06 19 382ms 20ms -
with wh_patitioned as ( ;
Date: 2026-01-21 06:55:01 Duration: 39ms Database: postgres parameters: $1 = '558', $2 = '558', $3 = '558', $4 = '558', $5 = '558', $6 = '558', $7 = '558', $8 = '558', $9 = '558'
-
with wh_patitioned as ( ;
Date: 2026-01-21 06:09:42 Duration: 38ms Database: postgres parameters: $1 = '558', $2 = '558', $3 = '558', $4 = '558', $5 = '558', $6 = '558', $7 = '558', $8 = '558', $9 = '558'
-
with wh_patitioned as ( ;
Date: 2026-01-21 06:01:02 Duration: 32ms Database: postgres parameters: $1 = '558', $2 = '558', $3 = '558', $4 = '558', $5 = '558', $6 = '558', $7 = '558', $8 = '558', $9 = '558'
8 353ms 71 0ms 18ms 4ms WITH /*Latest.JapSticks*/ all_results AS ( SELECT ;Times Reported Time consuming bind #8
Day Hour Count Duration Avg duration 06 71 353ms 4ms -
WITH /*Latest.JapSticks*/ all_results AS ( SELECT ;
Date: 2026-01-21 06:45:50 Duration: 18ms Database: postgres parameters: $1 = '667', $2 = '0', $3 = '0', $4 = '0', $5 = '', $6 = '0', $7 = '', $8 = '0', $9 = '', $10 = '0', $11 = '0'
-
WITH /*Latest.JapSticks*/ all_results AS ( SELECT ;
Date: 2026-01-21 06:06:50 Duration: 18ms Database: postgres parameters: $1 = '667', $2 = '0', $3 = '0', $4 = '0', $5 = '', $6 = '0', $7 = '', $8 = '0', $9 = '', $10 = '0', $11 = '0'
-
WITH /*Latest.JapSticks*/ all_results AS ( SELECT ;
Date: 2026-01-21 06:34:15 Duration: 18ms Database: postgres parameters: $1 = '689', $2 = '0', $3 = '0', $4 = '0', $5 = '', $6 = '0', $7 = '', $8 = '0', $9 = '', $10 = '0', $11 = '0'
9 325ms 13,713 0ms 2ms 0ms select 1;Times Reported Time consuming bind #9
Day Hour Count Duration Avg duration 06 13,713 325ms 0ms -
select 1;
Date: 2026-01-21 06:24:02 Duration: 2ms Database: postgres
-
select 1;
Date: 2026-01-21 06:21:10 Duration: 1ms Database: postgres
-
select 1;
Date: 2026-01-21 06:20:56 Duration: 1ms Database: postgres
10 270ms 5,640 0ms 0ms 0ms INSERT INTO T15 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;Times Reported Time consuming bind #10
Day Hour Count Duration Avg duration 06 5,640 270ms 0ms -
INSERT INTO T15 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;
Date: 2026-01-21 06:00:48 Duration: 0ms Database: postgres parameters: $1 = '2026-01-21 05:30:00', $2 = '2.00897', $3 = '2.009505', $4 = '2.0089', $5 = '2.009265', $6 = '543', $7 = '515840230400034300', $8 = '0', $9 = '2026-01-21 06:00:48.365', $10 = '2026-01-21 06:00:48.09', $11 = '2.00897', $12 = '2.009505', $13 = '2.0089', $14 = '2.009265', $15 = '543', $16 = '0', $17 = '2026-01-21 06:00:48.365', $18 = '2026-01-21 06:00:48.09'
-
INSERT INTO T15 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;
Date: 2026-01-21 06:11:53 Duration: 0ms Database: postgres parameters: $1 = '2026-01-21 05:45:00', $2 = '48574.05', $3 = '48580.05', $4 = '48559.55', $5 = '48579.35', $6 = '2515', $7 = '515840248000537300', $8 = '0', $9 = '2026-01-21 06:11:53.418', $10 = '2026-01-21 06:11:53.34', $11 = '48574.05', $12 = '48580.05', $13 = '48559.55', $14 = '48579.35', $15 = '2515', $16 = '0', $17 = '2026-01-21 06:11:53.418', $18 = '2026-01-21 06:11:53.34'
-
INSERT INTO T15 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;
Date: 2026-01-21 06:47:41 Duration: 0ms Database: postgres parameters: $1 = '2026-01-21 06:30:00', $2 = '199.768', $3 = '199.874', $4 = '199.756', $5 = '199.862', $6 = '572', $7 = '515840230445516300', $8 = '0', $9 = '2026-01-21 06:47:41.782', $10 = '2026-01-21 06:47:41.699', $11 = '199.768', $12 = '199.874', $13 = '199.756', $14 = '199.862', $15 = '572', $16 = '0', $17 = '2026-01-21 06:47:41.782', $18 = '2026-01-21 06:47:41.699'
11 254ms 3,226 0ms 0ms 0ms INSERT INTO T30 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;Times Reported Time consuming bind #11
Day Hour Count Duration Avg duration 06 3,226 254ms 0ms -
INSERT INTO T30 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;
Date: 2026-01-21 06:40:53 Duration: 0ms Database: postgres parameters: $1 = '2026-01-21 05:30:00', $2 = '48583.15', $3 = '48590.55', $4 = '48559.55', $5 = '48579.35', $6 = '5172', $7 = '515840248000726300', $8 = '0', $9 = '2026-01-21 06:40:53.425', $10 = '2026-01-21 06:40:53.325', $11 = '48583.15', $12 = '48590.55', $13 = '48559.55', $14 = '48579.35', $15 = '5172', $16 = '0', $17 = '2026-01-21 06:40:53.425', $18 = '2026-01-21 06:40:53.325'
-
INSERT INTO T30 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;
Date: 2026-01-21 06:11:53 Duration: 0ms Database: postgres parameters: $1 = '2026-01-21 05:30:00', $2 = '48583.15', $3 = '48590.55', $4 = '48559.55', $5 = '48579.35', $6 = '5172', $7 = '515840248000726300', $8 = '0', $9 = '2026-01-21 06:11:53.454', $10 = '2026-01-21 06:11:53.356', $11 = '48583.15', $12 = '48590.55', $13 = '48559.55', $14 = '48579.35', $15 = '5172', $16 = '0', $17 = '2026-01-21 06:11:53.454', $18 = '2026-01-21 06:11:53.356'
-
INSERT INTO T30 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;
Date: 2026-01-21 06:41:51 Duration: 0ms Database: postgres parameters: $1 = '2026-01-21 06:00:00', $2 = '25055.28', $3 = '25058.78', $4 = '25028.4', $5 = '25049.4', $6 = '6497', $7 = '515840248039147300', $8 = '0', $9 = '2026-01-21 06:41:51.885', $10 = '2026-01-21 06:41:51.805', $11 = '25055.28', $12 = '25058.78', $13 = '25028.4', $14 = '25049.4', $15 = '6497', $16 = '0', $17 = '2026-01-21 06:41:51.885', $18 = '2026-01-21 06:41:51.805'
12 185ms 2,114 0ms 0ms 0ms INSERT INTO T60 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;Times Reported Time consuming bind #12
Day Hour Count Duration Avg duration 06 2,114 185ms 0ms -
INSERT INTO T60 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;
Date: 2026-01-21 06:02:39 Duration: 0ms Database: postgres parameters: $1 = '2026-01-21 04:00:00', $2 = '8779.15', $3 = '8793.55', $4 = '8774.2', $5 = '8783.65', $6 = '5405', $7 = '515840248015562300', $8 = '0', $9 = '2026-01-21 06:02:39.848', $10 = '2026-01-21 06:02:39.739', $11 = '8779.15', $12 = '8793.55', $13 = '8774.2', $14 = '8783.65', $15 = '5405', $16 = '0', $17 = '2026-01-21 06:02:39.848', $18 = '2026-01-21 06:02:39.739'
-
INSERT INTO T60 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;
Date: 2026-01-21 06:00:55 Duration: 0ms Database: postgres parameters: $1 = '2026-01-21 05:00:00', $2 = '0.91267', $3 = '0.91285', $4 = '0.91242', $5 = '0.91272', $6 = '1828', $7 = '515840217506621300', $8 = '0', $9 = '2026-01-21 06:00:55.019', $10 = '2026-01-21 06:00:55.019', $11 = '0.91267', $12 = '0.91285', $13 = '0.91242', $14 = '0.91272', $15 = '1828', $16 = '0', $17 = '2026-01-21 06:00:55.019', $18 = '2026-01-21 06:00:55.019'
-
INSERT INTO T60 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;
Date: 2026-01-21 06:18:00 Duration: 0ms Database: postgres parameters: $1 = '2026-01-21 05:00:00', $2 = '8783.65', $3 = '8791.65', $4 = '8772.65', $5 = '8783.65', $6 = '659', $7 = '500991628266153200', $8 = '0', $9 = '2026-01-21 06:18:00.149', $10 = '2026-01-21 06:18:00.149', $11 = '8783.65', $12 = '8791.65', $13 = '8772.65', $14 = '8783.65', $15 = '659', $16 = '0', $17 = '2026-01-21 06:18:00.149', $18 = '2026-01-21 06:18:00.149'
13 164ms 278 0ms 2ms 0ms SELECT NULL AS TABLE_CAT, n.nspname AS TABLE_SCHEM, c.relname AS TABLE_NAME, CASE n.nspname ~ '^pg_' OR n.nspname = 'information_schema' WHEN true THEN CASE WHEN n.nspname = 'pg_catalog' OR n.nspname = 'information_schema' THEN CASE c.relkind WHEN 'r' THEN 'SYSTEM TABLE' WHEN 'v' THEN 'SYSTEM VIEW' WHEN 'i' THEN 'SYSTEM INDEX' ELSE NULL END WHEN n.nspname = 'pg_toast' THEN CASE c.relkind WHEN 'r' THEN 'SYSTEM TOAST TABLE' WHEN 'i' THEN 'SYSTEM TOAST INDEX' ELSE NULL END ELSE CASE c.relkind WHEN 'r' THEN 'TEMPORARY TABLE' WHEN 'p' THEN 'TEMPORARY TABLE' WHEN 'i' THEN 'TEMPORARY INDEX' WHEN 'S' THEN 'TEMPORARY SEQUENCE' WHEN 'v' THEN 'TEMPORARY VIEW' ELSE NULL END END WHEN false THEN CASE c.relkind WHEN 'r' THEN 'TABLE' WHEN 'p' THEN 'PARTITIONED TABLE' WHEN 'i' THEN 'INDEX' WHEN 'S' THEN 'SEQUENCE' WHEN 'v' THEN 'VIEW' WHEN 'c' THEN 'TYPE' WHEN 'f' THEN 'FOREIGN TABLE' WHEN 'm' THEN 'MATERIALIZED VIEW' ELSE NULL END ELSE NULL END AS TABLE_TYPE, d.description AS REMARKS, '' as TYPE_CAT, '' as TYPE_SCHEM, '' as TYPE_NAME, '' AS SELF_REFERENCING_COL_NAME, '' AS REF_GENERATION FROM pg_catalog.pg_namespace n, pg_catalog.pg_class c LEFT JOIN pg_catalog.pg_description d ON (c.oid = d.objoid AND d.objsubid = 0) LEFT JOIN pg_catalog.pg_class dc ON (d.classoid = dc.oid AND dc.relname = 'pg_class') LEFT JOIN pg_catalog.pg_namespace dn ON (dn.oid = dc.relnamespace AND dn.nspname = 'pg_catalog') WHERE c.relnamespace = n.oid AND c.relname LIKE 'PROBABLYNOT' AND (false OR (c.relkind = 'r' AND n.nspname !~ '^pg_' AND n.nspname <> 'information_schema')) ORDER BY TABLE_TYPE, TABLE_SCHEM, TABLE_NAME;Times Reported Time consuming bind #13
Day Hour Count Duration Avg duration 06 278 164ms 0ms -
SELECT NULL AS TABLE_CAT, n.nspname AS TABLE_SCHEM, c.relname AS TABLE_NAME, CASE n.nspname ~ '^pg_' OR n.nspname = 'information_schema' WHEN true THEN CASE WHEN n.nspname = 'pg_catalog' OR n.nspname = 'information_schema' THEN CASE c.relkind WHEN 'r' THEN 'SYSTEM TABLE' WHEN 'v' THEN 'SYSTEM VIEW' WHEN 'i' THEN 'SYSTEM INDEX' ELSE NULL END WHEN n.nspname = 'pg_toast' THEN CASE c.relkind WHEN 'r' THEN 'SYSTEM TOAST TABLE' WHEN 'i' THEN 'SYSTEM TOAST INDEX' ELSE NULL END ELSE CASE c.relkind WHEN 'r' THEN 'TEMPORARY TABLE' WHEN 'p' THEN 'TEMPORARY TABLE' WHEN 'i' THEN 'TEMPORARY INDEX' WHEN 'S' THEN 'TEMPORARY SEQUENCE' WHEN 'v' THEN 'TEMPORARY VIEW' ELSE NULL END END WHEN false THEN CASE c.relkind WHEN 'r' THEN 'TABLE' WHEN 'p' THEN 'PARTITIONED TABLE' WHEN 'i' THEN 'INDEX' WHEN 'S' THEN 'SEQUENCE' WHEN 'v' THEN 'VIEW' WHEN 'c' THEN 'TYPE' WHEN 'f' THEN 'FOREIGN TABLE' WHEN 'm' THEN 'MATERIALIZED VIEW' ELSE NULL END ELSE NULL END AS TABLE_TYPE, d.description AS REMARKS, '' as TYPE_CAT, '' as TYPE_SCHEM, '' as TYPE_NAME, '' AS SELF_REFERENCING_COL_NAME, '' AS REF_GENERATION FROM pg_catalog.pg_namespace n, pg_catalog.pg_class c LEFT JOIN pg_catalog.pg_description d ON (c.oid = d.objoid AND d.objsubid = 0) LEFT JOIN pg_catalog.pg_class dc ON (d.classoid = dc.oid AND dc.relname = 'pg_class') LEFT JOIN pg_catalog.pg_namespace dn ON (dn.oid = dc.relnamespace AND dn.nspname = 'pg_catalog') WHERE c.relnamespace = n.oid AND c.relname LIKE 'PROBABLYNOT' AND (false OR (c.relkind = 'r' AND n.nspname !~ '^pg_' AND n.nspname <> 'information_schema')) ORDER BY TABLE_TYPE, TABLE_SCHEM, TABLE_NAME;
Date: 2026-01-21 06:11:49 Duration: 2ms Database: postgres
-
SELECT NULL AS TABLE_CAT, n.nspname AS TABLE_SCHEM, c.relname AS TABLE_NAME, CASE n.nspname ~ '^pg_' OR n.nspname = 'information_schema' WHEN true THEN CASE WHEN n.nspname = 'pg_catalog' OR n.nspname = 'information_schema' THEN CASE c.relkind WHEN 'r' THEN 'SYSTEM TABLE' WHEN 'v' THEN 'SYSTEM VIEW' WHEN 'i' THEN 'SYSTEM INDEX' ELSE NULL END WHEN n.nspname = 'pg_toast' THEN CASE c.relkind WHEN 'r' THEN 'SYSTEM TOAST TABLE' WHEN 'i' THEN 'SYSTEM TOAST INDEX' ELSE NULL END ELSE CASE c.relkind WHEN 'r' THEN 'TEMPORARY TABLE' WHEN 'p' THEN 'TEMPORARY TABLE' WHEN 'i' THEN 'TEMPORARY INDEX' WHEN 'S' THEN 'TEMPORARY SEQUENCE' WHEN 'v' THEN 'TEMPORARY VIEW' ELSE NULL END END WHEN false THEN CASE c.relkind WHEN 'r' THEN 'TABLE' WHEN 'p' THEN 'PARTITIONED TABLE' WHEN 'i' THEN 'INDEX' WHEN 'S' THEN 'SEQUENCE' WHEN 'v' THEN 'VIEW' WHEN 'c' THEN 'TYPE' WHEN 'f' THEN 'FOREIGN TABLE' WHEN 'm' THEN 'MATERIALIZED VIEW' ELSE NULL END ELSE NULL END AS TABLE_TYPE, d.description AS REMARKS, '' as TYPE_CAT, '' as TYPE_SCHEM, '' as TYPE_NAME, '' AS SELF_REFERENCING_COL_NAME, '' AS REF_GENERATION FROM pg_catalog.pg_namespace n, pg_catalog.pg_class c LEFT JOIN pg_catalog.pg_description d ON (c.oid = d.objoid AND d.objsubid = 0) LEFT JOIN pg_catalog.pg_class dc ON (d.classoid = dc.oid AND dc.relname = 'pg_class') LEFT JOIN pg_catalog.pg_namespace dn ON (dn.oid = dc.relnamespace AND dn.nspname = 'pg_catalog') WHERE c.relnamespace = n.oid AND c.relname LIKE 'PROBABLYNOT' AND (false OR (c.relkind = 'r' AND n.nspname !~ '^pg_' AND n.nspname <> 'information_schema')) ORDER BY TABLE_TYPE, TABLE_SCHEM, TABLE_NAME;
Date: 2026-01-21 06:11:50 Duration: 1ms Database: postgres
-
SELECT NULL AS TABLE_CAT, n.nspname AS TABLE_SCHEM, c.relname AS TABLE_NAME, CASE n.nspname ~ '^pg_' OR n.nspname = 'information_schema' WHEN true THEN CASE WHEN n.nspname = 'pg_catalog' OR n.nspname = 'information_schema' THEN CASE c.relkind WHEN 'r' THEN 'SYSTEM TABLE' WHEN 'v' THEN 'SYSTEM VIEW' WHEN 'i' THEN 'SYSTEM INDEX' ELSE NULL END WHEN n.nspname = 'pg_toast' THEN CASE c.relkind WHEN 'r' THEN 'SYSTEM TOAST TABLE' WHEN 'i' THEN 'SYSTEM TOAST INDEX' ELSE NULL END ELSE CASE c.relkind WHEN 'r' THEN 'TEMPORARY TABLE' WHEN 'p' THEN 'TEMPORARY TABLE' WHEN 'i' THEN 'TEMPORARY INDEX' WHEN 'S' THEN 'TEMPORARY SEQUENCE' WHEN 'v' THEN 'TEMPORARY VIEW' ELSE NULL END END WHEN false THEN CASE c.relkind WHEN 'r' THEN 'TABLE' WHEN 'p' THEN 'PARTITIONED TABLE' WHEN 'i' THEN 'INDEX' WHEN 'S' THEN 'SEQUENCE' WHEN 'v' THEN 'VIEW' WHEN 'c' THEN 'TYPE' WHEN 'f' THEN 'FOREIGN TABLE' WHEN 'm' THEN 'MATERIALIZED VIEW' ELSE NULL END ELSE NULL END AS TABLE_TYPE, d.description AS REMARKS, '' as TYPE_CAT, '' as TYPE_SCHEM, '' as TYPE_NAME, '' AS SELF_REFERENCING_COL_NAME, '' AS REF_GENERATION FROM pg_catalog.pg_namespace n, pg_catalog.pg_class c LEFT JOIN pg_catalog.pg_description d ON (c.oid = d.objoid AND d.objsubid = 0) LEFT JOIN pg_catalog.pg_class dc ON (d.classoid = dc.oid AND dc.relname = 'pg_class') LEFT JOIN pg_catalog.pg_namespace dn ON (dn.oid = dc.relnamespace AND dn.nspname = 'pg_catalog') WHERE c.relnamespace = n.oid AND c.relname LIKE 'PROBABLYNOT' AND (false OR (c.relkind = 'r' AND n.nspname !~ '^pg_' AND n.nspname <> 'information_schema')) ORDER BY TABLE_TYPE, TABLE_SCHEM, TABLE_NAME;
Date: 2026-01-21 06:11:50 Duration: 1ms Database: postgres
14 92ms 93 0ms 1ms 0ms SELECT timegranularity FROM brokersymbollist bsl INNER JOIN symbols s ON bsl.symbolid = s.symbolid INNER JOIN downloadersymbolsettings dss on s.symbolid = dss.symbolid LEFT OUTER JOIN brokerinstrumentmapping bdfi ON bdfi.brokerid = $1 AND dss.datafeedinstrumentid = bdfi.datafeedinstrumentid WHERE s.nonliquid = 0 and s.deleted = 0 and dss.enabled = 1 AND s.symbol ILIKE $2 AND bsl.brokerid = $3 AND timegranularity >= 15 ORDER BY timegranularity LIMIT 1;Times Reported Time consuming bind #14
Day Hour Count Duration Avg duration 06 93 92ms 0ms -
SELECT timegranularity FROM brokersymbollist bsl INNER JOIN symbols s ON bsl.symbolid = s.symbolid INNER JOIN downloadersymbolsettings dss on s.symbolid = dss.symbolid LEFT OUTER JOIN brokerinstrumentmapping bdfi ON bdfi.brokerid = $1 AND dss.datafeedinstrumentid = bdfi.datafeedinstrumentid WHERE s.nonliquid = 0 and s.deleted = 0 and dss.enabled = 1 AND s.symbol ILIKE $2 AND bsl.brokerid = $3 AND timegranularity >= 15 ORDER BY timegranularity LIMIT 1;
Date: 2026-01-21 06:15:46 Duration: 1ms Database: postgres parameters: $1 = '558', $2 = 'EURAUD', $3 = '558'
-
SELECT timegranularity FROM brokersymbollist bsl INNER JOIN symbols s ON bsl.symbolid = s.symbolid INNER JOIN downloadersymbolsettings dss on s.symbolid = dss.symbolid LEFT OUTER JOIN brokerinstrumentmapping bdfi ON bdfi.brokerid = $1 AND dss.datafeedinstrumentid = bdfi.datafeedinstrumentid WHERE s.nonliquid = 0 and s.deleted = 0 and dss.enabled = 1 AND s.symbol ILIKE $2 AND bsl.brokerid = $3 AND timegranularity >= 15 ORDER BY timegranularity LIMIT 1;
Date: 2026-01-21 06:46:32 Duration: 1ms Database: postgres parameters: $1 = '689', $2 = 'XAUUSD', $3 = '689'
-
SELECT timegranularity FROM brokersymbollist bsl INNER JOIN symbols s ON bsl.symbolid = s.symbolid INNER JOIN downloadersymbolsettings dss on s.symbolid = dss.symbolid LEFT OUTER JOIN brokerinstrumentmapping bdfi ON bdfi.brokerid = $1 AND dss.datafeedinstrumentid = bdfi.datafeedinstrumentid WHERE s.nonliquid = 0 and s.deleted = 0 and dss.enabled = 1 AND s.symbol ILIKE $2 AND bsl.brokerid = $3 AND timegranularity >= 15 ORDER BY timegranularity LIMIT 1;
Date: 2026-01-21 06:20:33 Duration: 1ms Database: postgres parameters: $1 = '689', $2 = 'XAUUSD', $3 = '689'
15 83ms 6 3ms 40ms 13ms with rankedmt4 as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors ), last_feed_entry as ( select * from rankedmt4 where r = 1 ), ok_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where status = 'OK' ), earliest_entry_after_ok as ( select m.datafeedname, min(m.eventtimestamp) as eventtimestamp from mt4datafeederrors m left outer join ( select datafeedname, eventtimestamp from ok_entries where r = 1) oo on m.datafeedname = oo.datafeedname where m.eventtimestamp > coalesce(oo.eventtimestamp, '1900-01-01'::timestamp without time zone) group by m.datafeedname ), notified_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where notified is not null and notified <> '' ), broker as ( select *, row_number() over (partition by feedname order by brokerid) r from ( select distinct b.brokerid, b.name as brokername, dss.classname as feedname from downloadersymbolsettings dss inner join brokersymbollist bsl on dss.symbolid = bsl.symbolid inner join broker b on bsl.brokerid = b.brokerid where dss.enabled = 1) a ) select last.id, last.datafeedname, last.eventtimestamp, last.status, last.errordescription, last.serveraddress, last.username, note.notified, note.eventtimestamp, broker.brokername from last_feed_entry last inner join earliest_entry_after_ok after_ok on last.datafeedname = after_ok.datafeedname inner join broker on last.datafeedname = broker.feedname left outer join ok_entries ok on ok.datafeedname = last.datafeedname left outer join notified_entries note on note.datafeedname = last.datafeedname and note.r = 1 where (ok.r is null or ok.r = 1) and last.datafeedname not in ( select distinct datafeedname from last_feed_entry where status = 'OK') and extract(epoch from (last.eventtimestamp - after_ok.eventtimestamp)) > 60 * 60 and last.eventtimestamp > current_timestamp - interval '1 day' and (note.eventtimestamp is null or note.eventtimestamp < current_timestamp - interval '10 hours') and last.eventtimestamp > current_timestamp - interval '1 hour' and broker.r = 1;Times Reported Time consuming bind #15
Day Hour Count Duration Avg duration 06 6 83ms 13ms -
with rankedmt4 as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors ), last_feed_entry as ( select * from rankedmt4 where r = 1 ), ok_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where status = 'OK' ), earliest_entry_after_ok as ( select m.datafeedname, min(m.eventtimestamp) as eventtimestamp from mt4datafeederrors m left outer join ( select datafeedname, eventtimestamp from ok_entries where r = 1) oo on m.datafeedname = oo.datafeedname where m.eventtimestamp > coalesce(oo.eventtimestamp, '1900-01-01'::timestamp without time zone) group by m.datafeedname ), notified_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where notified is not null and notified <> '' ), broker as ( select *, row_number() over (partition by feedname order by brokerid) r from ( select distinct b.brokerid, b.name as brokername, dss.classname as feedname from downloadersymbolsettings dss inner join brokersymbollist bsl on dss.symbolid = bsl.symbolid inner join broker b on bsl.brokerid = b.brokerid where dss.enabled = 1) a ) select last.id, last.datafeedname, last.eventtimestamp, last.status, last.errordescription, last.serveraddress, last.username, note.notified, note.eventtimestamp, broker.brokername from last_feed_entry last inner join earliest_entry_after_ok after_ok on last.datafeedname = after_ok.datafeedname inner join broker on last.datafeedname = broker.feedname left outer join ok_entries ok on ok.datafeedname = last.datafeedname left outer join notified_entries note on note.datafeedname = last.datafeedname and note.r = 1 where (ok.r is null or ok.r = 1) and last.datafeedname not in ( select distinct datafeedname from last_feed_entry where status = 'OK') and extract(epoch from (last.eventtimestamp - after_ok.eventtimestamp)) > 60 * 60 and last.eventtimestamp > current_timestamp - interval '1 day' and (note.eventtimestamp is null or note.eventtimestamp < current_timestamp - interval '10 hours') and last.eventtimestamp > current_timestamp - interval '1 hour' and broker.r = 1;
Date: 2026-01-21 06:40:03 Duration: 40ms Database: postgres
-
with rankedmt4 as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors ), last_feed_entry as ( select * from rankedmt4 where r = 1 ), ok_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where status = 'OK' ), earliest_entry_after_ok as ( select m.datafeedname, min(m.eventtimestamp) as eventtimestamp from mt4datafeederrors m left outer join ( select datafeedname, eventtimestamp from ok_entries where r = 1) oo on m.datafeedname = oo.datafeedname where m.eventtimestamp > coalesce(oo.eventtimestamp, '1900-01-01'::timestamp without time zone) group by m.datafeedname ), notified_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where notified is not null and notified <> '' ), broker as ( select *, row_number() over (partition by feedname order by brokerid) r from ( select distinct b.brokerid, b.name as brokername, dss.classname as feedname from downloadersymbolsettings dss inner join brokersymbollist bsl on dss.symbolid = bsl.symbolid inner join broker b on bsl.brokerid = b.brokerid where dss.enabled = 1) a ) select last.id, last.datafeedname, last.eventtimestamp, last.status, last.errordescription, last.serveraddress, last.username, note.notified, note.eventtimestamp, broker.brokername from last_feed_entry last inner join earliest_entry_after_ok after_ok on last.datafeedname = after_ok.datafeedname inner join broker on last.datafeedname = broker.feedname left outer join ok_entries ok on ok.datafeedname = last.datafeedname left outer join notified_entries note on note.datafeedname = last.datafeedname and note.r = 1 where (ok.r is null or ok.r = 1) and last.datafeedname not in ( select distinct datafeedname from last_feed_entry where status = 'OK') and extract(epoch from (last.eventtimestamp - after_ok.eventtimestamp)) > 60 * 60 and last.eventtimestamp > current_timestamp - interval '1 day' and (note.eventtimestamp is null or note.eventtimestamp < current_timestamp - interval '10 hours') and last.eventtimestamp > current_timestamp - interval '1 hour' and broker.r = 1;
Date: 2026-01-21 06:20:02 Duration: 11ms Database: postgres
-
with rankedmt4 as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors ), last_feed_entry as ( select * from rankedmt4 where r = 1 ), ok_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where status = 'OK' ), earliest_entry_after_ok as ( select m.datafeedname, min(m.eventtimestamp) as eventtimestamp from mt4datafeederrors m left outer join ( select datafeedname, eventtimestamp from ok_entries where r = 1) oo on m.datafeedname = oo.datafeedname where m.eventtimestamp > coalesce(oo.eventtimestamp, '1900-01-01'::timestamp without time zone) group by m.datafeedname ), notified_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where notified is not null and notified <> '' ), broker as ( select *, row_number() over (partition by feedname order by brokerid) r from ( select distinct b.brokerid, b.name as brokername, dss.classname as feedname from downloadersymbolsettings dss inner join brokersymbollist bsl on dss.symbolid = bsl.symbolid inner join broker b on bsl.brokerid = b.brokerid where dss.enabled = 1) a ) select last.id, last.datafeedname, last.eventtimestamp, last.status, last.errordescription, last.serveraddress, last.username, note.notified, note.eventtimestamp, broker.brokername from last_feed_entry last inner join earliest_entry_after_ok after_ok on last.datafeedname = after_ok.datafeedname inner join broker on last.datafeedname = broker.feedname left outer join ok_entries ok on ok.datafeedname = last.datafeedname left outer join notified_entries note on note.datafeedname = last.datafeedname and note.r = 1 where (ok.r is null or ok.r = 1) and last.datafeedname not in ( select distinct datafeedname from last_feed_entry where status = 'OK') and extract(epoch from (last.eventtimestamp - after_ok.eventtimestamp)) > 60 * 60 and last.eventtimestamp > current_timestamp - interval '1 day' and (note.eventtimestamp is null or note.eventtimestamp < current_timestamp - interval '10 hours') and last.eventtimestamp > current_timestamp - interval '1 hour' and broker.r = 1;
Date: 2026-01-21 06:50:02 Duration: 11ms Database: postgres
16 66ms 8 4ms 24ms 8ms WITH pre_symbols AS ( /* find relevant symbols */ ;Times Reported Time consuming bind #16
Day Hour Count Duration Avg duration 06 8 66ms 8ms -
WITH pre_symbols AS ( /* find relevant symbols */ ;
Date: 2026-01-21 06:11:48 Duration: 24ms Database: postgres parameters: $1 = '1018', $2 = 'ICMARKETS-AU-MT5', $3 = 'AAPL.NAS', $4 = 'ABBV.NYSE', $5 = 'AMCR.NYSE', $6 = 'AMP.NYSE', $7 = 'AMZN.NAS', $8 = 'ANZ.ASX', $9 = 'AUDJPY', $10 = 'AUDUSD', $11 = 'AUS200', $12 = 'BABA.NYSE', $13 = 'BIIB.NAS', $14 = 'BXB.ASX', $15 = 'CBA.ASX', $16 = 'CHINA50', $17 = 'CSL.ASX', $18 = 'DE30', $19 = 'ES35', $20 = 'EURCHF', $21 = 'EURGBP', $22 = 'EURUSD', $23 = 'F40', $24 = 'FMG.ASX', $25 = 'GBPJPY', $26 = 'GBPUSD', $27 = 'GOOG.NAS', $28 = 'HK50', $29 = 'IT40', $30 = 'JP225', $31 = 'KO.NYSE', $32 = 'MQG.ASX', $33 = 'MSFT.NAS', $34 = 'NAB.ASX', $35 = 'NFLX.NAS', $36 = 'PYPL.NAS', $37 = 'QBE.ASX', $38 = 'STOXX50', $39 = 'SUN.ASX', $40 = 'TCL.ASX', $41 = 'TLS.ASX', $42 = 'TSLA.NAS', $43 = 'UK100', $44 = 'UNH.NYSE', $45 = 'US2000', $46 = 'US30', $47 = 'US500', $48 = 'USDCAD', $49 = 'USDCHF', $50 = 'USDCNH', $51 = 'USDJPY', $52 = 'USTEC', $53 = 'WBC.ASX', $54 = 'WES.ASX', $55 = 'WOW.ASX', $56 = 'WPL.ASX', $57 = 'XAUEUR', $58 = 'XAUUSD', $59 = 'XBRUSD', $60 = 'XTIUSD', $61 = 'AAPL.NAS', $62 = 'ABBV.NYSE', $63 = 'AMCR.NYSE', $64 = 'AMP.NYSE', $65 = 'AMZN.NAS', $66 = 'ANZ.ASX', $67 = 'AUDJPY', $68 = 'AUDUSD', $69 = 'AUS200', $70 = 'BABA.NYSE', $71 = 'BIIB.NAS', $72 = 'BXB.ASX', $73 = 'CBA.ASX', $74 = 'CHINA50', $75 = 'CSL.ASX', $76 = 'DE30', $77 = 'ES35', $78 = 'EURCHF', $79 = 'EURGBP', $80 = 'EURUSD', $81 = 'F40', $82 = 'FMG.ASX', $83 = 'GBPJPY', $84 = 'GBPUSD', $85 = 'GOOG.NAS', $86 = 'HK50', $87 = 'IT40', $88 = 'JP225', $89 = 'KO.NYSE', $90 = 'MQG.ASX', $91 = 'MSFT.NAS', $92 = 'NAB.ASX', $93 = 'NFLX.NAS', $94 = 'PYPL.NAS', $95 = 'QBE.ASX', $96 = 'STOXX50', $97 = 'SUN.ASX', $98 = 'TCL.ASX', $99 = 'TLS.ASX', $100 = 'TSLA.NAS', $101 = 'UK100', $102 = 'UNH.NYSE', $103 = 'US2000', $104 = 'US30', $105 = 'US500', $106 = 'USDCAD', $107 = 'USDCHF', $108 = 'USDCNH', $109 = 'USDJPY', $110 = 'USTEC', $111 = 'WBC.ASX', $112 = 'WES.ASX', $113 = 'WOW.ASX', $114 = 'WPL.ASX', $115 = 'XAUEUR', $116 = 'XAUUSD', $117 = 'XBRUSD', $118 = 'XTIUSD', $119 = '5'
-
WITH pre_symbols AS ( /* find relevant symbols */ ;
Date: 2026-01-21 06:11:48 Duration: 10ms Database: postgres parameters: $1 = '1018', $2 = 'ICMARKETS-AU-MT5', $3 = 'AAPL.NAS', $4 = 'ABBV.NYSE', $5 = 'AMCR.NYSE', $6 = 'AMP.NYSE', $7 = 'AMZN.NAS', $8 = 'ANZ.ASX', $9 = 'AUDJPY', $10 = 'AUDUSD', $11 = 'AUS200', $12 = 'BABA.NYSE', $13 = 'BIIB.NAS', $14 = 'BXB.ASX', $15 = 'CBA.ASX', $16 = 'CHINA50', $17 = 'CSL.ASX', $18 = 'DE30', $19 = 'ES35', $20 = 'EURCHF', $21 = 'EURGBP', $22 = 'EURUSD', $23 = 'F40', $24 = 'FMG.ASX', $25 = 'GBPJPY', $26 = 'GBPUSD', $27 = 'GOOG.NAS', $28 = 'HK50', $29 = 'IT40', $30 = 'JP225', $31 = 'KO.NYSE', $32 = 'MQG.ASX', $33 = 'MSFT.NAS', $34 = 'NAB.ASX', $35 = 'NFLX.NAS', $36 = 'PYPL.NAS', $37 = 'QBE.ASX', $38 = 'STOXX50', $39 = 'SUN.ASX', $40 = 'TCL.ASX', $41 = 'TLS.ASX', $42 = 'TSLA.NAS', $43 = 'UK100', $44 = 'UNH.NYSE', $45 = 'US2000', $46 = 'US30', $47 = 'US500', $48 = 'USDCAD', $49 = 'USDCHF', $50 = 'USDCNH', $51 = 'USDJPY', $52 = 'USTEC', $53 = 'WBC.ASX', $54 = 'WES.ASX', $55 = 'WOW.ASX', $56 = 'WPL.ASX', $57 = 'XAUEUR', $58 = 'XAUUSD', $59 = 'XBRUSD', $60 = 'XTIUSD', $61 = 'AAPL.NAS', $62 = 'ABBV.NYSE', $63 = 'AMCR.NYSE', $64 = 'AMP.NYSE', $65 = 'AMZN.NAS', $66 = 'ANZ.ASX', $67 = 'AUDJPY', $68 = 'AUDUSD', $69 = 'AUS200', $70 = 'BABA.NYSE', $71 = 'BIIB.NAS', $72 = 'BXB.ASX', $73 = 'CBA.ASX', $74 = 'CHINA50', $75 = 'CSL.ASX', $76 = 'DE30', $77 = 'ES35', $78 = 'EURCHF', $79 = 'EURGBP', $80 = 'EURUSD', $81 = 'F40', $82 = 'FMG.ASX', $83 = 'GBPJPY', $84 = 'GBPUSD', $85 = 'GOOG.NAS', $86 = 'HK50', $87 = 'IT40', $88 = 'JP225', $89 = 'KO.NYSE', $90 = 'MQG.ASX', $91 = 'MSFT.NAS', $92 = 'NAB.ASX', $93 = 'NFLX.NAS', $94 = 'PYPL.NAS', $95 = 'QBE.ASX', $96 = 'STOXX50', $97 = 'SUN.ASX', $98 = 'TCL.ASX', $99 = 'TLS.ASX', $100 = 'TSLA.NAS', $101 = 'UK100', $102 = 'UNH.NYSE', $103 = 'US2000', $104 = 'US30', $105 = 'US500', $106 = 'USDCAD', $107 = 'USDCHF', $108 = 'USDCNH', $109 = 'USDJPY', $110 = 'USTEC', $111 = 'WBC.ASX', $112 = 'WES.ASX', $113 = 'WOW.ASX', $114 = 'WPL.ASX', $115 = 'XAUEUR', $116 = 'XAUUSD', $117 = 'XBRUSD', $118 = 'XTIUSD', $119 = '5'
-
WITH pre_symbols AS ( /* find relevant symbols */ ;
Date: 2026-01-21 06:11:49 Duration: 8ms Database: postgres parameters: $1 = '1018', $2 = 'ICMARKETS-AU-MT5', $3 = 'AAPL.NAS', $4 = 'ABBV.NYSE', $5 = 'AMCR.NYSE', $6 = 'AMP.NYSE', $7 = 'AMZN.NAS', $8 = 'ANZ.ASX', $9 = 'AUDJPY', $10 = 'AUDUSD', $11 = 'AUS200', $12 = 'BABA.NYSE', $13 = 'BIIB.NAS', $14 = 'BXB.ASX', $15 = 'CBA.ASX', $16 = 'CHINA50', $17 = 'CSL.ASX', $18 = 'DE30', $19 = 'ES35', $20 = 'EURCHF', $21 = 'EURGBP', $22 = 'EURUSD', $23 = 'F40', $24 = 'FMG.ASX', $25 = 'GBPJPY', $26 = 'GBPUSD', $27 = 'GOOG.NAS', $28 = 'HK50', $29 = 'IT40', $30 = 'JP225', $31 = 'KO.NYSE', $32 = 'MQG.ASX', $33 = 'MSFT.NAS', $34 = 'NAB.ASX', $35 = 'NFLX.NAS', $36 = 'PYPL.NAS', $37 = 'QBE.ASX', $38 = 'STOXX50', $39 = 'SUN.ASX', $40 = 'TCL.ASX', $41 = 'TLS.ASX', $42 = 'TSLA.NAS', $43 = 'UK100', $44 = 'UNH.NYSE', $45 = 'US2000', $46 = 'US30', $47 = 'US500', $48 = 'USDCAD', $49 = 'USDCHF', $50 = 'USDCNH', $51 = 'USDJPY', $52 = 'USTEC', $53 = 'WBC.ASX', $54 = 'WES.ASX', $55 = 'WOW.ASX', $56 = 'WPL.ASX', $57 = 'XAUEUR', $58 = 'XAUUSD', $59 = 'XBRUSD', $60 = 'XTIUSD', $61 = 'AAPL.NAS', $62 = 'ABBV.NYSE', $63 = 'AMCR.NYSE', $64 = 'AMP.NYSE', $65 = 'AMZN.NAS', $66 = 'ANZ.ASX', $67 = 'AUDJPY', $68 = 'AUDUSD', $69 = 'AUS200', $70 = 'BABA.NYSE', $71 = 'BIIB.NAS', $72 = 'BXB.ASX', $73 = 'CBA.ASX', $74 = 'CHINA50', $75 = 'CSL.ASX', $76 = 'DE30', $77 = 'ES35', $78 = 'EURCHF', $79 = 'EURGBP', $80 = 'EURUSD', $81 = 'F40', $82 = 'FMG.ASX', $83 = 'GBPJPY', $84 = 'GBPUSD', $85 = 'GOOG.NAS', $86 = 'HK50', $87 = 'IT40', $88 = 'JP225', $89 = 'KO.NYSE', $90 = 'MQG.ASX', $91 = 'MSFT.NAS', $92 = 'NAB.ASX', $93 = 'NFLX.NAS', $94 = 'PYPL.NAS', $95 = 'QBE.ASX', $96 = 'STOXX50', $97 = 'SUN.ASX', $98 = 'TCL.ASX', $99 = 'TLS.ASX', $100 = 'TSLA.NAS', $101 = 'UK100', $102 = 'UNH.NYSE', $103 = 'US2000', $104 = 'US30', $105 = 'US500', $106 = 'USDCAD', $107 = 'USDCHF', $108 = 'USDCNH', $109 = 'USDJPY', $110 = 'USTEC', $111 = 'WBC.ASX', $112 = 'WES.ASX', $113 = 'WOW.ASX', $114 = 'WPL.ASX', $115 = 'XAUEUR', $116 = 'XAUUSD', $117 = 'XBRUSD', $118 = 'XTIUSD'
17 50ms 1 50ms 50ms 50ms with maxwhid as ( ;Times Reported Time consuming bind #17
Day Hour Count Duration Avg duration 06 1 50ms 50ms -
with maxwhid as ( ;
Date: 2026-01-21 06:11:46 Duration: 50ms Database: postgres parameters: $1 = '335', $2 = '621', $3 = '637', $4 = '642', $5 = '666', $6 = '660', $7 = '643', $8 = '630', $9 = '680', $10 = '641', $11 = '431', $12 = '622', $13 = '489', $14 = '529', $15 = '576', $16 = '665', $17 = '667', $18 = '558', $19 = '620', $20 = '125', $21 = '488', $22 = '567', $23 = '689', $24 = '700', $25 = '758', $26 = '763', $27 = '765', $28 = '817', $29 = '914', $30 = '972'
18 38ms 4 0ms 36ms 9ms select distinct replace(recognitionengine, ' ', '_') || '%' || replace(datetimeupdate::text, ' ', '_') as t from latest_candle_datetime_per_receng order by t;Times Reported Time consuming bind #18
Day Hour Count Duration Avg duration 06 4 38ms 9ms -
select distinct replace(recognitionengine, ' ', '_') || '%' || replace(datetimeupdate::text, ' ', '_') as t from latest_candle_datetime_per_receng order by t;
Date: 2026-01-21 06:40:10 Duration: 36ms Database: postgres
-
select distinct replace(recognitionengine, ' ', '_') || '%' || replace(datetimeupdate::text, ' ', '_') as t from latest_candle_datetime_per_receng order by t;
Date: 2026-01-21 06:10:11 Duration: 1ms Database: postgres
-
select distinct replace(recognitionengine, ' ', '_') || '%' || replace(datetimeupdate::text, ' ', '_') as t from latest_candle_datetime_per_receng order by t;
Date: 2026-01-21 06:40:10 Duration: 0ms Database: postgres
19 37ms 7 3ms 6ms 5ms SELECT DISTINCT ON (basegroupname, symbol) ;Times Reported Time consuming bind #19
Day Hour Count Duration Avg duration 06 7 37ms 5ms -
SELECT DISTINCT ON (basegroupname, symbol) ;
Date: 2026-01-21 06:17:04 Duration: 6ms Database: postgres parameters: $1 = '958', $2 = '958'
-
SELECT DISTINCT ON (basegroupname, symbol) ;
Date: 2026-01-21 06:17:03 Duration: 6ms Database: postgres parameters: $1 = '958', $2 = '958'
-
SELECT DISTINCT ON (basegroupname, symbol) ;
Date: 2026-01-21 06:52:58 Duration: 6ms Database: postgres parameters: $1 = '667', $2 = '667'
20 27ms 31 0ms 3ms 0ms WITH tr_max AS ( ;Times Reported Time consuming bind #20
Day Hour Count Duration Avg duration 06 31 27ms 0ms -
WITH tr_max AS ( ;
Date: 2026-01-21 06:11:49 Duration: 3ms Database: postgres parameters: $1 = '606715250232663300', $2 = '606715250232663300', $3 = '606715250232663300', $4 = '2', $5 = '2', $6 = '1320', $7 = 't', $8 = '1320'
-
WITH tr_max AS ( ;
Date: 2026-01-21 06:11:49 Duration: 1ms Database: postgres parameters: $1 = '606715250640565300', $2 = '606715250640565300', $3 = '606715250640565300', $4 = '2', $5 = '2', $6 = '1320', $7 = 't', $8 = '1320'
-
WITH tr_max AS ( ;
Date: 2026-01-21 06:11:50 Duration: 1ms Database: postgres parameters: $1 = '605679104132143300', $2 = '605679104132143300', $3 = '605679104132143300', $4 = '3', $5 = '3', $6 = '240', $7 = 't', $8 = '240'
-
Events
Log levels
Key values
- 322,318 Log entries
Events distribution
Key values
- 0 PANIC entries
- 0 FATAL entries
- 283 ERROR entries
- 0 WARNING entries
Most Frequent Errors/Events
Key values
- 239 Max number of times the same event was reported
- 283 Total events found
Rank Times reported Error 1 239 ERROR: permission denied for table datafeed_restarter_events
Times Reported Most Frequent Error / Event #1
Day Hour Count Jan 21 06 239 - ERROR: permission denied for table datafeed_restarter_events
Statement: /* service='datadog-agent' */ select count(*) from datafeed_restarter_events where is_current_entry = 1;
Date: 2026-01-21 06:00:03
2 34 ERROR: function fixcandlegaps(...) is not unique
Times Reported Most Frequent Error / Event #2
Day Hour Count Jan 21 06 34 - ERROR: function fixcandlegaps(unknown, boolean) is not unique at character 8
Hint: Could not choose a best candidate function. You might need to add explicit type casts.
Statement: select fixcandlegaps('GLOBALFXMT5', false);Date: 2026-01-21 06:06:01
3 6 LOG: process ... still waiting for AccessShareLock on relation ... of database ... after ... ms
Times Reported Most Frequent Error / Event #3
Day Hour Count Jan 21 06 6 - LOG: process 11696 still waiting for AccessShareLock on relation 5894441 of database 5881926 after 1000.052 ms at character 77
- LOG: process 9711 still waiting for AccessShareLock on relation 5894441 of database 5881926 after 1000.060 ms at character 28
- LOG: process 11812 still waiting for AccessShareLock on relation 5894441 of database 5881926 after 1000.055 ms at character 77
Detail: Process holding the lock: 7968. Wait queue: 11687, 11696.
Statement: select recognitionengine,to_char(datetimeupdate, 'yyyy-mm-dd HH24:MI') from latest_candle_datetime_per_recengDate: 2026-01-21 06:17:05
Detail: Process holding the lock: 7968. Wait queue: 11687, 11696, 9711.
Statement: SELECT datetimeupdate FROM latest_candle_datetime_per_receng WHERE recognitionengine ILIKE 'ICMARKETS-AU-MT5 - 1'Date: 2026-01-21 06:17:20
Detail: Process holding the lock: 7968. Wait queue: 11687, 11696, 9711, 11812.
Statement: select recognitionengine,to_char(datetimeupdate, 'yyyy-mm-dd HH24:MI') from latest_candle_datetime_per_recengDate: 2026-01-21 06:17:40
4 4 LOG: process ... still waiting for AccessExclusiveLock on relation ... of database ... after ... ms
Times Reported Most Frequent Error / Event #4
Day Hour Count Jan 21 06 4 - LOG: process 8023 still waiting for AccessExclusiveLock on relation 5883477 of database 5881926 after 1000.047 ms
- LOG: process 8023 still waiting for AccessExclusiveLock on relation 5883477 of database 5881926 after 51755.076 ms
- LOG: process 11687 still waiting for AccessExclusiveLock on relation 5894441 of database 5881926 after 1000.726 ms
Detail: Process holding the lock: 7968. Wait queue: 8023.
Statement: TRUNCATE TABLE solr_relevance_old;Date: 2026-01-21 06:05:12
Detail: Process holding the lock: 7968. Wait queue: 8023.
Statement: TRUNCATE TABLE solr_relevance_old;Date: 2026-01-21 06:06:02
Detail: Process holding the lock: 7968. Wait queue: 11687.
Statement: refresh materialized view latest_candle_datetime_per_recengDate: 2026-01-21 06:17:02