-
Global information
- Generated on Sun Jan 11 23:59:53 2026
- Log file: /home/postgres/pg_data/data/pg_log/postgresql-2026-01-12_010000.log
- Parsed 2,413,496 log entries in 51s
- Log start from 2026-01-12 01:00:00 to 2026-01-12 01:59:50
-
Overview
Global Stats
- 231 Number of unique normalized queries
- 233,840 Number of queries
- 1h32m49s Total query duration
- 2026-01-12 01:00:00 First query
- 2026-01-12 01:59:50 Last query
- 1,961 queries/s at 2026-01-12 01:15:04 Query peak
- 1h32m49s Total query duration
- 14s441ms Prepare/parse total duration
- 1m3s Bind total duration
- 1h31m32s Execute total duration
- 0 Number of events
- 0 Number of unique normalized events
- 0 Max number of times the same event was reported
- 0 Number of cancellation
- 34 Total number of automatic vacuums
- 48 Total number of automatic analyzes
- 521 Number temporary file
- 188.66 MiB Max size of temporary file
- 8.54 MiB Average size of temporary file
- 5,535 Total number of sessions
- 13 sessions at 2026-01-12 01:57:44 Session peak
- 2d9h39m44s Total duration of sessions
- 37s503ms Average duration of sessions
- 42 Average queries per session
- 1s6ms Average queries duration per session
- 36s497ms Average idle time per session
- 5,535 Total number of connections
- 50 connections/s at 2026-01-12 01:16:34 Connection peak
- 3 Total number of databases
SQL Traffic
Key values
- 1,961 queries/s Query Peak
- 2026-01-12 01:15:04 Date
SELECT Traffic
Key values
- 886 queries/s Query Peak
- 2026-01-12 01:15:04 Date
INSERT/UPDATE/DELETE Traffic
Key values
- 219 queries/s Query Peak
- 2026-01-12 01:00:56 Date
Queries duration
Key values
- 1h32m49s Total query duration
Prepared queries ratio
Key values
- 0.00 Ratio of bind vs prepare
- 0.00 % Ratio between prepared and "usual" statements
General Activity
↑ Back to the top of the General Activity tableDay Hour Count Min duration Max duration Avg duration Latency Percentile(90) Latency Percentile(95) Latency Percentile(99) Jan 12 01 233,840 0ms 20s909ms 23ms 3m42s 3m57s 4m26s Day Hour SELECT COPY TO Average Duration Latency Percentile(90) Latency Percentile(95) Latency Percentile(99) Jan 12 01 62,980 26 0ms 0ms 0ms 0ms Day Hour INSERT UPDATE DELETE COPY FROM Average Duration Latency Percentile(90) Latency Percentile(95) Latency Percentile(99) Jan 12 01 31,692 2,402 16 96 0ms 0ms 0ms 0ms Day Hour Prepare Bind Bind/Prepare Percentage of prepare Jan 12 01 40,407 88,817 2.20 37.68% Day Hour Count Average / Second Jan 12 01 5,535 1.54/s Day Hour Count Average Duration Average idle time Jan 12 01 5,535 37s503ms 36s511ms -
Connections
Established Connections
Key values
- 50 connections Connection Peak
- 2026-01-12 01:16:34 Date
Connections per database
Key values
- acaweb_fx Main Database
- 5,535 connections Total
Connections per user
Key values
- postgres Main User
- 5,535 connections Total
Connections per host
Key values
- 192.168.1.15 Main host with 2451 connections
- 5,535 Total connections
-
Sessions
Simultaneous sessions
Key values
- 13 sessions Session Peak
- 2026-01-12 01:57:44 Date
Histogram of session times
Key values
- 4,690 0-500ms duration
Sessions per database
Key values
- acaweb_fx Main Database
- 5,535 sessions Total
Sessions per user
Key values
- postgres Main User
- 5,535 sessions Total
Sessions per host
Key values
- 192.168.1.15 Main Host
- 5,535 sessions Total
Host Count Total Duration Average Duration 127.0.0.1 109 12s64ms 110ms 192.168.0.114 2 10m 5m 192.168.0.216 110 3m53s 2s124ms 192.168.0.74 882 7h35m41s 30s999ms 192.168.1.145 25 1d13h7m17s 1h29m5s 192.168.1.15 2,451 2h4m46s 3s54ms 192.168.1.20 42 10h17m32s 14m42s 192.168.1.239 9 51ms 5ms 192.168.1.90 12 32s883ms 2s740ms 192.168.2.126 18 5s622ms 312ms 192.168.2.182 12 826ms 68ms 192.168.2.82 48 12s677ms 264ms 192.168.3.199 36 1s258ms 34ms 192.168.4.142 1,080 13m17s 738ms 192.168.4.222 1 1m1s 1m1s 192.168.4.241 7 31s646ms 4s520ms 192.168.4.33 95 1m25s 902ms 192.168.4.98 330 15s775ms 47ms [local] 266 2m55s 659ms -
Checkpoints / Restartpoints
Checkpoints Buffers
Key values
- 8,940 buffers Checkpoint Peak
- 2026-01-12 01:05:03 Date
- 210.056 seconds Highest write time
- 0.091 seconds Sync time
Checkpoints Wal files
Key values
- 4 files Wal files usage Peak
- 2026-01-12 01:35:04 Date
Checkpoints distance
Key values
- 110.68 Mo Distance Peak
- 2026-01-12 01:10:03 Date
Checkpoints Activity
↑ Back to the top of the Checkpoint Activity tableDay Hour Written buffers Write time Sync time Total time Jan 12 01 51,597 2,098.462s 0.166s 2,098.954s Day Hour Added Removed Recycled Synced files Longest sync Average sync Jan 12 01 0 0 25 2,078 0.056s 0s Day Hour Count Avg time (sec) Jan 12 01 0 0s Day Hour Mean distance Mean estimate Jan 12 01 33,588.67 kB 48,633.25 kB -
Temporary Files
Size of temporary files
Key values
- 184.88 MiB Temp Files size Peak
- 2026-01-12 01:50:06 Date
Number of temporary files
Key values
- 31 per second Temp Files Peak
- 2026-01-12 01:17:10 Date
Temporary Files Activity
↑ Back to the top of the Temporary Files Activity tableDay Hour Count Total size Average size Jan 12 01 521 4.35 GiB 8.54 MiB Queries generating the most temporary files (N)
Rank Count Total size Min size Max size Avg size Query 1 30 93.07 MiB 3.04 MiB 3.31 MiB 3.10 MiB select resultuid from relevance_consecutivecandles_results order by resultuid desc limit ?), all_results as ( select ccr.resultuid as resultuid, ccr.direction as direction, s.exchange as exchange, s.symbolid as symbolid, coalesce(bim.code, s.symbol) as symbol_code, s.longname as symbol_name, s.timegranularity as interval, ccr.patternendtime as identified, dtt.timezone as timezone, ccr.qtyconsecutivecandles as length, g.basegroupname, case when rcr.age is not null then rcr.age when ccr.resultuid <= rm.resultuid then ? else ? end as age, case when rcr.relevant is not null then rcr.relevant when ccr.resultuid <= rm.resultuid then ? else ? end as relevant, cps.pip, newlevels.filtered from consecutivecandles_results ccr inner join brokersymbollist bsl on bsl.brokerid = ? and bsl.symbolid = ccr.symbolid inner join symbols s on ccr.symbolid = s.symbolid and s.nonliquid = ? inner join downloadersymbolsettings dss on ccr.symbolid = dss.symbolid inner join datafeedstimetable dtt on dss.classname = dtt.classname and dtt.dayofweek = ? inner join symbolgroup sg on ccr.symbolid = sg.symbolid inner join groups g on sg.groupid = g.groupid inner join brokergroups bg on g.groupid = bg.groupid and bsl.brokerid = bg.brokerid inner join rar_max rm on ? = ? left outer join relevance_consecutivecandles_results rcr on rcr.resultuid = ccr.resultuid left join currencypips cps on cps.symbol = s.symbol left outer join brokerinstrumentmap bim on dss.datafeedinstrumentid = bim.datafeedinstrumentid and bim.brokerid = bsl.brokerid and bim.type = ? left join lateral calc_cc_signal_filter (ccr.resultuid) newlevels on true where ccr.gmttimefound > now() - interval ? and s.deleted = ? and (ccr.simulation = ? or ccr.simulation is null) and (? = ? or s.timegranularity in (...)) and (? = ? or s.exchange in (...)) and (? = ? or coalesce(bim.code, s.symbol) in (...)) and (? = ? or ccr.patternlengthbars <= ?)), results as ( select distinct on (symbolid) * from all_results where (false = ? or relevant = ?) and (? = ? or age <= ?) order by symbolid, resultuid ) select * from results order by identified desc, length desc;-
SELECT resultuid FROM relevance_consecutivecandles_results ORDER BY resultuid DESC LIMIT 1), all_results AS ( SELECT ccr.resultuid AS resultuid, ccr.direction AS direction, s.exchange AS exchange, s.symbolid AS symbolid, coalesce(bim.code, s.symbol) AS symbol_code, s.longname AS symbol_name, s.timegranularity AS interval, ccr.patternendtime AS identified, dtt.timezone AS timezone, ccr.qtyconsecutivecandles AS length, g.basegroupname, CASE WHEN rcr.age IS NOT NULL THEN rcr.age WHEN ccr.resultuid <= rm.resultuid THEN 1 ELSE 0 END as age, CASE WHEN rcr.relevant IS NOT NULL THEN rcr.relevant WHEN ccr.resultuid <= rm.resultuid THEN 0 ELSE 1 END as relevant, cps.pip, newLevels.filtered FROM consecutivecandles_results ccr INNER JOIN brokersymbollist bsl ON bsl.brokerid = $1 AND bsl.symbolid = ccr.symbolid INNER JOIN symbols s ON ccr.symbolid = s.symbolid AND s.nonliquid = 0 INNER JOIN downloadersymbolsettings dss ON ccr.symbolid = dss.symbolid INNER JOIN datafeedstimetable dtt ON dss.classname = dtt.classname AND dtt.dayofweek = 3 INNER JOIN symbolgroup sg on ccr.symbolid = sg.symbolid INNER JOIN groups g ON sg.groupid = g.groupid INNER JOIN brokergroups bg on g.groupid = bg.groupid AND bsl.brokerid = bg.brokerid INNER JOIN rar_max rm ON 1 = 1 LEFT OUTER JOIN relevance_consecutivecandles_results rcr ON rcr.resultuid = ccr.resultuid LEFT JOIN currencypips cps on cps.symbol = s.symbol LEFT OUTER JOIN brokerinstrumentmap bim ON dss.datafeedinstrumentid = bim.datafeedinstrumentid AND bim.brokerid = bsl.brokerid AND bim.TYPE = 'OUTBOUND' LEFT JOIN LATERAL calc_cc_signal_filter (ccr.resultuid) newLevels on true WHERE ccr.gmttimefound > now() - INTERVAL '7 DAYS' AND s.deleted = 0 AND (ccr.simulation = 0 OR ccr.simulation IS NULL) AND ($2 = 0 OR s.timegranularity in ($3, $4, $5, $6, $7, $8, $9)) AND ($10 = 0 OR s.exchange in ($11)) AND ($12 = 0 OR coalesce(bim.code, s.symbol) in ($13, $14, $15, $16, $17, $18, $19, $20, $21, $22, $23, $24, $25, $26, $27, $28, $29, $30, $31, $32, $33, $34, $35, $36, $37, $38, $39, $40, $41, $42, $43, $44, $45, $46, $47, $48, $49, $50, $51, $52, $53, $54, $55, $56, $57, $58, $59, $60, $61, $62, $63, $64, $65, $66, $67, $68, $69, $70, $71, $72, $73, $74, $75, $76, $77, $78, $79, $80, $81, $82, $83, $84, $85, $86, $87, $88, $89, $90, $91, $92, $93, $94, $95, $96, $97, $98, $99, $100, $101, $102, $103, $104, $105, $106, $107, $108, $109, $110, $111, $112, $113, $114, $115, $116, $117, $118, $119, $120, $121, $122, $123, $124, $125, $126, $127, $128, $129, $130, $131, $132, $133, $134, $135, $136, $137, $138, $139, $140, $141, $142, $143, $144, $145, $146, $147, $148, $149, $150, $151, $152, $153, $154, $155, $156, $157, $158, $159, $160, $161, $162, $163, $164, $165, $166, $167, $168, $169, $170, $171, $172, $173, $174, $175, $176, $177, $178, $179, $180, $181, $182, $183, $184, $185, $186, $187, $188, $189, $190, $191, $192, $193, $194, $195, $196, $197, $198, $199, $200, $201, $202, $203, $204, $205, $206, $207, $208, $209, $210, $211, $212, $213, $214, $215, $216, $217, $218, $219, $220, $221, $222, $223, $224, $225)) AND ($226 = 0 OR ccr.patternlengthbars <= $227)), results AS ( SELECT DISTINCT ON (symbolid) * FROM all_results WHERE (FALSE = $228 OR relevant = 1) AND ($229 = 0 OR age <= $230) ORDER BY symbolid, resultuid ) SELECT * from results ORDER BY identified DESC, length DESC;
Date: 2026-01-12 01:01:23 Duration: 0ms
2 30 1.67 GiB 7.03 MiB 188.66 MiB 56.89 MiB with rankedmt4 as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors ), last_feed_entry as ( select * from rankedmt4 where r = ? ), ok_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where status = ? ), earliest_entry_after_ok as ( select m.datafeedname, min(m.eventtimestamp) as eventtimestamp from mt4datafeederrors m left outer join ( select datafeedname, eventtimestamp from ok_entries where r = ?) oo on m.datafeedname = oo.datafeedname where m.eventtimestamp > coalesce(oo.eventtimestamp, ?::timestamp without time zone) group by m.datafeedname ), notified_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where notified is not null and notified <> ? ), broker as ( select *, row_number() over (partition by feedname order by brokerid) r from ( select distinct b.brokerid, b.name as brokername, dss.classname as feedname from downloadersymbolsettings dss inner join brokersymbollist bsl on dss.symbolid = bsl.symbolid inner join broker b on bsl.brokerid = b.brokerid where dss.enabled = ?) a ) select last.id, last.datafeedname, last.eventtimestamp, last.status, last.errordescription, last.serveraddress, last.username, note.notified, note.eventtimestamp, broker.brokername from last_feed_entry last inner join earliest_entry_after_ok after_ok on last.datafeedname = after_ok.datafeedname inner join broker on last.datafeedname = broker.feedname left outer join ok_entries ok on ok.datafeedname = last.datafeedname left outer join notified_entries note on note.datafeedname = last.datafeedname and note.r = ? where (ok.r is null or ok.r = ?) and last.datafeedname not in ( select distinct datafeedname from last_feed_entry where status = ?) and extract(epoch from (last.eventtimestamp - after_ok.eventtimestamp)) > ? * ? and last.eventtimestamp > current_timestamp - interval ? and (note.eventtimestamp is null or note.eventtimestamp < current_timestamp - interval ?) and last.eventtimestamp > current_timestamp - interval ? and broker.r = ?;-
with rankedmt4 as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors ), last_feed_entry as ( select * from rankedmt4 where r = 1 ), ok_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where status = 'OK' ), earliest_entry_after_ok as ( select m.datafeedname, min(m.eventtimestamp) as eventtimestamp from mt4datafeederrors m left outer join ( select datafeedname, eventtimestamp from ok_entries where r = 1) oo on m.datafeedname = oo.datafeedname where m.eventtimestamp > coalesce(oo.eventtimestamp, '1900-01-01'::timestamp without time zone) group by m.datafeedname ), notified_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where notified is not null and notified <> '' ), broker as ( select *, row_number() over (partition by feedname order by brokerid) r from ( select distinct b.brokerid, b.name as brokername, dss.classname as feedname from downloadersymbolsettings dss inner join brokersymbollist bsl on dss.symbolid = bsl.symbolid inner join broker b on bsl.brokerid = b.brokerid where dss.enabled = 1) a ) select last.id, last.datafeedname, last.eventtimestamp, last.status, last.errordescription, last.serveraddress, last.username, note.notified, note.eventtimestamp, broker.brokername from last_feed_entry last inner join earliest_entry_after_ok after_ok on last.datafeedname = after_ok.datafeedname inner join broker on last.datafeedname = broker.feedname left outer join ok_entries ok on ok.datafeedname = last.datafeedname left outer join notified_entries note on note.datafeedname = last.datafeedname and note.r = 1 where (ok.r is null or ok.r = 1) and last.datafeedname not in ( select distinct datafeedname from last_feed_entry where status = 'OK') and extract(epoch from (last.eventtimestamp - after_ok.eventtimestamp)) > 60 * 60 and last.eventtimestamp > current_timestamp - interval '1 day' and (note.eventtimestamp is null or note.eventtimestamp < current_timestamp - interval '10 hours') and last.eventtimestamp > current_timestamp - interval '1 hour' and broker.r = 1;
Date: 2026-01-12 01:00:05 Duration: 0ms
3 23 71.11 MiB 3.09 MiB 3.09 MiB 3.09 MiB select resultuid from relevance_autochartist_results order by resultuid desc limit ?), ar as ( select a.*, rr.age, rr.relevant from autochartist_results a left outer join relevance_autochartist_results rr on a.resultuid = rr.resultuid where case when false = ? then true else a.resultuid > ( select min(resultuid) from relevance_autochartist_results) end), all_results as ( select ar.resultuid as resultuid, ar.direction as direction, ar.predictiontimeto as predictiontimeto, ar.predictionpricefrom as predictionpricefrom, ar.predictionpriceto as predictionpriceto, cp.pip as pip, s.exchange as exchange, s.symbolid as symbolid, coalesce(bim.code, s.symbol) as symbol_code, s.longname as symbol_name, s.timegranularity as interval, ar.pattern as pattern_name, ar.breakout as breakout, ar.patternendtime as identified, dtt.timezone as timezone, ar.patternlengthbars as length, g.basegroupname, newlevels.profit, newlevels.stop, newlevels.filtered, case when ar.age is not null then ar.age when ar.resultuid <= rm.resultuid then ? else ? end as age, case when ar.relevant is not null then ar.relevant when ar.resultuid <= rm.resultuid then ? else ? end as relevant from ar inner join symbols s on ar.symbolid = s.symbolid and s.nonliquid = ? inner join brokersymbollist bsl on bsl.brokerid = ? and bsl.symbolid = s.symbolid inner join symbolgroup sg on bsl.symbolid = sg.symbolid inner join groups g on sg.groupid = g.groupid inner join brokergroups bg on g.groupid = bg.groupid and bsl.brokerid = bg.brokerid inner join downloadersymbolsettings dss on sg.symbolid = dss.symbolid inner join datafeedstimetable dtt on dss.classname = dtt.classname and dtt.dayofweek = ? inner join rar_max rm on ? = ? left outer join autochartist_symbolupdates au on dss.symbolid = au.symbolid left outer join currencypips cp on s.symbol = cp.symbol left join lateral calc_cp_signal (ar.resultuid) newlevels on true left outer join brokerinstrumentmap bim on dss.datafeedinstrumentid = bim.datafeedinstrumentid and bim.brokerid = bsl.brokerid and bim.type = ? where ar.gmttimefound > now() - interval ? and dss.enabled = ? and s.deleted = ? and (ar.simulation = ? or ar.simulation is null) and (? = ? or s.timegranularity in (...)) and (? = ? or s.exchange in (...)) and (? = ? or coalesce(bim.code, s.symbol) in (...)) and (? = ? or ar.pattern in (...)) and (? = ? or (? = ? and ar.breakout >= ?) or (? = ? and ar.breakout < ?)) and (? = ? or ar.patternlengthbars <= ?) and newlevels.filtered = false and ar.patternstarttime >= coalesce(au.earliestpricedatetime, ?::timestamp without time zone) -- to make sure patternstarttime is in our t-tables ), results as ( select distinct on (symbolid) * from all_results where (false = ? or relevant = ?) and (? = ? or age <= ?) order by symbolid, resultuid ) select * from results order by identified desc, length desc ;-
SELECT resultuid FROM relevance_autochartist_results ORDER BY resultuid DESC LIMIT 1), ar AS ( SELECT a.*, rr.age, rr.relevant from autochartist_results a LEFT OUTER JOIN relevance_autochartist_results rr on a.resultuid = rr.resultuid WHERE CASE WHEN FALSE = $1 THEN true ELSE a.resultuid > ( select min(resultuid) from relevance_autochartist_results) END), all_results AS ( SELECT ar.resultuid AS resultuid, ar.direction AS direction, ar.predictiontimeto AS predictiontimeto, ar.predictionpricefrom AS predictionpricefrom, ar.predictionpriceto AS predictionpriceto, cp.pip AS pip, s.exchange AS exchange, s.symbolid AS symbolid, coalesce(bim.code, s.symbol) AS symbol_code, s.longname AS symbol_name, s.timegranularity AS interval, ar.pattern AS pattern_name, ar.breakout AS breakout, ar.patternendtime AS identified, dtt.timezone AS timezone, ar.patternlengthbars AS length, g.basegroupname, newLevels.profit, newLevels.stop, newLevels.filtered, CASE WHEN ar.age IS NOT NULL THEN ar.age WHEN ar.resultuid <= rm.resultuid THEN 11 ELSE 0 END as age, CASE WHEN ar.relevant IS NOT NULL THEN ar.relevant WHEN ar.resultuid <= rm.resultuid THEN 0 ELSE 1 END as relevant FROM ar INNER JOIN symbols s ON ar.symbolid = s.symbolid AND s.nonliquid = 0 INNER JOIN brokersymbollist bsl ON bsl.brokerid = $2 AND bsl.symbolid = s.symbolid INNER JOIN symbolgroup sg on bsl.symbolid = sg.symbolid INNER JOIN groups g ON sg.groupid = g.groupid INNER JOIN brokergroups bg on g.groupid = bg.groupid AND bsl.brokerid = bg.brokerid INNER JOIN downloadersymbolsettings dss ON sg.symbolid = dss.symbolid INNER JOIN datafeedstimetable dtt ON dss.classname = dtt.classname AND dtt.dayofweek = 3 INNER JOIN rar_max rm ON 1 = 1 LEFT OUTER JOIN autochartist_symbolupdates au on dss.symbolid = au.symbolid LEFT OUTER JOIN currencypips cp ON s.symbol = cp.symbol LEFT JOIN LATERAL calc_cp_signal (ar.resultuid) newLevels on true LEFT OUTER JOIN brokerinstrumentmap bim ON dss.datafeedinstrumentid = bim.datafeedinstrumentid AND bim.brokerid = bsl.brokerid AND bim.TYPE = 'OUTBOUND' WHERE ar.gmttimefound > now() - INTERVAL '7 DAYS' AND dss.enabled = 1 AND s.deleted = 0 AND (ar.simulation = 0 OR ar.simulation IS NULL) AND ($3 = 0 OR s.timegranularity in ($4)) AND ($5 = 0 OR s.exchange in ($6)) AND ($7 = 0 OR coalesce(bim.code, s.symbol) in ($8)) AND ($9 = 0 OR ar.pattern in ($10)) AND ($11 = 0 OR ($12 = 1 AND ar.breakout >= 0) OR ($13 = 2 AND ar.breakout < 0)) AND ($14 = 0 OR ar.patternlengthbars <= $15) and newLevels.filtered = false AND ar.patternstarttime >= coalesce(au.earliestpricedatetime, '1900-01-01'::timestamp without time zone) -- To make sure patternstarttime is in our t-tables ), results AS ( SELECT DISTINCT ON (symbolid) * FROM all_results WHERE (FALSE = $16 OR relevant = 1) AND ($17 = 0 OR age <= $18) ORDER BY symbolid, resultuid ) SELECT * from results ORDER BY identified DESC, length DESC;
Date: 2026-01-12 01:02:11 Duration: 0ms
4 16 617.25 MiB 38.58 MiB 38.58 MiB 38.58 MiB update solr_relevance_old set new_hod_correct = sub.hod_correct, new_hod_percent = sub.hod_percent, new_hod_total = sub.hod_total, new_pattern_correct = sub.pattern_correct, new_pattern_percent = sub.pattern_percent, new_pattern_total = sub.pattern_total, new_percent = sub.percent, new_symbol_correct = sub.symbol_correct, new_symbol_percent = sub.symbol_percent, new_symbol_total = sub.symbol_total from ( select distinct resultuid, hod_correct, hod_percent, hod_total, hod, pattern_correct, pattern_percent, pattern_total, percent, symbol_correct, symbol_percent, symbol_total from whatshot_probability where type = ?) sub where result_uid = sub.resultuid;-
UPDATE solr_relevance_old SET new_hod_correct = sub.hod_correct, new_hod_percent = sub.hod_percent, new_hod_total = sub.hod_total, new_pattern_correct = sub.pattern_correct, new_pattern_percent = sub.pattern_percent, new_pattern_total = sub.pattern_total, new_percent = sub.percent, new_symbol_correct = sub.symbol_correct, new_symbol_percent = sub.symbol_percent, new_symbol_total = sub.symbol_total FROM ( select distinct resultuid, hod_correct, hod_percent, hod_total, hod, pattern_correct, pattern_percent, pattern_total, percent, symbol_correct, symbol_percent, symbol_total FROM whatshot_probability WHERE type = 'cp') sub WHERE result_uid = sub.resultuid;
Date: 2026-01-12 01:01:14 Duration: 0ms
5 16 1.11 GiB 70.80 MiB 70.80 MiB 70.80 MiB with max_ra as ( select resultuid from relevance_keylevels_results order by resultuid desc limit ?) update solr_relevance_old set newrelevant = sub.relevant, newage = sub.age from ( select so.uuid, case when ra.relevant is not null then ra.relevant when so.result_uid < max_ra.resultuid then ? else ? end as relevant, case when ra.age is not null then ra.age when so.result_uid < max_ra.resultuid then ? else ? end as age, so.result_uid from max_ra, solr_relevance_old so inner join keylevels_results k on so.result_uid = k.resultuid and so.uuid ilike ? inner join downloadersymbolsettings dss on k.symbolid = dss.symbolid left outer join relevance_keylevels_results ra on so.result_uid = ra.resultuid and so.uuid ilike ?) sub where solr_relevance_old.result_uid = sub.result_uid and solr_relevance_old.uuid ilike ?; update solr_relevance_old set newrelevant = ? where result_uid in ( select result_uid from solr_relevance_old s left outer join keylevels_results a on a.resultuid = s.result_uid where s.uuid ilike ? and a.resultuid is null); update solr_relevance_old set new_hod_correct = sub.hod_correct, new_hod_percent = sub.hod_percent, new_hod_total = sub.hod_total, new_pattern_correct = sub.pattern_correct, new_pattern_percent = sub.pattern_percent, new_pattern_total = sub.pattern_total, new_percent = sub.percent, new_symbol_correct = sub.symbol_correct, new_symbol_percent = sub.symbol_percent, new_symbol_total = sub.symbol_total from ( select distinct resultuid, hod_correct, hod_percent, hod_total, hod, pattern_correct, pattern_percent, pattern_total, percent, symbol_correct, symbol_percent, symbol_total from whatshot_probability where type in (...)) sub where result_uid = sub.resultuid;-
with max_ra as ( select resultuid from relevance_keylevels_results order by resultuid desc limit 1) update solr_relevance_old set newrelevant = sub.relevant, newage = sub.age from ( select so.uuid, case when ra.relevant is not null then ra.relevant when so.result_uid < max_ra.resultuid then 0 else 1 end as relevant, case when ra.age is not null then ra.age when so.result_uid < max_ra.resultuid then 11 else 0 end as age, so.result_uid from max_ra, solr_relevance_old so inner join keylevels_results k on so.result_uid = k.resultuid and so.uuid ilike 'kl_%' inner join downloadersymbolsettings dss on k.symbolid = dss.symbolid left outer join relevance_keylevels_results ra on so.result_uid = ra.resultuid and so.uuid ilike 'kl_%') sub where solr_relevance_old.result_uid = sub.result_uid and solr_relevance_old.uuid ilike 'kl_%'; update solr_relevance_old set newrelevant = 0 where result_uid in ( select result_uid from solr_relevance_old s left outer join keylevels_results a on a.resultuid = s.result_uid where s.uuid ilike 'kl_%' and a.resultuid is null); UPDATE solr_relevance_old SET new_hod_correct = sub.hod_correct, new_hod_percent = sub.hod_percent, new_hod_total = sub.hod_total, new_pattern_correct = sub.pattern_correct, new_pattern_percent = sub.pattern_percent, new_pattern_total = sub.pattern_total, new_percent = sub.percent, new_symbol_correct = sub.symbol_correct, new_symbol_percent = sub.symbol_percent, new_symbol_total = sub.symbol_total FROM ( select distinct resultuid, hod_correct, hod_percent, hod_total, hod, pattern_correct, pattern_percent, pattern_total, percent, symbol_correct, symbol_percent, symbol_total FROM whatshot_probability WHERE type in ('kl', 'ekl')) sub WHERE result_uid = sub.resultuid;
Date: 2026-01-12 01:01:17 Duration: 0ms
6 4 490.72 MiB 122.64 MiB 122.71 MiB 122.68 MiB select updateresultsmaterializedview ();-
select updateresultsmaterializedview ();
Date: 2026-01-12 01:02:23 Duration: 0ms
7 4 216.19 MiB 54.01 MiB 54.08 MiB 54.05 MiB select updateageforrelevantresults ();-
select updateageforrelevantresults ();
Date: 2026-01-12 01:02:10 Duration: 0ms
Queries generating the largest temporary files
Rank Size Query 1 188.66 MiB with rankedmt4 as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors ), last_feed_entry as ( select * from rankedmt4 where r = 1 ), ok_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where status = 'OK' ), earliest_entry_after_ok as ( select m.datafeedname, min(m.eventtimestamp) as eventtimestamp from mt4datafeederrors m left outer join ( select datafeedname, eventtimestamp from ok_entries where r = 1) oo on m.datafeedname = oo.datafeedname where m.eventtimestamp > coalesce(oo.eventtimestamp, '1900-01-01'::timestamp without time zone) group by m.datafeedname ), notified_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where notified is not null and notified <> '' ), broker as ( select *, row_number() over (partition by feedname order by brokerid) r from ( select distinct b.brokerid, b.name as brokername, dss.classname as feedname from downloadersymbolsettings dss inner join brokersymbollist bsl on dss.symbolid = bsl.symbolid inner join broker b on bsl.brokerid = b.brokerid where dss.enabled = 1) a ) select last.id, last.datafeedname, last.eventtimestamp, last.status, last.errordescription, last.serveraddress, last.username, note.notified, note.eventtimestamp, broker.brokername from last_feed_entry last inner join earliest_entry_after_ok after_ok on last.datafeedname = after_ok.datafeedname inner join broker on last.datafeedname = broker.feedname left outer join ok_entries ok on ok.datafeedname = last.datafeedname left outer join notified_entries note on note.datafeedname = last.datafeedname and note.r = 1 where (ok.r is null or ok.r = 1) and last.datafeedname not in ( select distinct datafeedname from last_feed_entry where status = 'OK') and extract(epoch from (last.eventtimestamp - after_ok.eventtimestamp)) > 60 * 60 and last.eventtimestamp > current_timestamp - interval '1 day' and (note.eventtimestamp is null or note.eventtimestamp < current_timestamp - interval '10 hours') and last.eventtimestamp > current_timestamp - interval '1 hour' and broker.r = 1;[ Date: 2026-01-12 01:40:05 ]
2 148.80 MiB with rankedmt4 as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors ), last_feed_entry as ( select * from rankedmt4 where r = 1 ), ok_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where status = 'OK' ), earliest_entry_after_ok as ( select m.datafeedname, min(m.eventtimestamp) as eventtimestamp from mt4datafeederrors m left outer join ( select datafeedname, eventtimestamp from ok_entries where r = 1) oo on m.datafeedname = oo.datafeedname where m.eventtimestamp > coalesce(oo.eventtimestamp, '1900-01-01'::timestamp without time zone) group by m.datafeedname ), notified_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where notified is not null and notified <> '' ), broker as ( select *, row_number() over (partition by feedname order by brokerid) r from ( select distinct b.brokerid, b.name as brokername, dss.classname as feedname from downloadersymbolsettings dss inner join brokersymbollist bsl on dss.symbolid = bsl.symbolid inner join broker b on bsl.brokerid = b.brokerid where dss.enabled = 1) a ) select last.id, last.datafeedname, last.eventtimestamp, last.status, last.errordescription, last.serveraddress, last.username, note.notified, note.eventtimestamp, broker.brokername from last_feed_entry last inner join earliest_entry_after_ok after_ok on last.datafeedname = after_ok.datafeedname inner join broker on last.datafeedname = broker.feedname left outer join ok_entries ok on ok.datafeedname = last.datafeedname left outer join notified_entries note on note.datafeedname = last.datafeedname and note.r = 1 where (ok.r is null or ok.r = 1) and last.datafeedname not in ( select distinct datafeedname from last_feed_entry where status = 'OK') and extract(epoch from (last.eventtimestamp - after_ok.eventtimestamp)) > 60 * 60 and last.eventtimestamp > current_timestamp - interval '1 day' and (note.eventtimestamp is null or note.eventtimestamp < current_timestamp - interval '10 hours') and last.eventtimestamp > current_timestamp - interval '1 hour' and broker.r = 1;[ Date: 2026-01-12 01:20:05 ]
3 122.71 MiB select updateresultsmaterializedview ();[ Date: 2026-01-12 01:47:14 ]
4 122.70 MiB select updateresultsmaterializedview ();[ Date: 2026-01-12 01:32:16 ]
5 122.67 MiB select updateresultsmaterializedview ();[ Date: 2026-01-12 01:17:15 ]
6 122.64 MiB select updateresultsmaterializedview ();[ Date: 2026-01-12 01:02:23 ]
7 114.37 MiB with rankedmt4 as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors ), last_feed_entry as ( select * from rankedmt4 where r = 1 ), ok_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where status = 'OK' ), earliest_entry_after_ok as ( select m.datafeedname, min(m.eventtimestamp) as eventtimestamp from mt4datafeederrors m left outer join ( select datafeedname, eventtimestamp from ok_entries where r = 1) oo on m.datafeedname = oo.datafeedname where m.eventtimestamp > coalesce(oo.eventtimestamp, '1900-01-01'::timestamp without time zone) group by m.datafeedname ), notified_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where notified is not null and notified <> '' ), broker as ( select *, row_number() over (partition by feedname order by brokerid) r from ( select distinct b.brokerid, b.name as brokername, dss.classname as feedname from downloadersymbolsettings dss inner join brokersymbollist bsl on dss.symbolid = bsl.symbolid inner join broker b on bsl.brokerid = b.brokerid where dss.enabled = 1) a ) select last.id, last.datafeedname, last.eventtimestamp, last.status, last.errordescription, last.serveraddress, last.username, note.notified, note.eventtimestamp, broker.brokername from last_feed_entry last inner join earliest_entry_after_ok after_ok on last.datafeedname = after_ok.datafeedname inner join broker on last.datafeedname = broker.feedname left outer join ok_entries ok on ok.datafeedname = last.datafeedname left outer join notified_entries note on note.datafeedname = last.datafeedname and note.r = 1 where (ok.r is null or ok.r = 1) and last.datafeedname not in ( select distinct datafeedname from last_feed_entry where status = 'OK') and extract(epoch from (last.eventtimestamp - after_ok.eventtimestamp)) > 60 * 60 and last.eventtimestamp > current_timestamp - interval '1 day' and (note.eventtimestamp is null or note.eventtimestamp < current_timestamp - interval '10 hours') and last.eventtimestamp > current_timestamp - interval '1 hour' and broker.r = 1;[ Date: 2026-01-12 01:30:03 ]
8 99.52 MiB with rankedmt4 as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors ), last_feed_entry as ( select * from rankedmt4 where r = 1 ), ok_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where status = 'OK' ), earliest_entry_after_ok as ( select m.datafeedname, min(m.eventtimestamp) as eventtimestamp from mt4datafeederrors m left outer join ( select datafeedname, eventtimestamp from ok_entries where r = 1) oo on m.datafeedname = oo.datafeedname where m.eventtimestamp > coalesce(oo.eventtimestamp, '1900-01-01'::timestamp without time zone) group by m.datafeedname ), notified_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where notified is not null and notified <> '' ), broker as ( select *, row_number() over (partition by feedname order by brokerid) r from ( select distinct b.brokerid, b.name as brokername, dss.classname as feedname from downloadersymbolsettings dss inner join brokersymbollist bsl on dss.symbolid = bsl.symbolid inner join broker b on bsl.brokerid = b.brokerid where dss.enabled = 1) a ) select last.id, last.datafeedname, last.eventtimestamp, last.status, last.errordescription, last.serveraddress, last.username, note.notified, note.eventtimestamp, broker.brokername from last_feed_entry last inner join earliest_entry_after_ok after_ok on last.datafeedname = after_ok.datafeedname inner join broker on last.datafeedname = broker.feedname left outer join ok_entries ok on ok.datafeedname = last.datafeedname left outer join notified_entries note on note.datafeedname = last.datafeedname and note.r = 1 where (ok.r is null or ok.r = 1) and last.datafeedname not in ( select distinct datafeedname from last_feed_entry where status = 'OK') and extract(epoch from (last.eventtimestamp - after_ok.eventtimestamp)) > 60 * 60 and last.eventtimestamp > current_timestamp - interval '1 day' and (note.eventtimestamp is null or note.eventtimestamp < current_timestamp - interval '10 hours') and last.eventtimestamp > current_timestamp - interval '1 hour' and broker.r = 1;[ Date: 2026-01-12 01:00:05 ]
9 96.40 MiB with rankedmt4 as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors ), last_feed_entry as ( select * from rankedmt4 where r = 1 ), ok_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where status = 'OK' ), earliest_entry_after_ok as ( select m.datafeedname, min(m.eventtimestamp) as eventtimestamp from mt4datafeederrors m left outer join ( select datafeedname, eventtimestamp from ok_entries where r = 1) oo on m.datafeedname = oo.datafeedname where m.eventtimestamp > coalesce(oo.eventtimestamp, '1900-01-01'::timestamp without time zone) group by m.datafeedname ), notified_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where notified is not null and notified <> '' ), broker as ( select *, row_number() over (partition by feedname order by brokerid) r from ( select distinct b.brokerid, b.name as brokername, dss.classname as feedname from downloadersymbolsettings dss inner join brokersymbollist bsl on dss.symbolid = bsl.symbolid inner join broker b on bsl.brokerid = b.brokerid where dss.enabled = 1) a ) select last.id, last.datafeedname, last.eventtimestamp, last.status, last.errordescription, last.serveraddress, last.username, note.notified, note.eventtimestamp, broker.brokername from last_feed_entry last inner join earliest_entry_after_ok after_ok on last.datafeedname = after_ok.datafeedname inner join broker on last.datafeedname = broker.feedname left outer join ok_entries ok on ok.datafeedname = last.datafeedname left outer join notified_entries note on note.datafeedname = last.datafeedname and note.r = 1 where (ok.r is null or ok.r = 1) and last.datafeedname not in ( select distinct datafeedname from last_feed_entry where status = 'OK') and extract(epoch from (last.eventtimestamp - after_ok.eventtimestamp)) > 60 * 60 and last.eventtimestamp > current_timestamp - interval '1 day' and (note.eventtimestamp is null or note.eventtimestamp < current_timestamp - interval '10 hours') and last.eventtimestamp > current_timestamp - interval '1 hour' and broker.r = 1;[ Date: 2026-01-12 01:10:04 ]
10 85.11 MiB with rankedmt4 as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors ), last_feed_entry as ( select * from rankedmt4 where r = 1 ), ok_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where status = 'OK' ), earliest_entry_after_ok as ( select m.datafeedname, min(m.eventtimestamp) as eventtimestamp from mt4datafeederrors m left outer join ( select datafeedname, eventtimestamp from ok_entries where r = 1) oo on m.datafeedname = oo.datafeedname where m.eventtimestamp > coalesce(oo.eventtimestamp, '1900-01-01'::timestamp without time zone) group by m.datafeedname ), notified_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where notified is not null and notified <> '' ), broker as ( select *, row_number() over (partition by feedname order by brokerid) r from ( select distinct b.brokerid, b.name as brokername, dss.classname as feedname from downloadersymbolsettings dss inner join brokersymbollist bsl on dss.symbolid = bsl.symbolid inner join broker b on bsl.brokerid = b.brokerid where dss.enabled = 1) a ) select last.id, last.datafeedname, last.eventtimestamp, last.status, last.errordescription, last.serveraddress, last.username, note.notified, note.eventtimestamp, broker.brokername from last_feed_entry last inner join earliest_entry_after_ok after_ok on last.datafeedname = after_ok.datafeedname inner join broker on last.datafeedname = broker.feedname left outer join ok_entries ok on ok.datafeedname = last.datafeedname left outer join notified_entries note on note.datafeedname = last.datafeedname and note.r = 1 where (ok.r is null or ok.r = 1) and last.datafeedname not in ( select distinct datafeedname from last_feed_entry where status = 'OK') and extract(epoch from (last.eventtimestamp - after_ok.eventtimestamp)) > 60 * 60 and last.eventtimestamp > current_timestamp - interval '1 day' and (note.eventtimestamp is null or note.eventtimestamp < current_timestamp - interval '10 hours') and last.eventtimestamp > current_timestamp - interval '1 hour' and broker.r = 1;[ Date: 2026-01-12 01:00:04 ]
11 85.09 MiB with rankedmt4 as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors ), last_feed_entry as ( select * from rankedmt4 where r = 1 ), ok_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where status = 'OK' ), earliest_entry_after_ok as ( select m.datafeedname, min(m.eventtimestamp) as eventtimestamp from mt4datafeederrors m left outer join ( select datafeedname, eventtimestamp from ok_entries where r = 1) oo on m.datafeedname = oo.datafeedname where m.eventtimestamp > coalesce(oo.eventtimestamp, '1900-01-01'::timestamp without time zone) group by m.datafeedname ), notified_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where notified is not null and notified <> '' ), broker as ( select *, row_number() over (partition by feedname order by brokerid) r from ( select distinct b.brokerid, b.name as brokername, dss.classname as feedname from downloadersymbolsettings dss inner join brokersymbollist bsl on dss.symbolid = bsl.symbolid inner join broker b on bsl.brokerid = b.brokerid where dss.enabled = 1) a ) select last.id, last.datafeedname, last.eventtimestamp, last.status, last.errordescription, last.serveraddress, last.username, note.notified, note.eventtimestamp, broker.brokername from last_feed_entry last inner join earliest_entry_after_ok after_ok on last.datafeedname = after_ok.datafeedname inner join broker on last.datafeedname = broker.feedname left outer join ok_entries ok on ok.datafeedname = last.datafeedname left outer join notified_entries note on note.datafeedname = last.datafeedname and note.r = 1 where (ok.r is null or ok.r = 1) and last.datafeedname not in ( select distinct datafeedname from last_feed_entry where status = 'OK') and extract(epoch from (last.eventtimestamp - after_ok.eventtimestamp)) > 60 * 60 and last.eventtimestamp > current_timestamp - interval '1 day' and (note.eventtimestamp is null or note.eventtimestamp < current_timestamp - interval '10 hours') and last.eventtimestamp > current_timestamp - interval '1 hour' and broker.r = 1;[ Date: 2026-01-12 01:50:03 ]
12 85.05 MiB with rankedmt4 as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors ), last_feed_entry as ( select * from rankedmt4 where r = 1 ), ok_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where status = 'OK' ), earliest_entry_after_ok as ( select m.datafeedname, min(m.eventtimestamp) as eventtimestamp from mt4datafeederrors m left outer join ( select datafeedname, eventtimestamp from ok_entries where r = 1) oo on m.datafeedname = oo.datafeedname where m.eventtimestamp > coalesce(oo.eventtimestamp, '1900-01-01'::timestamp without time zone) group by m.datafeedname ), notified_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where notified is not null and notified <> '' ), broker as ( select *, row_number() over (partition by feedname order by brokerid) r from ( select distinct b.brokerid, b.name as brokername, dss.classname as feedname from downloadersymbolsettings dss inner join brokersymbollist bsl on dss.symbolid = bsl.symbolid inner join broker b on bsl.brokerid = b.brokerid where dss.enabled = 1) a ) select last.id, last.datafeedname, last.eventtimestamp, last.status, last.errordescription, last.serveraddress, last.username, note.notified, note.eventtimestamp, broker.brokername from last_feed_entry last inner join earliest_entry_after_ok after_ok on last.datafeedname = after_ok.datafeedname inner join broker on last.datafeedname = broker.feedname left outer join ok_entries ok on ok.datafeedname = last.datafeedname left outer join notified_entries note on note.datafeedname = last.datafeedname and note.r = 1 where (ok.r is null or ok.r = 1) and last.datafeedname not in ( select distinct datafeedname from last_feed_entry where status = 'OK') and extract(epoch from (last.eventtimestamp - after_ok.eventtimestamp)) > 60 * 60 and last.eventtimestamp > current_timestamp - interval '1 day' and (note.eventtimestamp is null or note.eventtimestamp < current_timestamp - interval '10 hours') and last.eventtimestamp > current_timestamp - interval '1 hour' and broker.r = 1;[ Date: 2026-01-12 01:50:03 ]
13 72.46 MiB with rankedmt4 as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors ), last_feed_entry as ( select * from rankedmt4 where r = 1 ), ok_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where status = 'OK' ), earliest_entry_after_ok as ( select m.datafeedname, min(m.eventtimestamp) as eventtimestamp from mt4datafeederrors m left outer join ( select datafeedname, eventtimestamp from ok_entries where r = 1) oo on m.datafeedname = oo.datafeedname where m.eventtimestamp > coalesce(oo.eventtimestamp, '1900-01-01'::timestamp without time zone) group by m.datafeedname ), notified_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where notified is not null and notified <> '' ), broker as ( select *, row_number() over (partition by feedname order by brokerid) r from ( select distinct b.brokerid, b.name as brokername, dss.classname as feedname from downloadersymbolsettings dss inner join brokersymbollist bsl on dss.symbolid = bsl.symbolid inner join broker b on bsl.brokerid = b.brokerid where dss.enabled = 1) a ) select last.id, last.datafeedname, last.eventtimestamp, last.status, last.errordescription, last.serveraddress, last.username, note.notified, note.eventtimestamp, broker.brokername from last_feed_entry last inner join earliest_entry_after_ok after_ok on last.datafeedname = after_ok.datafeedname inner join broker on last.datafeedname = broker.feedname left outer join ok_entries ok on ok.datafeedname = last.datafeedname left outer join notified_entries note on note.datafeedname = last.datafeedname and note.r = 1 where (ok.r is null or ok.r = 1) and last.datafeedname not in ( select distinct datafeedname from last_feed_entry where status = 'OK') and extract(epoch from (last.eventtimestamp - after_ok.eventtimestamp)) > 60 * 60 and last.eventtimestamp > current_timestamp - interval '1 day' and (note.eventtimestamp is null or note.eventtimestamp < current_timestamp - interval '10 hours') and last.eventtimestamp > current_timestamp - interval '1 hour' and broker.r = 1;[ Date: 2026-01-12 01:30:05 ]
14 70.80 MiB with max_ra as ( select resultuid from relevance_keylevels_results order by resultuid desc limit 1) update solr_relevance_old set newrelevant = sub.relevant, newage = sub.age from ( select so.uuid, case when ra.relevant is not null then ra.relevant when so.result_uid < max_ra.resultuid then 0 else 1 end as relevant, case when ra.age is not null then ra.age when so.result_uid < max_ra.resultuid then 11 else 0 end as age, so.result_uid from max_ra, solr_relevance_old so inner join keylevels_results k on so.result_uid = k.resultuid and so.uuid ilike 'kl_%' inner join downloadersymbolsettings dss on k.symbolid = dss.symbolid left outer join relevance_keylevels_results ra on so.result_uid = ra.resultuid and so.uuid ilike 'kl_%') sub where solr_relevance_old.result_uid = sub.result_uid and solr_relevance_old.uuid ilike 'kl_%'; update solr_relevance_old set newrelevant = 0 where result_uid in ( select result_uid from solr_relevance_old s left outer join keylevels_results a on a.resultuid = s.result_uid where s.uuid ilike 'kl_%' and a.resultuid is null); UPDATE solr_relevance_old SET new_hod_correct = sub.hod_correct, new_hod_percent = sub.hod_percent, new_hod_total = sub.hod_total, new_pattern_correct = sub.pattern_correct, new_pattern_percent = sub.pattern_percent, new_pattern_total = sub.pattern_total, new_percent = sub.percent, new_symbol_correct = sub.symbol_correct, new_symbol_percent = sub.symbol_percent, new_symbol_total = sub.symbol_total FROM ( select distinct resultuid, hod_correct, hod_percent, hod_total, hod, pattern_correct, pattern_percent, pattern_total, percent, symbol_correct, symbol_percent, symbol_total FROM whatshot_probability WHERE type in ('kl', 'ekl')) sub WHERE result_uid = sub.resultuid;[ Date: 2026-01-12 01:01:17 ]
15 70.80 MiB with max_ra as ( select resultuid from relevance_keylevels_results order by resultuid desc limit 1) update solr_relevance_old set newrelevant = sub.relevant, newage = sub.age from ( select so.uuid, case when ra.relevant is not null then ra.relevant when so.result_uid < max_ra.resultuid then 0 else 1 end as relevant, case when ra.age is not null then ra.age when so.result_uid < max_ra.resultuid then 11 else 0 end as age, so.result_uid from max_ra, solr_relevance_old so inner join keylevels_results k on so.result_uid = k.resultuid and so.uuid ilike 'kl_%' inner join downloadersymbolsettings dss on k.symbolid = dss.symbolid left outer join relevance_keylevels_results ra on so.result_uid = ra.resultuid and so.uuid ilike 'kl_%') sub where solr_relevance_old.result_uid = sub.result_uid and solr_relevance_old.uuid ilike 'kl_%'; update solr_relevance_old set newrelevant = 0 where result_uid in ( select result_uid from solr_relevance_old s left outer join keylevels_results a on a.resultuid = s.result_uid where s.uuid ilike 'kl_%' and a.resultuid is null); UPDATE solr_relevance_old SET new_hod_correct = sub.hod_correct, new_hod_percent = sub.hod_percent, new_hod_total = sub.hod_total, new_pattern_correct = sub.pattern_correct, new_pattern_percent = sub.pattern_percent, new_pattern_total = sub.pattern_total, new_percent = sub.percent, new_symbol_correct = sub.symbol_correct, new_symbol_percent = sub.symbol_percent, new_symbol_total = sub.symbol_total FROM ( select distinct resultuid, hod_correct, hod_percent, hod_total, hod, pattern_correct, pattern_percent, pattern_total, percent, symbol_correct, symbol_percent, symbol_total FROM whatshot_probability WHERE type in ('kl', 'ekl')) sub WHERE result_uid = sub.resultuid;[ Date: 2026-01-12 01:03:15 ]
16 70.80 MiB with max_ra as ( select resultuid from relevance_keylevels_results order by resultuid desc limit 1) update solr_relevance_old set newrelevant = sub.relevant, newage = sub.age from ( select so.uuid, case when ra.relevant is not null then ra.relevant when so.result_uid < max_ra.resultuid then 0 else 1 end as relevant, case when ra.age is not null then ra.age when so.result_uid < max_ra.resultuid then 11 else 0 end as age, so.result_uid from max_ra, solr_relevance_old so inner join keylevels_results k on so.result_uid = k.resultuid and so.uuid ilike 'kl_%' inner join downloadersymbolsettings dss on k.symbolid = dss.symbolid left outer join relevance_keylevels_results ra on so.result_uid = ra.resultuid and so.uuid ilike 'kl_%') sub where solr_relevance_old.result_uid = sub.result_uid and solr_relevance_old.uuid ilike 'kl_%'; update solr_relevance_old set newrelevant = 0 where result_uid in ( select result_uid from solr_relevance_old s left outer join keylevels_results a on a.resultuid = s.result_uid where s.uuid ilike 'kl_%' and a.resultuid is null); UPDATE solr_relevance_old SET new_hod_correct = sub.hod_correct, new_hod_percent = sub.hod_percent, new_hod_total = sub.hod_total, new_pattern_correct = sub.pattern_correct, new_pattern_percent = sub.pattern_percent, new_pattern_total = sub.pattern_total, new_percent = sub.percent, new_symbol_correct = sub.symbol_correct, new_symbol_percent = sub.symbol_percent, new_symbol_total = sub.symbol_total FROM ( select distinct resultuid, hod_correct, hod_percent, hod_total, hod, pattern_correct, pattern_percent, pattern_total, percent, symbol_correct, symbol_percent, symbol_total FROM whatshot_probability WHERE type in ('kl', 'ekl')) sub WHERE result_uid = sub.resultuid;[ Date: 2026-01-12 01:05:15 ]
17 70.80 MiB with max_ra as ( select resultuid from relevance_keylevels_results order by resultuid desc limit 1) update solr_relevance_old set newrelevant = sub.relevant, newage = sub.age from ( select so.uuid, case when ra.relevant is not null then ra.relevant when so.result_uid < max_ra.resultuid then 0 else 1 end as relevant, case when ra.age is not null then ra.age when so.result_uid < max_ra.resultuid then 11 else 0 end as age, so.result_uid from max_ra, solr_relevance_old so inner join keylevels_results k on so.result_uid = k.resultuid and so.uuid ilike 'kl_%' inner join downloadersymbolsettings dss on k.symbolid = dss.symbolid left outer join relevance_keylevels_results ra on so.result_uid = ra.resultuid and so.uuid ilike 'kl_%') sub where solr_relevance_old.result_uid = sub.result_uid and solr_relevance_old.uuid ilike 'kl_%'; update solr_relevance_old set newrelevant = 0 where result_uid in ( select result_uid from solr_relevance_old s left outer join keylevels_results a on a.resultuid = s.result_uid where s.uuid ilike 'kl_%' and a.resultuid is null); UPDATE solr_relevance_old SET new_hod_correct = sub.hod_correct, new_hod_percent = sub.hod_percent, new_hod_total = sub.hod_total, new_pattern_correct = sub.pattern_correct, new_pattern_percent = sub.pattern_percent, new_pattern_total = sub.pattern_total, new_percent = sub.percent, new_symbol_correct = sub.symbol_correct, new_symbol_percent = sub.symbol_percent, new_symbol_total = sub.symbol_total FROM ( select distinct resultuid, hod_correct, hod_percent, hod_total, hod, pattern_correct, pattern_percent, pattern_total, percent, symbol_correct, symbol_percent, symbol_total FROM whatshot_probability WHERE type in ('kl', 'ekl')) sub WHERE result_uid = sub.resultuid;[ Date: 2026-01-12 01:11:18 ]
18 70.80 MiB with max_ra as ( select resultuid from relevance_keylevels_results order by resultuid desc limit 1) update solr_relevance_old set newrelevant = sub.relevant, newage = sub.age from ( select so.uuid, case when ra.relevant is not null then ra.relevant when so.result_uid < max_ra.resultuid then 0 else 1 end as relevant, case when ra.age is not null then ra.age when so.result_uid < max_ra.resultuid then 11 else 0 end as age, so.result_uid from max_ra, solr_relevance_old so inner join keylevels_results k on so.result_uid = k.resultuid and so.uuid ilike 'kl_%' inner join downloadersymbolsettings dss on k.symbolid = dss.symbolid left outer join relevance_keylevels_results ra on so.result_uid = ra.resultuid and so.uuid ilike 'kl_%') sub where solr_relevance_old.result_uid = sub.result_uid and solr_relevance_old.uuid ilike 'kl_%'; update solr_relevance_old set newrelevant = 0 where result_uid in ( select result_uid from solr_relevance_old s left outer join keylevels_results a on a.resultuid = s.result_uid where s.uuid ilike 'kl_%' and a.resultuid is null); UPDATE solr_relevance_old SET new_hod_correct = sub.hod_correct, new_hod_percent = sub.hod_percent, new_hod_total = sub.hod_total, new_pattern_correct = sub.pattern_correct, new_pattern_percent = sub.pattern_percent, new_pattern_total = sub.pattern_total, new_percent = sub.percent, new_symbol_correct = sub.symbol_correct, new_symbol_percent = sub.symbol_percent, new_symbol_total = sub.symbol_total FROM ( select distinct resultuid, hod_correct, hod_percent, hod_total, hod, pattern_correct, pattern_percent, pattern_total, percent, symbol_correct, symbol_percent, symbol_total FROM whatshot_probability WHERE type in ('kl', 'ekl')) sub WHERE result_uid = sub.resultuid;[ Date: 2026-01-12 01:16:17 ]
19 70.80 MiB with max_ra as ( select resultuid from relevance_keylevels_results order by resultuid desc limit 1) update solr_relevance_old set newrelevant = sub.relevant, newage = sub.age from ( select so.uuid, case when ra.relevant is not null then ra.relevant when so.result_uid < max_ra.resultuid then 0 else 1 end as relevant, case when ra.age is not null then ra.age when so.result_uid < max_ra.resultuid then 11 else 0 end as age, so.result_uid from max_ra, solr_relevance_old so inner join keylevels_results k on so.result_uid = k.resultuid and so.uuid ilike 'kl_%' inner join downloadersymbolsettings dss on k.symbolid = dss.symbolid left outer join relevance_keylevels_results ra on so.result_uid = ra.resultuid and so.uuid ilike 'kl_%') sub where solr_relevance_old.result_uid = sub.result_uid and solr_relevance_old.uuid ilike 'kl_%'; update solr_relevance_old set newrelevant = 0 where result_uid in ( select result_uid from solr_relevance_old s left outer join keylevels_results a on a.resultuid = s.result_uid where s.uuid ilike 'kl_%' and a.resultuid is null); UPDATE solr_relevance_old SET new_hod_correct = sub.hod_correct, new_hod_percent = sub.hod_percent, new_hod_total = sub.hod_total, new_pattern_correct = sub.pattern_correct, new_pattern_percent = sub.pattern_percent, new_pattern_total = sub.pattern_total, new_percent = sub.percent, new_symbol_correct = sub.symbol_correct, new_symbol_percent = sub.symbol_percent, new_symbol_total = sub.symbol_total FROM ( select distinct resultuid, hod_correct, hod_percent, hod_total, hod, pattern_correct, pattern_percent, pattern_total, percent, symbol_correct, symbol_percent, symbol_total FROM whatshot_probability WHERE type in ('kl', 'ekl')) sub WHERE result_uid = sub.resultuid;[ Date: 2026-01-12 01:18:15 ]
20 70.80 MiB with max_ra as ( select resultuid from relevance_keylevels_results order by resultuid desc limit 1) update solr_relevance_old set newrelevant = sub.relevant, newage = sub.age from ( select so.uuid, case when ra.relevant is not null then ra.relevant when so.result_uid < max_ra.resultuid then 0 else 1 end as relevant, case when ra.age is not null then ra.age when so.result_uid < max_ra.resultuid then 11 else 0 end as age, so.result_uid from max_ra, solr_relevance_old so inner join keylevels_results k on so.result_uid = k.resultuid and so.uuid ilike 'kl_%' inner join downloadersymbolsettings dss on k.symbolid = dss.symbolid left outer join relevance_keylevels_results ra on so.result_uid = ra.resultuid and so.uuid ilike 'kl_%') sub where solr_relevance_old.result_uid = sub.result_uid and solr_relevance_old.uuid ilike 'kl_%'; update solr_relevance_old set newrelevant = 0 where result_uid in ( select result_uid from solr_relevance_old s left outer join keylevels_results a on a.resultuid = s.result_uid where s.uuid ilike 'kl_%' and a.resultuid is null); UPDATE solr_relevance_old SET new_hod_correct = sub.hod_correct, new_hod_percent = sub.hod_percent, new_hod_total = sub.hod_total, new_pattern_correct = sub.pattern_correct, new_pattern_percent = sub.pattern_percent, new_pattern_total = sub.pattern_total, new_percent = sub.percent, new_symbol_correct = sub.symbol_correct, new_symbol_percent = sub.symbol_percent, new_symbol_total = sub.symbol_total FROM ( select distinct resultuid, hod_correct, hod_percent, hod_total, hod, pattern_correct, pattern_percent, pattern_total, percent, symbol_correct, symbol_percent, symbol_total FROM whatshot_probability WHERE type in ('kl', 'ekl')) sub WHERE result_uid = sub.resultuid;[ Date: 2026-01-12 01:20:15 ]
-
Vacuums
Vacuums / Analyzes Distribution
Key values
- 0 sec Highest CPU-cost vacuum
Table
Database - Date
- 0 sec Highest CPU-cost analyze
Table
Database - Date
Analyzes per table
Key values
- public.solr_relevance_old (16) Main table analyzed (database acaweb_fx)
- 48 analyzes Total
Table Number of analyzes acaweb_fx.public.solr_relevance_old 16 acaweb_fx.public.datafeeds_latestrun 4 acaweb_fx.pg_catalog.pg_attribute 4 acaweb_fx.public.relevance_keylevels_results 4 acaweb_fx.public.relevance_autochartist_results 4 acaweb_fx.public.relevance_fibonacci_results 4 acaweb_fx.pg_catalog.pg_type 2 acaweb_fx.public.latest_t15_candle_view 2 acaweb_fx.pg_catalog.pg_class 2 acaweb_fx.pg_catalog.pg_index 1 acaweb_fx.public.autochartist_symbolupdates 1 socialmedia.public.processes 1 acaweb_fx.public.relevance_consecutivecandles_results 1 acaweb_fx.public.symbollatestupdatetime 1 acaweb_fx.public.latest_candle_datetime_per_receng 1 Total 48 Vacuums per table
Key values
- public.solr_relevance_old (16) Main table vacuumed on database acaweb_fx
- 34 vacuums Total
Index Buffer usage Skipped WAL usage Table Vacuums scans hits misses dirtied pins frozen records full page bytes acaweb_fx.public.solr_relevance_old 16 16 12,804 0 83 0 0 9,205 1,980 8,588,686 acaweb_fx.public.datafeeds_latestrun 4 0 482 0 12 0 0 60 12 65,788 acaweb_fx.public.relevance_keylevels_results 2 2 7,450 0 468 5 143 2,093 454 1,753,607 acaweb_fx.public.relevance_autochartist_results 2 2 6,773 0 326 2 499 1,352 315 993,927 acaweb_fx.public.relevance_fibonacci_results 2 2 2,500 0 47 2 116 400 34 126,155 acaweb_fx.pg_toast.pg_toast_2619 1 1 162 0 20 0 0 98 20 84,436 acaweb_fx.pg_catalog.pg_type 1 1 111 0 21 0 0 49 15 102,532 acaweb_fx.public.autochartist_symbolupdates 1 1 22,195 0 2,814 4 38,242 5,544 2,792 1,490,643 acaweb_fx.public.solr_imports 1 1 65 0 4 0 0 6 4 16,259 acaweb_fx.pg_catalog.pg_attribute 1 1 778 0 197 0 67 375 152 861,882 acaweb_fx.pg_catalog.pg_statistic 1 1 1,001 0 145 0 582 453 115 506,044 acaweb_fx.public.latest_t15_candle_view 1 1 66 0 1 0 0 6 4 12,161 acaweb_fx.pg_catalog.pg_class 1 1 461 0 51 0 0 156 53 274,687 Total 34 30 54,848 46,472 4,189 13 39,649 19,797 5,950 14,876,807 Tuples removed per table
Key values
- public.solr_relevance_old (70155) Main table with removed tuples on database acaweb_fx
- 80221 tuples Total removed
Index Tuples Pages Table Vacuums scans removed remain not yet removable removed remain acaweb_fx.public.solr_relevance_old 16 16 70,155 95,347 0 0 3,142 acaweb_fx.public.autochartist_symbolupdates 1 1 5,104 54,792 281 0 40,691 acaweb_fx.pg_catalog.pg_attribute 1 1 1,478 10,778 0 0 260 acaweb_fx.public.relevance_keylevels_results 2 2 1,013 28,003 0 0 558 acaweb_fx.public.relevance_autochartist_results 2 2 916 17,293 0 0 760 acaweb_fx.pg_catalog.pg_statistic 1 1 666 3,763 0 0 1,194 acaweb_fx.public.datafeeds_latestrun 4 0 231 56 0 0 64 acaweb_fx.public.relevance_fibonacci_results 2 2 200 3,063 0 0 204 acaweb_fx.pg_catalog.pg_class 1 1 166 1,649 0 0 150 acaweb_fx.pg_catalog.pg_type 1 1 105 1,446 0 0 38 acaweb_fx.pg_toast.pg_toast_2619 1 1 70 168 0 4 45 acaweb_fx.public.latest_t15_candle_view 1 1 65 14 0 0 1 acaweb_fx.public.solr_imports 1 1 52 1 0 0 2 Total 34 30 80,221 216,373 281 4 47,109 Pages removed per table
Key values
- pg_toast.pg_toast_2619 (4) Main table with removed pages on database acaweb_fx
- 4 pages Total removed
Table Number of vacuums Index scans Tuples removed Pages removed acaweb_fx.pg_toast.pg_toast_2619 1 1 70 4 acaweb_fx.pg_catalog.pg_type 1 1 105 0 acaweb_fx.public.autochartist_symbolupdates 1 1 5104 0 acaweb_fx.public.datafeeds_latestrun 4 0 231 0 acaweb_fx.public.solr_imports 1 1 52 0 acaweb_fx.pg_catalog.pg_attribute 1 1 1478 0 acaweb_fx.pg_catalog.pg_statistic 1 1 666 0 acaweb_fx.public.latest_t15_candle_view 1 1 65 0 acaweb_fx.public.relevance_keylevels_results 2 2 1013 0 acaweb_fx.public.solr_relevance_old 16 16 70155 0 acaweb_fx.public.relevance_autochartist_results 2 2 916 0 acaweb_fx.pg_catalog.pg_class 1 1 166 0 acaweb_fx.public.relevance_fibonacci_results 2 2 200 0 Total 34 30 80,221 4 Autovacuum Activity
↑ Back to the top of the Autovacuum Activity tableDay Hour VACUUMs ANALYZEs Jan 12 01 34 48 - 0 sec Highest CPU-cost vacuum
-
Locks
Locks by types
Key values
- unknown Main Lock Type
- 0 locks Total
Most frequent waiting queries (N)
Rank Count Total time Min time Max time Avg duration Query NO DATASET
Queries that waited the most
Rank Wait time Query NO DATASET
-
Queries
Queries by type
Key values
- 62,980 Total read queries
- 44,258 Total write queries
Queries by database
Key values
- unknown Main database
- 232,823 Requests
- 1h31m32s (unknown)
- Main time consuming database
Database Request type Count Duration acaweb_fx Total 922 0ms copy from 80 0ms copy to 26 0ms cte 100 0ms ddl 16 0ms delete 16 0ms others 219 0ms select 94 0ms tcl 332 0ms update 39 0ms socialmedia Total 95 0ms select 95 0ms unknown Total 232,823 1h31m32s copy from 16 0ms cte 9,246 0ms insert 31,692 0ms others 9,463 0ms select 62,791 0ms tcl 332 0ms update 2,363 0ms Queries by user
Key values
- unknown Main user
- 232,823 Requests
User Request type Count Duration postgres Total 1,017 0ms copy from 80 0ms copy to 26 0ms cte 100 0ms ddl 16 0ms delete 16 0ms others 219 0ms select 189 0ms tcl 332 0ms update 39 0ms unknown Total 232,823 1h31m32s copy from 16 0ms cte 9,246 0ms insert 31,692 0ms others 9,463 0ms select 62,791 0ms tcl 332 0ms update 2,363 0ms Duration by user
Key values
- 1h31m32s (unknown) Main time consuming user
User Request type Count Duration postgres Total 1,017 0ms copy from 80 0ms copy to 26 0ms cte 100 0ms ddl 16 0ms delete 16 0ms others 219 0ms select 189 0ms tcl 332 0ms update 39 0ms unknown Total 232,823 1h31m32s copy from 16 0ms cte 9,246 0ms insert 31,692 0ms others 9,463 0ms select 62,791 0ms tcl 332 0ms update 2,363 0ms Queries by host
Key values
- unknown Main host
- 233,840 Requests
- 1h31m32s (unknown)
- Main time consuming host
Queries by application
Key values
- unknown Main application
- 233,465 Requests
- 1h31m32s (unknown)
- Main time consuming application
Number of cancelled queries
Key values
- 0 per second Cancelled query Peak
- 2026-01-12 01:41:52 Date
Number of cancelled queries (5 minutes period)
NO DATASET
-
Top Queries
Histogram of query times
Key values
- 80,708 0-1ms duration
Slowest individual queries
Rank Duration Query NO DATASET
Time consuming queries
Rank Total duration Times executed Min duration Max duration Avg duration Query 1 0ms 49 0ms 0ms 0ms select key, value from datasources ds inner join datasourceparams dsp on ds.id = dsp.datasourceid where ds.name = ?;Times Reported Time consuming queries #1
Day Hour Count Duration Avg duration Jan 12 01 49 0ms 0ms 2 0ms 227 0ms 0ms 0ms with rar_max as ( select resultuid from relevance_consecutivecandles_results order by resultuid desc limit ? ), all_results as ( select ccr.resultuid as resultuid, ccr.direction as direction, s.exchange as exchange, s.symbolid as symbolid, coalesce(bim.code, s.symbol) as symbol_code, s.longname as symbol_name, s.timegranularity as interval, ccr.patternendtime as identified, dtt.timezone as timezone, ccr.qtyconsecutivecandles as length, g.basegroupname, case when rcr.age is not null then rcr.age when ccr.resultuid <= rm.resultuid then ? else ? end as age, case when rcr.relevant is not null then rcr.relevant when ccr.resultuid <= rm.resultuid then ? else ? end as relevant, cps.pip, newlevels.filtered from consecutivecandles_results ccr inner join brokersymbollist bsl on bsl.brokerid = ? and bsl.symbolid = ccr.symbolid inner join symbols s on ccr.symbolid = s.symbolid and s.nonliquid = ? inner join downloadersymbolsettings dss on ccr.symbolid = dss.symbolid inner join datafeedstimetable dtt on dss.classname = dtt.classname and dtt.dayofweek = ? inner join symbolgroup sg on ccr.symbolid = sg.symbolid inner join groups g on sg.groupid = g.groupid inner join brokergroups bg on g.groupid = bg.groupid and bsl.brokerid = bg.brokerid inner join rar_max rm on ? = ? left outer join relevance_consecutivecandles_results rcr on rcr.resultuid = ccr.resultuid left join currencypips cps on cps.symbol = s.symbol left outer join brokerinstrumentmap bim on dss.datafeedinstrumentid = bim.datafeedinstrumentid and bim.brokerid = bsl.brokerid and bim.type = ? left join lateral calc_cc_signal_filter (ccr.resultuid) newlevels on true where ccr.gmttimefound > now() - interval ? and s.deleted = ? and (ccr.simulation = ? or ccr.simulation is null) and (? = ? or s.timegranularity in (...)) and (? = ? or s.exchange in (...)) and (? = ? or coalesce(bim.code, s.symbol) in (...)) and (? = ? or ccr.patternlengthbars <= ?) ), results as ( select distinct on (symbolid) * from all_results where (false = ? or relevant = ?) and (? = ? or age <= ?) order by symbolid, resultuid ) select * from results order by identified desc, length desc;Times Reported Time consuming queries #2
Day Hour Count Duration Avg duration Jan 12 01 227 0ms 0ms 3 0ms 1 0ms 0ms 0ms select trumpetsymbolid as sid, trumpettimegranularity as tg from brokersymbollist bsl left join powerstats_symboldata psd on bsl.symbolid = psd.symbolid left join downloadersymbolsettings dss on psd.symbolid = dss.symbolid left join symbols s on dss.symbolid = s.symbolid left outer join brokerinstrumentmap bdfi on code = ? and bdfi.brokerid = ? and dss.datafeedinstrumentid = bdfi.datafeedinstrumentid where (code = ? or s.symbol = ?) and bsl.brokerid = ? and dss.classname <> ? group by trumpetsymbolid, trumpettimegranularity;Times Reported Time consuming queries #3
Day Hour Count Duration Avg duration Jan 12 01 1 0ms 0ms 4 0ms 1 0ms 0ms 0ms insert into t15 (symbolid, pricedatetime, open, high, low, close, volume, bsf, sastdatetimereceived) values (?, ?::timestamp without time zone, ?, ?, ?, ?.?, ?, ?, ?::timestamp without time zone) on conflict (symbolid, pricedatetime) do nothing;Times Reported Time consuming queries #4
Day Hour Count Duration Avg duration Jan 12 01 1 0ms 0ms 5 0ms 14 0ms 0ms 0ms with rar_max as ( select resultuid from relevance_bigmovement_results order by resultuid desc limit ? ) select bmr.symbolid, patternstarttime, patternendtime, timegranularity, ? as direction, case when bmr.old_resultuid = ? then bmr.old_resultuid else bmr.resultuid end as uid, s.exchange, s.symbol, s.longname, s.shortname, dtt.timezone, bmr.patternmovement, bmr.statisticalmovement, bmr.fromprice, bmr.toprice, bmr.percentile, bmr.patternlengthbars, case when rbr.age is not null then rbr.age when bmr.resultuid <= rm.resultuid then ? else ? end as age, case when rbr.relevant is not null then rbr.relevant when bmr.resultuid <= rm.resultuid then ? else ? end as relevant, cps.pip from bigmovement_results bmr inner join downloadersymbolsettings dss on bmr.symbolid = dss.symbolid inner join datafeedstimetable dtt on dss.classname = dtt.classname inner join symbols s on bmr.symbolid = s.symbolid inner join rar_max rm on ? = ? left outer join relevance_bigmovement_results rbr on rbr.resultuid = bmr.resultuid left join currencypips cps on cps.symbol = s.symbol where (bmr.old_resultuid = ? or bmr.resultuid = ?) and dtt.dayofweek = ?;Times Reported Time consuming queries #5
Day Hour Count Duration Avg duration Jan 12 01 14 0ms 0ms 6 0ms 3,447 0ms 0ms 0ms insert into t60 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) values (?, ?, ?, ?, ?, ?, ?, ?, ?, ?) on conflict (pricedatetime, symbolid) do update set open = ?, high = ?, low = ?, close = ?, volume = ?, bsf = ?, sastdatetimewritten = ?, sastdatetimereceived = ?;Times Reported Time consuming queries #6
Day Hour Count Duration Avg duration Jan 12 01 3,447 0ms 0ms 7 0ms 4 0ms 0ms 0ms select psd.symbolid, psd.hourlysymbolid as intervalsymbolid, s.symbol, dtt.timezone, dtt.absolutetimezoneoffset, psh.hour as index, psh.ave, psh.stddev, (psh.ave - psh.stddev / ?.?) as low, (psh.ave + psh.stddev / ?.?) as high, ee.timezone as exchangetimezone, ee.mon_t1start as exchangestart, ee.mon_t1end as exchangeend from powerstats_symboldata psd inner join powerstats_hourly psh on psd.hourlysymbolid = psh.symbolid inner join symbols s on psh.symbolid = s.symbolid inner join downloadersymbolsettings dss on s.symbolid = dss.symbolid inner join datafeedstimetable dtt on dss.classname = dtt.classname inner join mat_ps_hourly_symbolid_max_enddate e on psh.enddate = e.enddate and psh.symbolid = e.symbolid inner join exchanges ee on ee.exchange = s.exchange where psd.symbolid = ? and dtt.dayofweek = ? order by hour asc;Times Reported Time consuming queries #7
Day Hour Count Duration Avg duration Jan 12 01 4 0ms 0ms 8 0ms 130 0ms 0ms 0ms select case when a.old_resultuid = ? then a.old_resultuid else a.resultuid end as resultuid, s.symbol, timegranularity as interval, direction as direction, patternendtime as patternendtime, patternstartprice as psp, patternendprice as pep, target03 as t03, target16 as t16, patternlengthbars as length, p.patternname as patternname, dtt.timezone, cps.pip from fibonacci_results a inner join downloadersymbolsettings dss on a.symbolid = dss.symbolid inner join datafeedstimetable dtt on dss.classname = dtt.classname inner join symbols s on a.symbolid = s.symbolid inner join fibonaccipatterns p on a.pattern = p.patternname left join currencypips cps on cps.symbol = s.symbol where (a.old_resultuid = ? or a.resultuid = ?) and dtt.dayofweek = ?;Times Reported Time consuming queries #8
Day Hour Count Duration Avg duration Jan 12 01 130 0ms 0ms 9 0ms 1,828 0ms 0ms 0ms update patternresultsrelevance set relevant = ?, saxo_relevant = ?, notrelevantpricedatetime = ?, reason = ? where uniqueindex = ? and relevant = ?;Times Reported Time consuming queries #9
Day Hour Count Duration Avg duration Jan 12 01 1,828 0ms 0ms 10 0ms 4 0ms 0ms 0ms select t.pricedatetime as pricedatetime, t.open as open, t.high as high, t.low as low, t.close "..." close, t.volume as volume, t.bsf as bsf from t1440 t where symbolid = ? and (bsf = ? or bsf is null) order by pricedatetime desc limit ?;Times Reported Time consuming queries #10
Day Hour Count Duration Avg duration Jan 12 01 4 0ms 0ms 11 0ms 18 0ms 0ms 0ms select cast(count(*) / cast(setting as numeric) * ? as int) from pg_stat_activity, pg_settings where name = ? group by setting;Times Reported Time consuming queries #11
Day Hour Count Duration Avg duration Jan 12 01 18 0ms 0ms 12 0ms 239 0ms 0ms 0ms select v.datname, c.relname, v.phase, v.heap_blks_total, v.heap_blks_scanned, v.heap_blks_vacuumed, v.index_vacuum_count, v.max_dead_tuples, v.num_dead_tuples from pg_stat_progress_vacuum as v join pg_class c on c.oid = v.relid;Times Reported Time consuming queries #12
Day Hour Count Duration Avg duration Jan 12 01 239 0ms 0ms 13 0ms 37 0ms 0ms 0ms select * from ( select pricedatetime, open, high, low, close, volume, bsf from t60 where symbolid = ? and (bsf = ? or bsf is null) order by pricedatetime desc limit ?) a order by pricedatetime asc;Times Reported Time consuming queries #13
Day Hour Count Duration Avg duration Jan 12 01 37 0ms 0ms 14 0ms 2 0ms 0ms 0ms with pre_symbols as ( select s.symbolid, s.symbol, s.timegranularity, dss.downloadersymbol, dtt.timezone, s.exchange, s.longname from symbols s inner join downloadersymbolsettings dss on dss.symbolid = s.symbolid inner join datafeedstimetable dtt on dss.classname = dtt.classname and dtt.dayofweek = ? inner join brokersymbollist bsl on bsl.symbolid = s.symbolid and brokerid = ? where dss.classname in (...) and (dss.downloadersymbol in (...) or s.symbol in (...)) and dss.enabled = ? and s.nonliquid = ? and s.deleted = ? ), report_symbols as ( select ps1.*, ps2.symbolid as price_symbol_id from pre_symbols ps1 inner join pre_symbols ps2 on ps1.symbol = ps2.symbol and ps1.downloadersymbol = ps2.downloadersymbol and ps2.timegranularity = ? ), rbr_max as ( select resultuid from relevance_bigmovement_results order by resultuid desc limit ? ) select bmr.resultuid as ruid, ? as direction, rs.symbol as sym, rs.downloadersymbol, rs.symbolid as sid, rs.timegranularity as tg, rs.timezone, rs.exchange as e, rs.longname, bmr.patternendtime as pet, lpi.latestpricedatetime as lpdt, bmr.patternlengthbars as l, bmr.breakout >= ? as complete, rbr.age as age from bigmovement_results bmr inner join report_symbols rs on rs.symbolid = bmr.symbolid inner join autochartist_symbolupdates lpi on lpi.symbolid = rs.price_symbol_id inner join rbr_max rm on ? = ? left outer join relevance_bigmovement_results rbr on rbr.resultuid = bmr.resultuid where patternendtime >= (now() - ? * interval ?) -- results can't be more than ? days old and (bmr.resultuid > rm.resultuid or rbr.relevant = ?) ;Times Reported Time consuming queries #14
Day Hour Count Duration Avg duration Jan 12 01 2 0ms 0ms 15 0ms 332 0ms 0ms 0ms commit;Times Reported Time consuming queries #15
Day Hour Count Duration Avg duration Jan 12 01 332 0ms 0ms 16 0ms 347 0ms 0ms 0ms with rar_max as ( select resultuid from relevance_keylevels_results order by resultuid desc limit ? ), kr as ( select a.*, rr.age, rr.relevant from keylevels_results a left outer join relevance_keylevels_results rr on a.resultuid = rr.resultuid where case when false = ? then true else a.resultuid > ( select min(resultuid) from relevance_keylevels_results) end ), all_results as ( select kr.resultuid as resultuid, kr.direction as direction, s.exchange as exchange, s.symbolid as symbolid, coalesce(bim.code, s.symbol) as symbol_code, s.longname as symbol_name, s.timegranularity as interval, p.patternname as pattern_name, kr.breakout as breakout, kr.atbaridentified as identified, dtt.timezone as timezone, kr.patternlengthbars as length, g.basegroupname, newlevels.filtered, case when kr.age is not null then kr.age when kr.resultuid <= rm.resultuid then ? else ? end as age, case when kr.relevant is not null then kr.relevant when kr.resultuid <= rm.resultuid then ? else ? end as relevant, cps.pip from kr inner join brokersymbollist bsl on bsl.brokerid = ? and bsl.symbolid = kr.symbolid inner join symbols s on bsl.symbolid = s.symbolid and s.nonliquid = ? inner join symbolgroup sg on s.symbolid = sg.symbolid inner join groups g on sg.groupid = g.groupid inner join brokergroups bg on g.groupid = bg.groupid and bsl.brokerid = bg.brokerid inner join hrspatterns p on kr.patternid = p.patternid inner join downloadersymbolsettings dss on s.symbolid = dss.symbolid inner join datafeedstimetable dtt on dss.classname = dtt.classname and dtt.dayofweek = ? inner join rar_max rm on ? = ? left outer join autochartist_symbolupdates au on dss.symbolid = au.symbolid left outer join relevance_keylevels_results rar on rar.resultuid = kr.resultuid left join lateral calc_kl_signal_filter (kr.resultuid) newlevels on true left join currencypips cps on cps.symbol = s.symbol left outer join brokerinstrumentmap bim on dss.datafeedinstrumentid = bim.datafeedinstrumentid and bim.brokerid = bsl.brokerid and bim.type = ? where kr.gmttimefound > now() - interval ? and dss.enabled = ? and s.deleted = ? and (kr.simulation = ? or kr.simulation is null) and (? = ? or s.timegranularity in (...)) and (? = ? or s.exchange in (...)) and (? = ? or coalesce(bim.code, s.symbol) in (...)) and (? = ? or p.patternname in (...)) and (? = ? or kr.patternclassid in (...)) and (? = ? or kr.patternlengthbars <= ?) and kr.patternstarttime::timestamp without time zone >= coalesce(au.earliestpricedatetime, ?::timestamp without time zone) -- to make sure patternstarttime is in our t-tables ), results as ( select distinct on (symbolid) * from all_results where (false = ? or relevant = ?) and (? = ? or age <= ?) order by symbolid, resultuid ) select * from results order by identified desc, length desc limit ?;Times Reported Time consuming queries #16
Day Hour Count Duration Avg duration Jan 12 01 347 0ms 0ms 17 0ms 15 0ms 0ms 0ms select psd.symbolid, psd.hourlysymbolid as intervalsymbolid, s.symbol, dtt.timezone, dtt.absolutetimezoneoffset, psh.hour as index, psh.ave, psh.stddev, ((psh.ave - psh.stddev) / ?.?) as low, ((psh.ave + psh.stddev) / ?.?) as high, ee.timezone as exchangetimezone, ee.mon_t1start as exchangestart, ee.mon_t1end as exchangeend from powerstats_symboldata psd inner join powerstats_hourly psh on psd.hourlysymbolid = psh.symbolid inner join symbols s on psh.symbolid = s.symbolid inner join downloadersymbolsettings dss on s.symbolid = dss.symbolid inner join datafeedstimetable dtt on dss.classname = dtt.classname inner join mat_ps_hourly_symbolid_max_enddate e on psh.enddate = e.enddate and psh.symbolid = e.symbolid inner join exchanges ee on ee.exchange = s.exchange where psd.symbolid = ? and dtt.dayofweek = ? order by hour asc;Times Reported Time consuming queries #17
Day Hour Count Duration Avg duration Jan 12 01 15 0ms 0ms 18 0ms 4 0ms 0ms 0ms select (cast(substring(tz.gmoffset from ? for ?) as float) * ? + cast(substring(tz.gmoffset from ? for ?) as float)) / ? as offset from timezones tz where tz.timezone = ?;Times Reported Time consuming queries #18
Day Hour Count Duration Avg duration Jan 12 01 4 0ms 0ms 19 0ms 239 0ms 0ms 0ms select count(*), sum(size), extract(epoch from now() - min(modification)) from pg_ls_waldir ();Times Reported Time consuming queries #19
Day Hour Count Duration Avg duration Jan 12 01 239 0ms 0ms 20 0ms 40 0ms 0ms 0ms select * from ( select pricedatetime, open, high, low, close, volume, bsf from t30 where symbolid = ? and (bsf = ? or bsf is null) order by pricedatetime desc limit ?) a order by pricedatetime asc;Times Reported Time consuming queries #20
Day Hour Count Duration Avg duration Jan 12 01 40 0ms 0ms Most frequent queries (N)
Rank Times executed Total duration Min duration Max duration Avg duration Query 1 29,140 0ms 0ms 0ms 0ms select ?;Times Reported Time consuming queries #1
Day Hour Count Duration Avg duration Jan 12 01 29,140 0ms 0ms 2 17,424 0ms 0ms 0ms 0ms select distinct on (coalesce(bim.code, s.symbol) , s.exchange, s.timegranularity, df.timezone) s.symbolid as id, coalesce(bim.code, s.symbol) as name, s.symbol as symbol, dss.downloadersymbol as ticker, s.exchange as exchange, s.timegranularity as interval, df.timezone as timezone from symbols s inner join downloadersymbolsettings dss on dss.symbolid = s.symbolid inner join datafeedstimetable df on df.classname ilike dss.classname left join brokersymbollist bsl on bsl.brokerid = ? and bsl.symbolid = s.symbolid left outer join brokerinstrumentmap bim on dss.datafeedinstrumentid = bim.datafeedinstrumentid and bim.brokerid = ? and bim.type = ? where s.symbolid = ?;Times Reported Time consuming queries #2
Day Hour Count Duration Avg duration Jan 12 01 17,424 0ms 0ms 3 8,230 0ms 0ms 0ms 0ms insert into executionlogs (executionid, status, message, details, detailtype) values (null, ?, ?, null, null);Times Reported Time consuming queries #3
Day Hour Count Duration Avg duration Jan 12 01 8,230 0ms 0ms 4 5,618 0ms 0ms 0ms 0ms insert into t15 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) values (?, ?, ?, ?, ?, ?, ?, ?, ?, ?) on conflict (pricedatetime, symbolid) do update set open = ?, high = ?, low = ?, close = ?, volume = ?, bsf = ?, sastdatetimewritten = ?, sastdatetimereceived = ?;Times Reported Time consuming queries #4
Day Hour Count Duration Avg duration Jan 12 01 5,618 0ms 0ms 5 4,989 0ms 0ms 0ms 0ms with rar_max as ( select resultuid from relevance_autochartist_results order by resultuid desc limit ? ) select a.symbolid, pattern, patternid, resy0, resy1, resx0, resx1, supporty0, supporty1, supportx0, supportx1, predictiontimeto, patternstarttime, timegranularity, patternendtime, direction, trendchange, patternlengthbars, patternquality, case when a.old_resultuid = ? then a.old_resultuid else a.resultuid end as uid, breakout, initialtrend, volumeincrease, symmetry as uniformity, predictionpricefrom, predictionpriceto, noise, s.exchange, s.symbol, s.longname, s.shortname, breakout, dtt.timezone, patternstartprice, patternendprice, qtytp, newlevels.profit, newlevels.stop, newlevels.filtered, case when rar.age is not null then rar.age when a.resultuid <= rm.resultuid then ? else ? end as age, case when rar.relevant is not null then rar.relevant when a.resultuid <= rm.resultuid then ? else ? end as relevant, cps.pip from autochartist_results a inner join downloadersymbolsettings dss on a.symbolid = dss.symbolid inner join datafeedstimetable dtt on dss.classname = dtt.classname inner join symbols s on a.symbolid = s.symbolid inner join patterns p on p.patternname = a.pattern inner join rar_max rm on ? = ? left outer join relevance_autochartist_results rar on rar.resultuid = a.resultuid left join lateral calc_cp_signal (a.resultuid) newlevels on true left join currencypips cps on cps.symbol = s.symbol where (a.old_resultuid = ? or a.resultuid = ?) and dtt.dayofweek = ?;Times Reported Time consuming queries #5
Day Hour Count Duration Avg duration Jan 12 01 4,989 0ms 0ms 6 4,506 0ms 0ms 0ms 0ms set extra_float_digits = ?;Times Reported Time consuming queries #6
Day Hour Count Duration Avg duration Jan 12 01 4,506 0ms 0ms 7 4,488 0ms 0ms 0ms 0ms set application_name = ?;Times Reported Time consuming queries #7
Day Hour Count Duration Avg duration Jan 12 01 4,488 0ms 0ms 8 4,054 0ms 0ms 0ms 0ms insert into t30 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) values (?, ?, ?, ?, ?, ?, ?, ?, ?, ?) on conflict (pricedatetime, symbolid) do update set open = ?, high = ?, low = ?, close = ?, volume = ?, bsf = ?, sastdatetimewritten = ?, sastdatetimereceived = ?;Times Reported Time consuming queries #8
Day Hour Count Duration Avg duration Jan 12 01 4,054 0ms 0ms 9 3,798 0ms 0ms 0ms 0ms select s.symbolid as id, s.symbol as name, s.exchange as exchange, s.timegranularity as interval, dtt.timezone as timezone from symbols s inner join downloadersymbolsettings dss on dss.symbolid = s.symbolid inner join datafeedstimetable dtt on dss.classname = dtt.classname and dtt.dayofweek = ? inner join brokersymbollist bsl on bsl.symbolid = s.symbolid where bsl.brokerid = ? and (? = ? or s.timegranularity = ?) and (s.symbol = ? or dss.downloadersymbol = ?) and dss.enabled = ?;Times Reported Time consuming queries #9
Day Hour Count Duration Avg duration Jan 12 01 3,798 0ms 0ms 10 3,447 0ms 0ms 0ms 0ms insert into t60 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) values (?, ?, ?, ?, ?, ?, ?, ?, ?, ?) on conflict (pricedatetime, symbolid) do update set open = ?, high = ?, low = ?, close = ?, volume = ?, bsf = ?, sastdatetimewritten = ?, sastdatetimereceived = ?;Times Reported Time consuming queries #10
Day Hour Count Duration Avg duration Jan 12 01 3,447 0ms 0ms 11 3,366 0ms 0ms 0ms 0ms insert into autochartist_results (resultid, symbolid, bandwidth, pattern, qtytp, gmttimefound, direction, initialtrend, breakout, volumeincrease, noise, symmetry, predictionpricefrom, predictionpriceto, predictiontimefrom, predictiontimeto, patternstarttime, patternendtime, patternstartprice, patternendprice, resx0, resx1, supportx0, supportx1, resy0, resy1, supporty0, supporty1, supportgradient, resgradient, riskreward, patternquality, trendchange, maxmovementafterbreakout, latestbaratbreakouttime, latestbaratbreakoutprice, patternlengthbars, temporarypattern, relevancestartdistance, simulation, writtendatetime) values (?, ?, ?.?, ?, ?, ?::timestamp without time zone, ?, ?.?, ?.?, ?.?, ?.?, ?.?, ?.?, ?.?, ?::timestamp without time zone, ?::timestamp without time zone, ?::timestamp without time zone, ?::timestamp without time zone, ?.?, ?.?, ?::timestamp without time zone, ?::timestamp without time zone, ?::timestamp without time zone, ?::timestamp without time zone, ?.?, ?.?, ?.?, ?.?, ?.?, ?.?, ?.?, ?.?, ?, ?.?, ?::timestamp without time zone, ?.?, ?, ?, ?.?, ?, current_timestamp::timestamp without time zone) on conflict do nothing;Times Reported Time consuming queries #11
Day Hour Count Duration Avg duration Jan 12 01 3,366 0ms 0ms 12 2,589 0ms 0ms 0ms 0ms insert into keylevels_results (bandwidth, breakout, patternid, gmttimefound, approachingtimestamp, approachingregion, qtytp, patternlengthbars, patternprice, x0, x1, x2, x3, x4, x5, x6, x7, x8, x9, breakoutbars, breakoutprice, patternendtime, atbaridentified, atpriceidentified, errormargin, direction, symbolid, predictionpricefrom, predictionpriceto, predictiontimefrom, predictiontimebars, uniquepointsvalue, furthestprice, relevancestartdistance, patternclassid, patternstarttime, stoplosslevel, simulation, writtendatetime) values (?.?, ?, ?, ?::timestamp without time zone, ?, ?.?, ?, ?, ?.?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?.?, ?::timestamp without time zone, ?, ?.?, ?.?, ?, ?, ?.?, ?.?, ?::timestamp without time zone, ?, ?, ?.?, ?.?, ?, ?, ?.?, ?, current_timestamp::timestamp without time zone) on conflict do nothing;Times Reported Time consuming queries #12
Day Hour Count Duration Avg duration Jan 12 01 2,589 0ms 0ms 13 1,943 0ms 0ms 0ms 0ms insert into fibonacci_results (bandwidth, pattern, gmttimefound, direction, patternstarttime, patternendtime, patternstartprice, patternendprice, qtytp, pricex, timex, pricea, timea, priceb, timeb, pricec, timec, priced, timed, averagequality, timequality, errormargin, patternlengthbars, target10, target06, target16, target07, target12, target05, target03, symbolid, noise, ratiosfound, temporarypattern, uniqueindex, completed, simulation, writtendatetime) values (?.?, ?, ?::timestamp without time zone, ?, ?::timestamp without time zone, ?::timestamp without time zone, ?.?, ?.?, ?, ?.?, ?::timestamp without time zone, ?.?, ?::timestamp without time zone, ?.?, ?::timestamp without time zone, ?.?, ?::timestamp without time zone, ?.?, ?::timestamp without time zone, ?.?, ?.?, ?.?, ?, ?.?, ?.?, ?.?, ?.?, ?.?, ?.?, ?.?, ?, ?.?, ?, ?, ?, ?, ?, current_timestamp::timestamp without time zone) on conflict do nothing;Times Reported Time consuming queries #13
Day Hour Count Duration Avg duration Jan 12 01 1,943 0ms 0ms 14 1,828 0ms 0ms 0ms 0ms update patternresultsrelevance set relevant = ?, saxo_relevant = ?, notrelevantpricedatetime = ?, reason = ? where uniqueindex = ? and relevant = ?;Times Reported Time consuming queries #14
Day Hour Count Duration Avg duration Jan 12 01 1,828 0ms 0ms 15 1,672 0ms 0ms 0ms 0ms with rar_max as ( select resultuid from relevance_keylevels_results order by resultuid desc limit ? ) select case when a.old_resultuid = ? then a.old_resultuid else a.resultuid end as ruid, s.symbolid as sid, s.symbol as sym, longname, shortname, exchange as e, timegranularity as tg, a.patternid as pid, a.direction as d, a.patternprice as pp, atbaridentified as pet, case when (x9 != ?) then x9 when (x8 != ?) then x8 when (x7 != ?) then x7 when (x6 != ?) then x6 when (x5 != ?) then x5 when (x4 != ?) then x4 when (x3 != ?) then x3 when (x2 != ?) then x2 end as pst, patternprice as patp, x0, x1, x2, case when (x3 != ?) then x3 else ? end as x3, case when (x4 != ?) then x4 else ? end as x4, case when (x5 != ?) then x5 else ? end as x5, case when (x6 != ?) then x6 else ? end as x6, case when (x7 != ?) then x7 else ? end as x7, case when (x8 != ?) then x8 else ? end as x8, errormargin as erm, breakoutprice as pe, breakoutbars as be, breakout, atbaridentified as atbar, atpriceidentified as atprice, patternlengthbars as l, bandwidth as bw, qtytp as qtp, p.patternname as patternname, dtt.absolutetimezoneoffset as tzos, dtt.timezone as timezone, approachingtimestamp as apt, approachingregion as apr, predictionpricefrom as ppf, predictionpriceto as ppt, predictiontimefrom as ptf, predictiontimebars as ptb, furthestprice as fp, newlevels.filtered, a.uniquepointsvalue as upv, case when rar.age is not null then rar.age when a.resultuid <= rm.resultuid then ? else ? end as age, case when rar.relevant is not null then rar.relevant when a.resultuid <= rm.resultuid then ? else ? end as relevant, cps.pip from keylevels_results a inner join downloadersymbolsettings dss on a.symbolid = dss.symbolid inner join datafeedstimetable dtt on dss.classname = dtt.classname inner join symbols s on a.symbolid = s.symbolid inner join hrspatterns p on a.patternid = p.patternid inner join rar_max rm on ? = ? left outer join relevance_keylevels_results rar on a.resultuid = rar.resultuid left join lateral calc_kl_signal_filter (a.resultuid) newlevels on true left join currencypips cps on cps.symbol = s.symbol where (a.old_resultuid = ? or a.resultuid = ?) and dtt.dayofweek = ?;Times Reported Time consuming queries #15
Day Hour Count Duration Avg duration Jan 12 01 1,672 0ms 0ms 16 895 0ms 0ms 0ms 0ms select symbolid, pricedatetime, classname, downloadfrequency, downloadersymbol, open, high, low, close, volume, bsf, sastdatetimereceived from ( select pricedatetime, dss.classname, dss.downloadfrequency, dss.symbolid, dss.downloadersymbol, t.open, t.high, t.low, t.close, t.volume, t.bsf, t.sastdatetimereceived, row_number() over (partition by t.symbolid order by t.pricedatetime desc) as rn from t15 t, downloadersymbolsettings dss, symbols s where dss.classname = ? and dss.downloadfrequency = ? and dss.symbolid = t.symbolid and s.symbolid = dss.symbolid and dss.enabled = ? and s.deleted = ? and dss.downloadersymbol in (...) and t.pricedatetime > now() - interval ?) as ranked_candles_table where rn = ?;Times Reported Time consuming queries #16
Day Hour Count Duration Avg duration Jan 12 01 895 0ms 0ms 17 849 0ms 0ms 0ms 0ms insert into t1440_underlying (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) values (?, ?, ?, ?, ?, ?, ?, ?, ?, ?) on conflict (pricedatetime, symbolid) do update set open = ?, high = ?, low = ?, close = ?, volume = ?, bsf = ?, sastdatetimewritten = ?, sastdatetimereceived = ?;Times Reported Time consuming queries #17
Day Hour Count Duration Avg duration Jan 12 01 849 0ms 0ms 18 728 0ms 0ms 0ms 0ms select t.pricedatetime, t.open, t.high, t.low, t.close, t.volume, t.symbolid, dss.downloadersymbol as symbol, dss.classname as datafeed, t.bsf, t.sastdatetimewritten, t.sastdatetimereceived from t15 t inner join downloadersymbolsettings dss on t.symbolid = dss.symbolid where dss.classname = ? and dss.downloadersymbol in (...) and dss.downloadfrequency = ? and t.pricedatetime between ?::timestamp and ?::timestamp order by t.pricedatetime;Times Reported Time consuming queries #18
Day Hour Count Duration Avg duration Jan 12 01 728 0ms 0ms 19 694 0ms 0ms 0ms 0ms select case when a.old_resultuid = ? then a.old_resultuid else a.resultuid end as resultuid, s.symbol, pattern as patternname, timegranularity as interval, patternlengthbars as length, patternendtime, direction, breakout, predictiontimeto, predictionpricefrom, predictionpriceto, patternstartprice, resy1, supporty1, dtt.timezone, cps.pip, newlevels.profit from autochartist_results a inner join downloadersymbolsettings dss on a.symbolid = dss.symbolid inner join datafeedstimetable dtt on dss.classname = dtt.classname inner join symbols s on a.symbolid = s.symbolid inner join patterns p on p.patternname = a.pattern left join currencypips cps on cps.symbol = s.symbol left join lateral calc_cp_signal (a.resultuid) newlevels on true where (a.old_resultuid = ? or a.resultuid = ?) and dtt.dayofweek = ?;Times Reported Time consuming queries #19
Day Hour Count Duration Avg duration Jan 12 01 694 0ms 0ms 20 659 0ms 0ms 0ms 0ms insert into t240 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) values (?, ?, ?, ?, ?, ?, ?, ?, ?, ?) on conflict (pricedatetime, symbolid) do update set open = ?, high = ?, low = ?, close = ?, volume = ?, bsf = ?, sastdatetimewritten = ?, sastdatetimereceived = ?;Times Reported Time consuming queries #20
Day Hour Count Duration Avg duration Jan 12 01 659 0ms 0ms Normalized slowest queries (N)
Rank Min duration Max duration Avg duration Times executed Total duration Query 1 0ms 0ms 0ms 49 0ms select key, value from datasources ds inner join datasourceparams dsp on ds.id = dsp.datasourceid where ds.name = ?;Times Reported Time consuming queries #1
Day Hour Count Duration Avg duration Jan 12 01 49 0ms 0ms 2 0ms 0ms 0ms 227 0ms with rar_max as ( select resultuid from relevance_consecutivecandles_results order by resultuid desc limit ? ), all_results as ( select ccr.resultuid as resultuid, ccr.direction as direction, s.exchange as exchange, s.symbolid as symbolid, coalesce(bim.code, s.symbol) as symbol_code, s.longname as symbol_name, s.timegranularity as interval, ccr.patternendtime as identified, dtt.timezone as timezone, ccr.qtyconsecutivecandles as length, g.basegroupname, case when rcr.age is not null then rcr.age when ccr.resultuid <= rm.resultuid then ? else ? end as age, case when rcr.relevant is not null then rcr.relevant when ccr.resultuid <= rm.resultuid then ? else ? end as relevant, cps.pip, newlevels.filtered from consecutivecandles_results ccr inner join brokersymbollist bsl on bsl.brokerid = ? and bsl.symbolid = ccr.symbolid inner join symbols s on ccr.symbolid = s.symbolid and s.nonliquid = ? inner join downloadersymbolsettings dss on ccr.symbolid = dss.symbolid inner join datafeedstimetable dtt on dss.classname = dtt.classname and dtt.dayofweek = ? inner join symbolgroup sg on ccr.symbolid = sg.symbolid inner join groups g on sg.groupid = g.groupid inner join brokergroups bg on g.groupid = bg.groupid and bsl.brokerid = bg.brokerid inner join rar_max rm on ? = ? left outer join relevance_consecutivecandles_results rcr on rcr.resultuid = ccr.resultuid left join currencypips cps on cps.symbol = s.symbol left outer join brokerinstrumentmap bim on dss.datafeedinstrumentid = bim.datafeedinstrumentid and bim.brokerid = bsl.brokerid and bim.type = ? left join lateral calc_cc_signal_filter (ccr.resultuid) newlevels on true where ccr.gmttimefound > now() - interval ? and s.deleted = ? and (ccr.simulation = ? or ccr.simulation is null) and (? = ? or s.timegranularity in (...)) and (? = ? or s.exchange in (...)) and (? = ? or coalesce(bim.code, s.symbol) in (...)) and (? = ? or ccr.patternlengthbars <= ?) ), results as ( select distinct on (symbolid) * from all_results where (false = ? or relevant = ?) and (? = ? or age <= ?) order by symbolid, resultuid ) select * from results order by identified desc, length desc;Times Reported Time consuming queries #2
Day Hour Count Duration Avg duration Jan 12 01 227 0ms 0ms 3 0ms 0ms 0ms 1 0ms select trumpetsymbolid as sid, trumpettimegranularity as tg from brokersymbollist bsl left join powerstats_symboldata psd on bsl.symbolid = psd.symbolid left join downloadersymbolsettings dss on psd.symbolid = dss.symbolid left join symbols s on dss.symbolid = s.symbolid left outer join brokerinstrumentmap bdfi on code = ? and bdfi.brokerid = ? and dss.datafeedinstrumentid = bdfi.datafeedinstrumentid where (code = ? or s.symbol = ?) and bsl.brokerid = ? and dss.classname <> ? group by trumpetsymbolid, trumpettimegranularity;Times Reported Time consuming queries #3
Day Hour Count Duration Avg duration Jan 12 01 1 0ms 0ms 4 0ms 0ms 0ms 1 0ms insert into t15 (symbolid, pricedatetime, open, high, low, close, volume, bsf, sastdatetimereceived) values (?, ?::timestamp without time zone, ?, ?, ?, ?.?, ?, ?, ?::timestamp without time zone) on conflict (symbolid, pricedatetime) do nothing;Times Reported Time consuming queries #4
Day Hour Count Duration Avg duration Jan 12 01 1 0ms 0ms 5 0ms 0ms 0ms 14 0ms with rar_max as ( select resultuid from relevance_bigmovement_results order by resultuid desc limit ? ) select bmr.symbolid, patternstarttime, patternendtime, timegranularity, ? as direction, case when bmr.old_resultuid = ? then bmr.old_resultuid else bmr.resultuid end as uid, s.exchange, s.symbol, s.longname, s.shortname, dtt.timezone, bmr.patternmovement, bmr.statisticalmovement, bmr.fromprice, bmr.toprice, bmr.percentile, bmr.patternlengthbars, case when rbr.age is not null then rbr.age when bmr.resultuid <= rm.resultuid then ? else ? end as age, case when rbr.relevant is not null then rbr.relevant when bmr.resultuid <= rm.resultuid then ? else ? end as relevant, cps.pip from bigmovement_results bmr inner join downloadersymbolsettings dss on bmr.symbolid = dss.symbolid inner join datafeedstimetable dtt on dss.classname = dtt.classname inner join symbols s on bmr.symbolid = s.symbolid inner join rar_max rm on ? = ? left outer join relevance_bigmovement_results rbr on rbr.resultuid = bmr.resultuid left join currencypips cps on cps.symbol = s.symbol where (bmr.old_resultuid = ? or bmr.resultuid = ?) and dtt.dayofweek = ?;Times Reported Time consuming queries #5
Day Hour Count Duration Avg duration Jan 12 01 14 0ms 0ms 6 0ms 0ms 0ms 3,447 0ms insert into t60 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) values (?, ?, ?, ?, ?, ?, ?, ?, ?, ?) on conflict (pricedatetime, symbolid) do update set open = ?, high = ?, low = ?, close = ?, volume = ?, bsf = ?, sastdatetimewritten = ?, sastdatetimereceived = ?;Times Reported Time consuming queries #6
Day Hour Count Duration Avg duration Jan 12 01 3,447 0ms 0ms 7 0ms 0ms 0ms 4 0ms select psd.symbolid, psd.hourlysymbolid as intervalsymbolid, s.symbol, dtt.timezone, dtt.absolutetimezoneoffset, psh.hour as index, psh.ave, psh.stddev, (psh.ave - psh.stddev / ?.?) as low, (psh.ave + psh.stddev / ?.?) as high, ee.timezone as exchangetimezone, ee.mon_t1start as exchangestart, ee.mon_t1end as exchangeend from powerstats_symboldata psd inner join powerstats_hourly psh on psd.hourlysymbolid = psh.symbolid inner join symbols s on psh.symbolid = s.symbolid inner join downloadersymbolsettings dss on s.symbolid = dss.symbolid inner join datafeedstimetable dtt on dss.classname = dtt.classname inner join mat_ps_hourly_symbolid_max_enddate e on psh.enddate = e.enddate and psh.symbolid = e.symbolid inner join exchanges ee on ee.exchange = s.exchange where psd.symbolid = ? and dtt.dayofweek = ? order by hour asc;Times Reported Time consuming queries #7
Day Hour Count Duration Avg duration Jan 12 01 4 0ms 0ms 8 0ms 0ms 0ms 130 0ms select case when a.old_resultuid = ? then a.old_resultuid else a.resultuid end as resultuid, s.symbol, timegranularity as interval, direction as direction, patternendtime as patternendtime, patternstartprice as psp, patternendprice as pep, target03 as t03, target16 as t16, patternlengthbars as length, p.patternname as patternname, dtt.timezone, cps.pip from fibonacci_results a inner join downloadersymbolsettings dss on a.symbolid = dss.symbolid inner join datafeedstimetable dtt on dss.classname = dtt.classname inner join symbols s on a.symbolid = s.symbolid inner join fibonaccipatterns p on a.pattern = p.patternname left join currencypips cps on cps.symbol = s.symbol where (a.old_resultuid = ? or a.resultuid = ?) and dtt.dayofweek = ?;Times Reported Time consuming queries #8
Day Hour Count Duration Avg duration Jan 12 01 130 0ms 0ms 9 0ms 0ms 0ms 1,828 0ms update patternresultsrelevance set relevant = ?, saxo_relevant = ?, notrelevantpricedatetime = ?, reason = ? where uniqueindex = ? and relevant = ?;Times Reported Time consuming queries #9
Day Hour Count Duration Avg duration Jan 12 01 1,828 0ms 0ms 10 0ms 0ms 0ms 4 0ms select t.pricedatetime as pricedatetime, t.open as open, t.high as high, t.low as low, t.close "..." close, t.volume as volume, t.bsf as bsf from t1440 t where symbolid = ? and (bsf = ? or bsf is null) order by pricedatetime desc limit ?;Times Reported Time consuming queries #10
Day Hour Count Duration Avg duration Jan 12 01 4 0ms 0ms 11 0ms 0ms 0ms 18 0ms select cast(count(*) / cast(setting as numeric) * ? as int) from pg_stat_activity, pg_settings where name = ? group by setting;Times Reported Time consuming queries #11
Day Hour Count Duration Avg duration Jan 12 01 18 0ms 0ms 12 0ms 0ms 0ms 239 0ms select v.datname, c.relname, v.phase, v.heap_blks_total, v.heap_blks_scanned, v.heap_blks_vacuumed, v.index_vacuum_count, v.max_dead_tuples, v.num_dead_tuples from pg_stat_progress_vacuum as v join pg_class c on c.oid = v.relid;Times Reported Time consuming queries #12
Day Hour Count Duration Avg duration Jan 12 01 239 0ms 0ms 13 0ms 0ms 0ms 37 0ms select * from ( select pricedatetime, open, high, low, close, volume, bsf from t60 where symbolid = ? and (bsf = ? or bsf is null) order by pricedatetime desc limit ?) a order by pricedatetime asc;Times Reported Time consuming queries #13
Day Hour Count Duration Avg duration Jan 12 01 37 0ms 0ms 14 0ms 0ms 0ms 2 0ms with pre_symbols as ( select s.symbolid, s.symbol, s.timegranularity, dss.downloadersymbol, dtt.timezone, s.exchange, s.longname from symbols s inner join downloadersymbolsettings dss on dss.symbolid = s.symbolid inner join datafeedstimetable dtt on dss.classname = dtt.classname and dtt.dayofweek = ? inner join brokersymbollist bsl on bsl.symbolid = s.symbolid and brokerid = ? where dss.classname in (...) and (dss.downloadersymbol in (...) or s.symbol in (...)) and dss.enabled = ? and s.nonliquid = ? and s.deleted = ? ), report_symbols as ( select ps1.*, ps2.symbolid as price_symbol_id from pre_symbols ps1 inner join pre_symbols ps2 on ps1.symbol = ps2.symbol and ps1.downloadersymbol = ps2.downloadersymbol and ps2.timegranularity = ? ), rbr_max as ( select resultuid from relevance_bigmovement_results order by resultuid desc limit ? ) select bmr.resultuid as ruid, ? as direction, rs.symbol as sym, rs.downloadersymbol, rs.symbolid as sid, rs.timegranularity as tg, rs.timezone, rs.exchange as e, rs.longname, bmr.patternendtime as pet, lpi.latestpricedatetime as lpdt, bmr.patternlengthbars as l, bmr.breakout >= ? as complete, rbr.age as age from bigmovement_results bmr inner join report_symbols rs on rs.symbolid = bmr.symbolid inner join autochartist_symbolupdates lpi on lpi.symbolid = rs.price_symbol_id inner join rbr_max rm on ? = ? left outer join relevance_bigmovement_results rbr on rbr.resultuid = bmr.resultuid where patternendtime >= (now() - ? * interval ?) -- results can't be more than ? days old and (bmr.resultuid > rm.resultuid or rbr.relevant = ?) ;Times Reported Time consuming queries #14
Day Hour Count Duration Avg duration Jan 12 01 2 0ms 0ms 15 0ms 0ms 0ms 332 0ms commit;Times Reported Time consuming queries #15
Day Hour Count Duration Avg duration Jan 12 01 332 0ms 0ms 16 0ms 0ms 0ms 347 0ms with rar_max as ( select resultuid from relevance_keylevels_results order by resultuid desc limit ? ), kr as ( select a.*, rr.age, rr.relevant from keylevels_results a left outer join relevance_keylevels_results rr on a.resultuid = rr.resultuid where case when false = ? then true else a.resultuid > ( select min(resultuid) from relevance_keylevels_results) end ), all_results as ( select kr.resultuid as resultuid, kr.direction as direction, s.exchange as exchange, s.symbolid as symbolid, coalesce(bim.code, s.symbol) as symbol_code, s.longname as symbol_name, s.timegranularity as interval, p.patternname as pattern_name, kr.breakout as breakout, kr.atbaridentified as identified, dtt.timezone as timezone, kr.patternlengthbars as length, g.basegroupname, newlevels.filtered, case when kr.age is not null then kr.age when kr.resultuid <= rm.resultuid then ? else ? end as age, case when kr.relevant is not null then kr.relevant when kr.resultuid <= rm.resultuid then ? else ? end as relevant, cps.pip from kr inner join brokersymbollist bsl on bsl.brokerid = ? and bsl.symbolid = kr.symbolid inner join symbols s on bsl.symbolid = s.symbolid and s.nonliquid = ? inner join symbolgroup sg on s.symbolid = sg.symbolid inner join groups g on sg.groupid = g.groupid inner join brokergroups bg on g.groupid = bg.groupid and bsl.brokerid = bg.brokerid inner join hrspatterns p on kr.patternid = p.patternid inner join downloadersymbolsettings dss on s.symbolid = dss.symbolid inner join datafeedstimetable dtt on dss.classname = dtt.classname and dtt.dayofweek = ? inner join rar_max rm on ? = ? left outer join autochartist_symbolupdates au on dss.symbolid = au.symbolid left outer join relevance_keylevels_results rar on rar.resultuid = kr.resultuid left join lateral calc_kl_signal_filter (kr.resultuid) newlevels on true left join currencypips cps on cps.symbol = s.symbol left outer join brokerinstrumentmap bim on dss.datafeedinstrumentid = bim.datafeedinstrumentid and bim.brokerid = bsl.brokerid and bim.type = ? where kr.gmttimefound > now() - interval ? and dss.enabled = ? and s.deleted = ? and (kr.simulation = ? or kr.simulation is null) and (? = ? or s.timegranularity in (...)) and (? = ? or s.exchange in (...)) and (? = ? or coalesce(bim.code, s.symbol) in (...)) and (? = ? or p.patternname in (...)) and (? = ? or kr.patternclassid in (...)) and (? = ? or kr.patternlengthbars <= ?) and kr.patternstarttime::timestamp without time zone >= coalesce(au.earliestpricedatetime, ?::timestamp without time zone) -- to make sure patternstarttime is in our t-tables ), results as ( select distinct on (symbolid) * from all_results where (false = ? or relevant = ?) and (? = ? or age <= ?) order by symbolid, resultuid ) select * from results order by identified desc, length desc limit ?;Times Reported Time consuming queries #16
Day Hour Count Duration Avg duration Jan 12 01 347 0ms 0ms 17 0ms 0ms 0ms 15 0ms select psd.symbolid, psd.hourlysymbolid as intervalsymbolid, s.symbol, dtt.timezone, dtt.absolutetimezoneoffset, psh.hour as index, psh.ave, psh.stddev, ((psh.ave - psh.stddev) / ?.?) as low, ((psh.ave + psh.stddev) / ?.?) as high, ee.timezone as exchangetimezone, ee.mon_t1start as exchangestart, ee.mon_t1end as exchangeend from powerstats_symboldata psd inner join powerstats_hourly psh on psd.hourlysymbolid = psh.symbolid inner join symbols s on psh.symbolid = s.symbolid inner join downloadersymbolsettings dss on s.symbolid = dss.symbolid inner join datafeedstimetable dtt on dss.classname = dtt.classname inner join mat_ps_hourly_symbolid_max_enddate e on psh.enddate = e.enddate and psh.symbolid = e.symbolid inner join exchanges ee on ee.exchange = s.exchange where psd.symbolid = ? and dtt.dayofweek = ? order by hour asc;Times Reported Time consuming queries #17
Day Hour Count Duration Avg duration Jan 12 01 15 0ms 0ms 18 0ms 0ms 0ms 4 0ms select (cast(substring(tz.gmoffset from ? for ?) as float) * ? + cast(substring(tz.gmoffset from ? for ?) as float)) / ? as offset from timezones tz where tz.timezone = ?;Times Reported Time consuming queries #18
Day Hour Count Duration Avg duration Jan 12 01 4 0ms 0ms 19 0ms 0ms 0ms 239 0ms select count(*), sum(size), extract(epoch from now() - min(modification)) from pg_ls_waldir ();Times Reported Time consuming queries #19
Day Hour Count Duration Avg duration Jan 12 01 239 0ms 0ms 20 0ms 0ms 0ms 40 0ms select * from ( select pricedatetime, open, high, low, close, volume, bsf from t30 where symbolid = ? and (bsf = ? or bsf is null) order by pricedatetime desc limit ?) a order by pricedatetime asc;Times Reported Time consuming queries #20
Day Hour Count Duration Avg duration Jan 12 01 40 0ms 0ms Time consuming prepare
Rank Total duration Times executed Min duration Max duration Avg duration Query 1 5s412ms 4,969 0ms 13ms 1ms WITH rar_max as ( ;Times Reported Time consuming prepare #1
Day Hour Count Duration Avg duration Jan 12 01 4,969 5s412ms 1ms -
WITH rar_max as ( ;
Date: 2026-01-12 01:45:43 Duration: 13ms Database: postgres
-
WITH rar_max as ( ;
Date: 2026-01-12 01:08:02 Duration: 12ms Database: postgres
-
WITH rar_max as ( ;
Date: 2026-01-12 01:46:43 Duration: 12ms Database: postgres
2 5s156ms 7,816 0ms 15ms 0ms SELECT ;Times Reported Time consuming prepare #2
Day Hour Count Duration Avg duration 01 7,816 5s156ms 0ms -
SELECT ;
Date: 2026-01-12 01:09:39 Duration: 15ms Database: postgres
-
SELECT ;
Date: 2026-01-12 01:11:33 Duration: 14ms Database: postgres
-
SELECT ;
Date: 2026-01-12 01:23:06 Duration: 12ms Database: postgres
3 1s425ms 1,046 0ms 5ms 1ms SELECT symbolid, ;Times Reported Time consuming prepare #3
Day Hour Count Duration Avg duration 01 1,046 1s425ms 1ms -
SELECT symbolid, ;
Date: 2026-01-12 01:16:33 Duration: 5ms Database: postgres
-
SELECT symbolid, ;
Date: 2026-01-12 01:01:02 Duration: 2ms Database: postgres
-
SELECT symbolid, ;
Date: 2026-01-12 01:00:56 Duration: 2ms Database: postgres
4 586ms 4,506 0ms 5ms 0ms SET extra_float_digits = 3;Times Reported Time consuming prepare #4
Day Hour Count Duration Avg duration 01 4,506 586ms 0ms -
SET extra_float_digits = 3;
Date: 2026-01-12 01:00:04 Duration: 5ms Database: postgres
-
SET extra_float_digits = 3;
Date: 2026-01-12 01:35:40 Duration: 4ms Database: postgres
-
SET extra_float_digits = 3;
Date: 2026-01-12 01:22:06 Duration: 4ms Database: postgres
5 388ms 7,685 0ms 10ms 0ms select 1;Times Reported Time consuming prepare #5
Day Hour Count Duration Avg duration 01 7,685 388ms 0ms -
select 1;
Date: 2026-01-12 01:49:13 Duration: 10ms Database: postgres
-
select 1;
Date: 2026-01-12 01:26:37 Duration: 8ms Database: postgres
-
select 1;
Date: 2026-01-12 01:02:00 Duration: 4ms Database: postgres
6 358ms 358 0ms 1ms 1ms SELECT s.symbolid, dss.downloadfrequency, dss.downloadersymbol;Times Reported Time consuming prepare #6
Day Hour Count Duration Avg duration 01 358 358ms 1ms -
SELECT s.symbolid, dss.downloadfrequency, dss.downloadersymbol;
Date: 2026-01-12 01:30:20 Duration: 1ms Database: postgres
-
SELECT s.symbolid, dss.downloadfrequency, dss.downloadersymbol;
Date: 2026-01-12 01:30:39 Duration: 1ms Database: postgres
-
SELECT s.symbolid, dss.downloadfrequency, dss.downloadersymbol;
Date: 2026-01-12 01:45:45 Duration: 1ms Database: postgres
7 319ms 3,227 0ms 0ms 0ms INSERT INTO T60 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;Times Reported Time consuming prepare #7
Day Hour Count Duration Avg duration 01 3,227 319ms 0ms -
INSERT INTO T60 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;
Date: 2026-01-12 01:11:48 Duration: 0ms Database: postgres
-
INSERT INTO T60 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;
Date: 2026-01-12 01:11:38 Duration: 0ms Database: postgres
-
INSERT INTO T60 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;
Date: 2026-01-12 01:10:28 Duration: 0ms Database: postgres
8 313ms 3,875 0ms 0ms 0ms INSERT INTO T30 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;Times Reported Time consuming prepare #8
Day Hour Count Duration Avg duration 01 3,875 313ms 0ms -
INSERT INTO T30 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;
Date: 2026-01-12 01:11:38 Duration: 0ms Database: postgres
-
INSERT INTO T30 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;
Date: 2026-01-12 01:41:51 Duration: 0ms Database: postgres
-
INSERT INTO T30 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;
Date: 2026-01-12 01:41:53 Duration: 0ms Database: postgres
9 139ms 921 0ms 0ms 0ms INSERT INTO T15 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;Times Reported Time consuming prepare #9
Day Hour Count Duration Avg duration 01 921 139ms 0ms -
INSERT INTO T15 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;
Date: 2026-01-12 01:26:52 Duration: 0ms Database: postgres
-
INSERT INTO T15 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;
Date: 2026-01-12 01:56:52 Duration: 0ms Database: postgres
-
INSERT INTO T15 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;
Date: 2026-01-12 01:55:52 Duration: 0ms Database: postgres
10 80ms 677 0ms 0ms 0ms INSERT INTO T1440_underlying (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;Times Reported Time consuming prepare #10
Day Hour Count Duration Avg duration 01 677 80ms 0ms -
INSERT INTO T1440_underlying (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;
Date: 2026-01-12 01:02:36 Duration: 0ms Database: postgres
-
INSERT INTO T1440_underlying (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;
Date: 2026-01-12 01:02:03 Duration: 0ms Database: postgres
-
INSERT INTO T1440_underlying (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;
Date: 2026-01-12 01:01:54 Duration: 0ms Database: postgres
11 65ms 4,488 0ms 3ms 0ms SET application_name = 'PostgreSQL JDBC Driver';Times Reported Time consuming prepare #11
Day Hour Count Duration Avg duration 01 4,488 65ms 0ms -
SET application_name = 'PostgreSQL JDBC Driver';
Date: 2026-01-12 01:21:05 Duration: 3ms Database: postgres
-
SET application_name = 'PostgreSQL JDBC Driver';
Date: 2026-01-12 01:38:10 Duration: 2ms Database: postgres
-
SET application_name = 'PostgreSQL JDBC Driver';
Date: 2026-01-12 01:31:08 Duration: 1ms Database: postgres
12 59ms 659 0ms 0ms 0ms INSERT INTO T240 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;Times Reported Time consuming prepare #12
Day Hour Count Duration Avg duration 01 659 59ms 0ms -
INSERT INTO T240 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;
Date: 2026-01-12 01:01:52 Duration: 0ms Database: postgres
-
INSERT INTO T240 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;
Date: 2026-01-12 01:30:50 Duration: 0ms Database: postgres
-
INSERT INTO T240 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;
Date: 2026-01-12 01:46:07 Duration: 0ms Database: postgres
13 46ms 18 1ms 3ms 2ms select cast(count(*) / cast(setting as numeric) * 100 as int) from pg_stat_activity, pg_settings WHERE name = 'max_connections' group by setting;Times Reported Time consuming prepare #13
Day Hour Count Duration Avg duration 01 18 46ms 2ms -
select cast(count(*) / cast(setting as numeric) * 100 as int) from pg_stat_activity, pg_settings WHERE name = 'max_connections' group by setting;
Date: 2026-01-12 01:51:00 Duration: 3ms Database: postgres
-
select cast(count(*) / cast(setting as numeric) * 100 as int) from pg_stat_activity, pg_settings WHERE name = 'max_connections' group by setting;
Date: 2026-01-12 01:01:03 Duration: 2ms Database: postgres
-
select cast(count(*) / cast(setting as numeric) * 100 as int) from pg_stat_activity, pg_settings WHERE name = 'max_connections' group by setting;
Date: 2026-01-12 01:31:01 Duration: 2ms Database: postgres
14 14ms 6 2ms 3ms 2ms select client_addr, count(1) from pg_stat_activity, pg_settings WHERE name = 'max_connections' group by client_addr, setting having (client_addr is not null OR (client_addr is null and count(1) > (cast(setting as numeric) / 3 * 2))) order by count desc;Times Reported Time consuming prepare #14
Day Hour Count Duration Avg duration 01 6 14ms 2ms -
select client_addr, count(1) from pg_stat_activity, pg_settings WHERE name = 'max_connections' group by client_addr, setting having (client_addr is not null OR (client_addr is null and count(1) > (cast(setting as numeric) / 3 * 2))) order by count desc;
Date: 2026-01-12 01:40:05 Duration: 3ms Database: postgres
-
select client_addr, count(1) from pg_stat_activity, pg_settings WHERE name = 'max_connections' group by client_addr, setting having (client_addr is not null OR (client_addr is null and count(1) > (cast(setting as numeric) / 3 * 2))) order by count desc;
Date: 2026-01-12 01:00:04 Duration: 2ms Database: postgres
-
select client_addr, count(1) from pg_stat_activity, pg_settings WHERE name = 'max_connections' group by client_addr, setting having (client_addr is not null OR (client_addr is null and count(1) > (cast(setting as numeric) / 3 * 2))) order by count desc;
Date: 2026-01-12 01:50:05 Duration: 2ms Database: postgres
15 14ms 15 0ms 3ms 0ms WITH last_candle AS ( ;Times Reported Time consuming prepare #15
Day Hour Count Duration Avg duration 01 15 14ms 0ms -
WITH last_candle AS ( ;
Date: 2026-01-12 01:36:00 Duration: 3ms Database: postgres
-
WITH last_candle AS ( ;
Date: 2026-01-12 01:52:00 Duration: 2ms Database: postgres
-
WITH last_candle AS ( ;
Date: 2026-01-12 01:48:00 Duration: 2ms Database: postgres
16 13ms 6 2ms 2ms 2ms with rankedmt4 as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors ), last_feed_entry as ( select * from rankedmt4 where r = 1 ), ok_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where status = 'OK' ), earliest_entry_after_ok as ( select m.datafeedname, min(m.eventtimestamp) as eventtimestamp from mt4datafeederrors m left outer join ( select datafeedname, eventtimestamp from ok_entries where r = 1) oo on m.datafeedname = oo.datafeedname where m.eventtimestamp > coalesce(oo.eventtimestamp, '1900-01-01'::timestamp without time zone) group by m.datafeedname ), notified_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where notified is not null and notified <> '' ), broker as ( select *, row_number() over (partition by feedname order by brokerid) r from ( select distinct b.brokerid, b.name as brokername, dss.classname as feedname from downloadersymbolsettings dss inner join brokersymbollist bsl on dss.symbolid = bsl.symbolid inner join broker b on bsl.brokerid = b.brokerid where dss.enabled = 1) a ) select last.id, last.datafeedname, last.eventtimestamp, last.status, last.errordescription, last.serveraddress, last.username, note.notified, note.eventtimestamp, broker.brokername from last_feed_entry last inner join earliest_entry_after_ok after_ok on last.datafeedname = after_ok.datafeedname inner join broker on last.datafeedname = broker.feedname left outer join ok_entries ok on ok.datafeedname = last.datafeedname left outer join notified_entries note on note.datafeedname = last.datafeedname and note.r = 1 where (ok.r is null or ok.r = 1) and last.datafeedname not in ( select distinct datafeedname from last_feed_entry where status = 'OK') and extract(epoch from (last.eventtimestamp - after_ok.eventtimestamp)) > 60 * 60 and last.eventtimestamp > current_timestamp - interval '1 day' and (note.eventtimestamp is null or note.eventtimestamp < current_timestamp - interval '10 hours') and last.eventtimestamp > current_timestamp - interval '1 hour' and broker.r = 1;Times Reported Time consuming prepare #16
Day Hour Count Duration Avg duration 01 6 13ms 2ms -
with rankedmt4 as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors ), last_feed_entry as ( select * from rankedmt4 where r = 1 ), ok_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where status = 'OK' ), earliest_entry_after_ok as ( select m.datafeedname, min(m.eventtimestamp) as eventtimestamp from mt4datafeederrors m left outer join ( select datafeedname, eventtimestamp from ok_entries where r = 1) oo on m.datafeedname = oo.datafeedname where m.eventtimestamp > coalesce(oo.eventtimestamp, '1900-01-01'::timestamp without time zone) group by m.datafeedname ), notified_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where notified is not null and notified <> '' ), broker as ( select *, row_number() over (partition by feedname order by brokerid) r from ( select distinct b.brokerid, b.name as brokername, dss.classname as feedname from downloadersymbolsettings dss inner join brokersymbollist bsl on dss.symbolid = bsl.symbolid inner join broker b on bsl.brokerid = b.brokerid where dss.enabled = 1) a ) select last.id, last.datafeedname, last.eventtimestamp, last.status, last.errordescription, last.serveraddress, last.username, note.notified, note.eventtimestamp, broker.brokername from last_feed_entry last inner join earliest_entry_after_ok after_ok on last.datafeedname = after_ok.datafeedname inner join broker on last.datafeedname = broker.feedname left outer join ok_entries ok on ok.datafeedname = last.datafeedname left outer join notified_entries note on note.datafeedname = last.datafeedname and note.r = 1 where (ok.r is null or ok.r = 1) and last.datafeedname not in ( select distinct datafeedname from last_feed_entry where status = 'OK') and extract(epoch from (last.eventtimestamp - after_ok.eventtimestamp)) > 60 * 60 and last.eventtimestamp > current_timestamp - interval '1 day' and (note.eventtimestamp is null or note.eventtimestamp < current_timestamp - interval '10 hours') and last.eventtimestamp > current_timestamp - interval '1 hour' and broker.r = 1;
Date: 2026-01-12 01:10:02 Duration: 2ms Database: postgres
-
with rankedmt4 as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors ), last_feed_entry as ( select * from rankedmt4 where r = 1 ), ok_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where status = 'OK' ), earliest_entry_after_ok as ( select m.datafeedname, min(m.eventtimestamp) as eventtimestamp from mt4datafeederrors m left outer join ( select datafeedname, eventtimestamp from ok_entries where r = 1) oo on m.datafeedname = oo.datafeedname where m.eventtimestamp > coalesce(oo.eventtimestamp, '1900-01-01'::timestamp without time zone) group by m.datafeedname ), notified_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where notified is not null and notified <> '' ), broker as ( select *, row_number() over (partition by feedname order by brokerid) r from ( select distinct b.brokerid, b.name as brokername, dss.classname as feedname from downloadersymbolsettings dss inner join brokersymbollist bsl on dss.symbolid = bsl.symbolid inner join broker b on bsl.brokerid = b.brokerid where dss.enabled = 1) a ) select last.id, last.datafeedname, last.eventtimestamp, last.status, last.errordescription, last.serveraddress, last.username, note.notified, note.eventtimestamp, broker.brokername from last_feed_entry last inner join earliest_entry_after_ok after_ok on last.datafeedname = after_ok.datafeedname inner join broker on last.datafeedname = broker.feedname left outer join ok_entries ok on ok.datafeedname = last.datafeedname left outer join notified_entries note on note.datafeedname = last.datafeedname and note.r = 1 where (ok.r is null or ok.r = 1) and last.datafeedname not in ( select distinct datafeedname from last_feed_entry where status = 'OK') and extract(epoch from (last.eventtimestamp - after_ok.eventtimestamp)) > 60 * 60 and last.eventtimestamp > current_timestamp - interval '1 day' and (note.eventtimestamp is null or note.eventtimestamp < current_timestamp - interval '10 hours') and last.eventtimestamp > current_timestamp - interval '1 hour' and broker.r = 1;
Date: 2026-01-12 01:00:02 Duration: 2ms Database: postgres
-
with rankedmt4 as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors ), last_feed_entry as ( select * from rankedmt4 where r = 1 ), ok_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where status = 'OK' ), earliest_entry_after_ok as ( select m.datafeedname, min(m.eventtimestamp) as eventtimestamp from mt4datafeederrors m left outer join ( select datafeedname, eventtimestamp from ok_entries where r = 1) oo on m.datafeedname = oo.datafeedname where m.eventtimestamp > coalesce(oo.eventtimestamp, '1900-01-01'::timestamp without time zone) group by m.datafeedname ), notified_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where notified is not null and notified <> '' ), broker as ( select *, row_number() over (partition by feedname order by brokerid) r from ( select distinct b.brokerid, b.name as brokername, dss.classname as feedname from downloadersymbolsettings dss inner join brokersymbollist bsl on dss.symbolid = bsl.symbolid inner join broker b on bsl.brokerid = b.brokerid where dss.enabled = 1) a ) select last.id, last.datafeedname, last.eventtimestamp, last.status, last.errordescription, last.serveraddress, last.username, note.notified, note.eventtimestamp, broker.brokername from last_feed_entry last inner join earliest_entry_after_ok after_ok on last.datafeedname = after_ok.datafeedname inner join broker on last.datafeedname = broker.feedname left outer join ok_entries ok on ok.datafeedname = last.datafeedname left outer join notified_entries note on note.datafeedname = last.datafeedname and note.r = 1 where (ok.r is null or ok.r = 1) and last.datafeedname not in ( select distinct datafeedname from last_feed_entry where status = 'OK') and extract(epoch from (last.eventtimestamp - after_ok.eventtimestamp)) > 60 * 60 and last.eventtimestamp > current_timestamp - interval '1 day' and (note.eventtimestamp is null or note.eventtimestamp < current_timestamp - interval '10 hours') and last.eventtimestamp > current_timestamp - interval '1 hour' and broker.r = 1;
Date: 2026-01-12 01:30:02 Duration: 2ms Database: postgres
17 13ms 8 0ms 3ms 1ms WITH pre_symbols AS ( /* find relevant symbols */ ;Times Reported Time consuming prepare #17
Day Hour Count Duration Avg duration 01 8 13ms 1ms -
WITH pre_symbols AS ( /* find relevant symbols */ ;
Date: 2026-01-12 01:01:08 Duration: 3ms Database: postgres
-
WITH pre_symbols AS ( /* find relevant symbols */ ;
Date: 2026-01-12 01:14:26 Duration: 2ms Database: postgres
-
WITH pre_symbols AS ( /* find relevant symbols */ ;
Date: 2026-01-12 01:14:26 Duration: 2ms Database: postgres
18 6ms 34 0ms 0ms 0ms SELECT NULL AS TABLE_CAT, n.nspname AS TABLE_SCHEM, c.relname AS TABLE_NAME, CASE n.nspname ~ '^pg_' OR n.nspname = 'information_schema' WHEN true THEN CASE WHEN n.nspname = 'pg_catalog' OR n.nspname = 'information_schema' THEN CASE c.relkind WHEN 'r' THEN 'SYSTEM TABLE' WHEN 'v' THEN 'SYSTEM VIEW' WHEN 'i' THEN 'SYSTEM INDEX' ELSE NULL END WHEN n.nspname = 'pg_toast' THEN CASE c.relkind WHEN 'r' THEN 'SYSTEM TOAST TABLE' WHEN 'i' THEN 'SYSTEM TOAST INDEX' ELSE NULL END ELSE CASE c.relkind WHEN 'r' THEN 'TEMPORARY TABLE' WHEN 'p' THEN 'TEMPORARY TABLE' WHEN 'i' THEN 'TEMPORARY INDEX' WHEN 'S' THEN 'TEMPORARY SEQUENCE' WHEN 'v' THEN 'TEMPORARY VIEW' ELSE NULL END END WHEN false THEN CASE c.relkind WHEN 'r' THEN 'TABLE' WHEN 'p' THEN 'PARTITIONED TABLE' WHEN 'i' THEN 'INDEX' WHEN 'S' THEN 'SEQUENCE' WHEN 'v' THEN 'VIEW' WHEN 'c' THEN 'TYPE' WHEN 'f' THEN 'FOREIGN TABLE' WHEN 'm' THEN 'MATERIALIZED VIEW' ELSE NULL END ELSE NULL END AS TABLE_TYPE, d.description AS REMARKS, '' as TYPE_CAT, '' as TYPE_SCHEM, '' as TYPE_NAME, '' AS SELF_REFERENCING_COL_NAME, '' AS REF_GENERATION FROM pg_catalog.pg_namespace n, pg_catalog.pg_class c LEFT JOIN pg_catalog.pg_description d ON (c.oid = d.objoid AND d.objsubid = 0) LEFT JOIN pg_catalog.pg_class dc ON (d.classoid = dc.oid AND dc.relname = 'pg_class') LEFT JOIN pg_catalog.pg_namespace dn ON (dn.oid = dc.relnamespace AND dn.nspname = 'pg_catalog') WHERE c.relnamespace = n.oid AND c.relname LIKE 'PROBABLYNOT' AND (false OR (c.relkind = 'r' AND n.nspname !~ '^pg_' AND n.nspname <> 'information_schema')) ORDER BY TABLE_TYPE, TABLE_SCHEM, TABLE_NAME;Times Reported Time consuming prepare #18
Day Hour Count Duration Avg duration 01 34 6ms 0ms -
SELECT NULL AS TABLE_CAT, n.nspname AS TABLE_SCHEM, c.relname AS TABLE_NAME, CASE n.nspname ~ '^pg_' OR n.nspname = 'information_schema' WHEN true THEN CASE WHEN n.nspname = 'pg_catalog' OR n.nspname = 'information_schema' THEN CASE c.relkind WHEN 'r' THEN 'SYSTEM TABLE' WHEN 'v' THEN 'SYSTEM VIEW' WHEN 'i' THEN 'SYSTEM INDEX' ELSE NULL END WHEN n.nspname = 'pg_toast' THEN CASE c.relkind WHEN 'r' THEN 'SYSTEM TOAST TABLE' WHEN 'i' THEN 'SYSTEM TOAST INDEX' ELSE NULL END ELSE CASE c.relkind WHEN 'r' THEN 'TEMPORARY TABLE' WHEN 'p' THEN 'TEMPORARY TABLE' WHEN 'i' THEN 'TEMPORARY INDEX' WHEN 'S' THEN 'TEMPORARY SEQUENCE' WHEN 'v' THEN 'TEMPORARY VIEW' ELSE NULL END END WHEN false THEN CASE c.relkind WHEN 'r' THEN 'TABLE' WHEN 'p' THEN 'PARTITIONED TABLE' WHEN 'i' THEN 'INDEX' WHEN 'S' THEN 'SEQUENCE' WHEN 'v' THEN 'VIEW' WHEN 'c' THEN 'TYPE' WHEN 'f' THEN 'FOREIGN TABLE' WHEN 'm' THEN 'MATERIALIZED VIEW' ELSE NULL END ELSE NULL END AS TABLE_TYPE, d.description AS REMARKS, '' as TYPE_CAT, '' as TYPE_SCHEM, '' as TYPE_NAME, '' AS SELF_REFERENCING_COL_NAME, '' AS REF_GENERATION FROM pg_catalog.pg_namespace n, pg_catalog.pg_class c LEFT JOIN pg_catalog.pg_description d ON (c.oid = d.objoid AND d.objsubid = 0) LEFT JOIN pg_catalog.pg_class dc ON (d.classoid = dc.oid AND dc.relname = 'pg_class') LEFT JOIN pg_catalog.pg_namespace dn ON (dn.oid = dc.relnamespace AND dn.nspname = 'pg_catalog') WHERE c.relnamespace = n.oid AND c.relname LIKE 'PROBABLYNOT' AND (false OR (c.relkind = 'r' AND n.nspname !~ '^pg_' AND n.nspname <> 'information_schema')) ORDER BY TABLE_TYPE, TABLE_SCHEM, TABLE_NAME;
Date: 2026-01-12 01:14:30 Duration: 0ms Database: postgres
-
SELECT NULL AS TABLE_CAT, n.nspname AS TABLE_SCHEM, c.relname AS TABLE_NAME, CASE n.nspname ~ '^pg_' OR n.nspname = 'information_schema' WHEN true THEN CASE WHEN n.nspname = 'pg_catalog' OR n.nspname = 'information_schema' THEN CASE c.relkind WHEN 'r' THEN 'SYSTEM TABLE' WHEN 'v' THEN 'SYSTEM VIEW' WHEN 'i' THEN 'SYSTEM INDEX' ELSE NULL END WHEN n.nspname = 'pg_toast' THEN CASE c.relkind WHEN 'r' THEN 'SYSTEM TOAST TABLE' WHEN 'i' THEN 'SYSTEM TOAST INDEX' ELSE NULL END ELSE CASE c.relkind WHEN 'r' THEN 'TEMPORARY TABLE' WHEN 'p' THEN 'TEMPORARY TABLE' WHEN 'i' THEN 'TEMPORARY INDEX' WHEN 'S' THEN 'TEMPORARY SEQUENCE' WHEN 'v' THEN 'TEMPORARY VIEW' ELSE NULL END END WHEN false THEN CASE c.relkind WHEN 'r' THEN 'TABLE' WHEN 'p' THEN 'PARTITIONED TABLE' WHEN 'i' THEN 'INDEX' WHEN 'S' THEN 'SEQUENCE' WHEN 'v' THEN 'VIEW' WHEN 'c' THEN 'TYPE' WHEN 'f' THEN 'FOREIGN TABLE' WHEN 'm' THEN 'MATERIALIZED VIEW' ELSE NULL END ELSE NULL END AS TABLE_TYPE, d.description AS REMARKS, '' as TYPE_CAT, '' as TYPE_SCHEM, '' as TYPE_NAME, '' AS SELF_REFERENCING_COL_NAME, '' AS REF_GENERATION FROM pg_catalog.pg_namespace n, pg_catalog.pg_class c LEFT JOIN pg_catalog.pg_description d ON (c.oid = d.objoid AND d.objsubid = 0) LEFT JOIN pg_catalog.pg_class dc ON (d.classoid = dc.oid AND dc.relname = 'pg_class') LEFT JOIN pg_catalog.pg_namespace dn ON (dn.oid = dc.relnamespace AND dn.nspname = 'pg_catalog') WHERE c.relnamespace = n.oid AND c.relname LIKE 'PROBABLYNOT' AND (false OR (c.relkind = 'r' AND n.nspname !~ '^pg_' AND n.nspname <> 'information_schema')) ORDER BY TABLE_TYPE, TABLE_SCHEM, TABLE_NAME;
Date: 2026-01-12 01:14:28 Duration: 0ms Database: postgres
-
SELECT NULL AS TABLE_CAT, n.nspname AS TABLE_SCHEM, c.relname AS TABLE_NAME, CASE n.nspname ~ '^pg_' OR n.nspname = 'information_schema' WHEN true THEN CASE WHEN n.nspname = 'pg_catalog' OR n.nspname = 'information_schema' THEN CASE c.relkind WHEN 'r' THEN 'SYSTEM TABLE' WHEN 'v' THEN 'SYSTEM VIEW' WHEN 'i' THEN 'SYSTEM INDEX' ELSE NULL END WHEN n.nspname = 'pg_toast' THEN CASE c.relkind WHEN 'r' THEN 'SYSTEM TOAST TABLE' WHEN 'i' THEN 'SYSTEM TOAST INDEX' ELSE NULL END ELSE CASE c.relkind WHEN 'r' THEN 'TEMPORARY TABLE' WHEN 'p' THEN 'TEMPORARY TABLE' WHEN 'i' THEN 'TEMPORARY INDEX' WHEN 'S' THEN 'TEMPORARY SEQUENCE' WHEN 'v' THEN 'TEMPORARY VIEW' ELSE NULL END END WHEN false THEN CASE c.relkind WHEN 'r' THEN 'TABLE' WHEN 'p' THEN 'PARTITIONED TABLE' WHEN 'i' THEN 'INDEX' WHEN 'S' THEN 'SEQUENCE' WHEN 'v' THEN 'VIEW' WHEN 'c' THEN 'TYPE' WHEN 'f' THEN 'FOREIGN TABLE' WHEN 'm' THEN 'MATERIALIZED VIEW' ELSE NULL END ELSE NULL END AS TABLE_TYPE, d.description AS REMARKS, '' as TYPE_CAT, '' as TYPE_SCHEM, '' as TYPE_NAME, '' AS SELF_REFERENCING_COL_NAME, '' AS REF_GENERATION FROM pg_catalog.pg_namespace n, pg_catalog.pg_class c LEFT JOIN pg_catalog.pg_description d ON (c.oid = d.objoid AND d.objsubid = 0) LEFT JOIN pg_catalog.pg_class dc ON (d.classoid = dc.oid AND dc.relname = 'pg_class') LEFT JOIN pg_catalog.pg_namespace dn ON (dn.oid = dc.relnamespace AND dn.nspname = 'pg_catalog') WHERE c.relnamespace = n.oid AND c.relname LIKE 'PROBABLYNOT' AND (false OR (c.relkind = 'r' AND n.nspname !~ '^pg_' AND n.nspname <> 'information_schema')) ORDER BY TABLE_TYPE, TABLE_SCHEM, TABLE_NAME;
Date: 2026-01-12 01:14:30 Duration: 0ms Database: postgres
19 4ms 24 0ms 0ms 0ms WITH tr_max AS ( ;Times Reported Time consuming prepare #19
Day Hour Count Duration Avg duration 01 24 4ms 0ms -
WITH tr_max AS ( ;
Date: 2026-01-12 01:14:29 Duration: 0ms Database: postgres
-
WITH tr_max AS ( ;
Date: 2026-01-12 01:14:28 Duration: 0ms Database: postgres
-
WITH tr_max AS ( ;
Date: 2026-01-12 01:14:27 Duration: 0ms Database: postgres
20 3ms 2 1ms 2ms 1ms select * from datafeed_symbols where classname ilike 'ICMARKETS-AU-MT5' limit 5000;Times Reported Time consuming prepare #20
Day Hour Count Duration Avg duration 01 2 3ms 1ms -
select * from datafeed_symbols where classname ilike 'ICMARKETS-AU-MT5' limit 5000;
Date: 2026-01-12 01:06:19 Duration: 2ms Database: postgres
-
select * from datafeed_symbols where classname ilike 'ICMARKETS-AU-MT5' limit 5000;
Date: 2026-01-12 01:50:17 Duration: 1ms Database: postgres
Time consuming bind
Rank Total duration Times executed Min duration Max duration Avg duration Query 1 43s534ms 8,306 0ms 57ms 5ms WITH rar_max as ( ;Times Reported Time consuming bind #1
Day Hour Count Duration Avg duration Jan 12 01 8,306 43s534ms 5ms -
WITH rar_max as ( ;
Date: 2026-01-12 01:00:04 Duration: 57ms Database: postgres parameters: $1 = '607475022556447301', $2 = '607475022556447301', $3 = '607475022556447301'
-
WITH rar_max as ( ;
Date: 2026-01-12 01:32:09 Duration: 34ms Database: postgres parameters: $1 = '689', $2 = '7', $3 = '15', $4 = '30', $5 = '60', $6 = '120', $7 = '240', $8 = '480', $9 = '1440', $10 = '0', $11 = '', $12 = '310', $13 = '#AAPL', $14 = '#ADS', $15 = '#AIG', $16 = '#ALV', $17 = '#AMZN', $18 = '#AXP', $19 = '#BA', $20 = '#BABA', $21 = '#BAC', $22 = '#BAS', $23 = '#BAYN', $24 = '#BEI', $25 = '#BIDU', $26 = '#BMW', $27 = '#C', $28 = '#CAT', $29 = '#CBK', $30 = '#CL', $31 = '#CSCO', $32 = '#CVX', $33 = '#DAI', $34 = '#DB1', $35 = '#DBK', $36 = '#DIS', $37 = '#DPW', $38 = '#DTE', $39 = '#EBAY', $40 = '#EON', $41 = '#F', $42 = '#FB', $43 = '#FDX', $44 = '#FME', $45 = '#GE', $46 = '#GM', $47 = '#GOOG', $48 = '#GS', $49 = '#HPQ', $50 = '#IBM', $51 = '#IFX', $52 = '#INTC', $53 = '#JD', $54 = '#JNJ', $55 = '#JPM', $56 = '#KO', $57 = '#LHA', $58 = '#LMT', $59 = '#MA', $60 = '#MCD', $61 = '#META', $62 = '#MMM', $63 = '#MSFT', $64 = '#MUV2', $65 = '#NFLX', $66 = '#NKE', $67 = '#NTES', $68 = '#ORCL', $69 = '#PFE', $70 = '#PG', $71 = '#QCOM', $72 = '#RACE', $73 = '#RWE', $74 = '#SAP', $75 = '#SIE', $76 = '#T', $77 = '#UBER', $78 = '#V', $79 = '#VOW', $80 = '#WB', $81 = '#XOM', $82 = 'AUDCAD', $83 = 'AUDCHF', $84 = 'AUDJPY', $85 = 'AUDNZD', $86 = 'AUDUSD', $87 = 'AUS200', $88 = 'BRENT', $89 = 'BTCUSD', $90 = 'CADCHF', $91 = 'CADJPY', $92 = 'CHFJPY', $93 = 'CHI50', $94 = 'ESP35', $95 = 'ETHUSD', $96 = 'EU50', $97 = 'EURAUD', $98 = 'EURCAD', $99 = 'EURCHF', $100 = 'EURGBP', $101 = 'EURHUF', $102 = 'EURJPY', $103 = 'EURNZD', $104 = 'EURPLN', $105 = 'EURUSD', $106 = 'FRA40', $107 = 'GBPAUD', $108 = 'GBPCAD', $109 = 'GBPCHF', $110 = 'GBPJPY', $111 = 'GBPNZD', $112 = 'GBPUSD', $113 = 'GER30', $114 = 'HK50', $115 = 'HKCH50', $116 = 'IT40', $117 = 'JP225', $118 = 'LTCUSD', $119 = 'NAS100', $120 = 'NZDCAD', $121 = 'NZDCHF', $122 = 'NZDJPY', $123 = 'NZDUSD', $124 = 'SPX500', $125 = 'UK100', $126 = 'US30', $127 = 'USDCAD', $128 = 'USDCHF', $129 = 'USDCNH', $130 = 'USDCZK', $131 = 'USDDKK', $132 = 'USDHKD', $133 = 'USDHUF', $134 = 'USDJPY', $135 = 'USDMXN', $136 = 'USDNOK', $137 = 'USDPLN', $138 = 'USDSEK', $139 = 'USDSGD', $140 = 'USDTRY', $141 = 'USDX', $142 = 'USDZAR', $143 = 'WTI', $144 = 'XAGUSD', $145 = 'XAUUSD', $146 = '#ADS', $147 = '#ALV', $148 = '#BAS', $149 = '#BAYN', $150 = '#BEI', $151 = '#BMW', $152 = '#CBK', $153 = '#DAI', $154 = '#DB1', $155 = '#DBK', $156 = '#DPW', $157 = '#DTE', $158 = '#EON', $159 = '#FME', $160 = '#IFX', $161 = '#LHA', $162 = '#MUV2', $163 = '#RWE', $164 = '#SAP', $165 = '#SIE', $166 = '#VOW', $167 = 'AUDCAD', $168 = 'AUDCHF', $169 = 'AUDJPY', $170 = 'AUDNZD', $171 = 'AUDUSD', $172 = 'CADCHF', $173 = 'CADJPY', $174 = 'CHFJPY', $175 = 'EURAUD', $176 = 'EURCAD', $177 = 'EURCHF', $178 = 'EURGBP', $179 = 'EURHUF', $180 = 'EURJPY', $181 = 'EURNZD', $182 = 'EURPLN', $183 = 'EURUSD', $184 = 'GBPAUD', $185 = 'GBPCAD', $186 = 'GBPCHF', $187 = 'GBPJPY', $188 = 'GBPNZD', $189 = 'GBPUSD', $190 = 'NZDCAD', $191 = 'NZDCHF', $192 = 'NZDJPY', $193 = 'NZDUSD', $194 = 'USDCAD', $195 = 'USDCHF', $196 = 'USDCNH', $197 = 'USDCZK', $198 = 'USDDKK', $199 = 'USDHKD', $200 = 'USDHUF', $201 = 'USDJPY', $202 = 'USDMXN', $203 = 'USDNOK', $204 = 'USDPLN', $205 = 'USDSEK', $206 = 'USDSGD', $207 = 'USDTRY', $208 = 'USDX', $209 = 'USDZAR', $210 = 'XAGUSD', $211 = 'XAUUSD', $212 = 'BTCUSD', $213 = 'ETHUSD', $214 = 'LTCUSD', $215 = 'AUDCAD', $216 = 'AUDCHF', $217 = 'AUDJPY', $218 = 'AUDNZD', $219 = 'CADCHF', $220 = 'CADJPY', $221 = 'CHFJPY', $222 = 'EURAUD', $223 = 'EURCAD', $224 = 'EURCHF', $225 = 'EURGBP', $226 = 'EURHUF', $227 = 'EURJPY', $228 = 'EURNZD', $229 = 'EURPLN', $230 = 'GBPAUD', $231 = 'GBPCAD', $232 = 'GBPCHF', $233 = 'GBPJPY', $234 = 'GBPNZD', $235 = 'NZDCAD', $236 = 'NZDCHF', $237 = 'NZDJPY', $238 = 'USDCNH', $239 = 'USDCZK', $240 = 'USDDKK', $241 = 'USDHKD', $242 = 'USDHUF', $243 = 'USDMXN', $244 = 'USDNOK', $245 = 'USDPLN', $246 = 'USDSEK', $247 = 'USDSGD', $248 = 'USDTRY', $249 = 'USDX', $250 = 'USDZAR', $251 = 'XAGUSD', $252 = 'XAUUSD', $253 = 'BRENT', $254 = 'WTI', $255 = 'AUS200', $256 = 'CHI50', $257 = 'ESP35', $258 = 'EU50', $259 = 'FRA40', $260 = 'GER30', $261 = 'HK50', $262 = 'HKCH50', $263 = 'IT40', $264 = 'JP225', $265 = 'NAS100', $266 = 'SPX500', $267 = 'UK100', $268 = 'US30', $269 = 'AUDUSD', $270 = 'EURUSD', $271 = 'GBPUSD', $272 = 'NZDUSD', $273 = 'USDCAD', $274 = 'USDCHF', $275 = 'USDJPY', $276 = '#AAPL', $277 = '#AIG', $278 = '#AMZN', $279 = '#AXP', $280 = '#BA', $281 = '#BABA', $282 = '#BAC', $283 = '#BIDU', $284 = '#C', $285 = '#CAT', $286 = '#CL', $287 = '#CSCO', $288 = '#CVX', $289 = '#DIS', $290 = '#EBAY', $291 = '#F', $292 = '#FB', $293 = '#FDX', $294 = '#GE', $295 = '#GM', $296 = '#GOOG', $297 = '#GS', $298 = '#HPQ', $299 = '#IBM', $300 = '#INTC', $301 = '#JD', $302 = '#JNJ', $303 = '#JPM', $304 = '#KO', $305 = '#LMT', $306 = '#MA', $307 = '#MCD', $308 = '#MMM', $309 = '#MSFT', $310 = '#NFLX', $311 = '#NKE', $312 = '#NTES', $313 = '#ORCL', $314 = '#PFE', $315 = '#PG', $316 = '#QCOM', $317 = '#RACE', $318 = '#T', $319 = '#UBER', $320 = '#V', $321 = '#WB', $322 = '#XOM', $323 = '700', $324 = '700', $325 = 't', $326 = '10', $327 = '10'
-
WITH rar_max as ( ;
Date: 2026-01-12 01:36:39 Duration: 34ms Database: postgres parameters: $1 = '558', $2 = '7', $3 = '15', $4 = '30', $5 = '60', $6 = '120', $7 = '240', $8 = '480', $9 = '1440', $10 = '0', $11 = '', $12 = '160', $13 = 'AUDSGD', $14 = 'CHFSGD', $15 = 'EURDKK', $16 = 'EURHKD', $17 = 'EURNOK', $18 = 'EURPLN', $19 = 'EURSEK', $20 = 'EURSGD', $21 = 'EURTRY', $22 = 'EURZAR', $23 = 'GBPDKK', $24 = 'GBPNOK', $25 = 'GBPSEK', $26 = 'GBPSGD', $27 = 'NOKJPY', $28 = 'NOKSEK', $29 = 'SEKJPY', $30 = 'SGDJPY', $31 = 'USDCNH', $32 = 'USDCZK', $33 = 'USDDKK', $34 = 'USDHKD', $35 = 'USDHUF', $36 = 'USDMXN', $37 = 'USDNOK', $38 = 'USDPLN', $39 = 'USDRUB', $40 = 'USDSEK', $41 = 'USDTHB', $42 = 'USDTRY', $43 = 'USDZAR', $44 = 'AUDUSD', $45 = 'EURUSD', $46 = 'GBPUSD', $47 = 'USDCAD', $48 = 'USDCHF', $49 = 'USDJPY', $50 = 'AUDCAD', $51 = 'AUDCHF', $52 = 'AUDJPY', $53 = 'AUDNZD', $54 = 'CADCHF', $55 = 'CADJPY', $56 = 'CHFJPY', $57 = 'EURAUD', $58 = 'EURCAD', $59 = 'EURCHF', $60 = 'EURGBP', $61 = 'EURJPY', $62 = 'EURNZD', $63 = 'GBPAUD', $64 = 'GBPCAD', $65 = 'GBPCHF', $66 = 'GBPJPY', $67 = 'GBPNZD', $68 = 'NZDCAD', $69 = 'NZDCHF', $70 = 'NZDJPY', $71 = 'NZDUSD', $72 = 'USDSGD', $73 = 'AUS200', $74 = 'DE30', $75 = 'ES35', $76 = 'F40', $77 = 'HK50', $78 = 'IT40', $79 = 'JP225', $80 = 'STOXX50', $81 = 'UK100', $82 = 'US2000', $83 = 'US30', $84 = 'US500', $85 = 'CHINA50', $86 = 'USTEC', $87 = 'XAGEUR', $88 = 'XAGUSD', $89 = 'XAUUSD', $90 = 'XAUEUR', $91 = 'XPDUSD', $92 = 'XPTUSD', $93 = 'AUDSGD', $94 = 'CHFSGD', $95 = 'EURDKK', $96 = 'EURHKD', $97 = 'EURNOK', $98 = 'EURPLN', $99 = 'EURSEK', $100 = 'EURSGD', $101 = 'EURTRY', $102 = 'EURZAR', $103 = 'GBPDKK', $104 = 'GBPNOK', $105 = 'GBPSEK', $106 = 'GBPSGD', $107 = 'NOKJPY', $108 = 'NOKSEK', $109 = 'SEKJPY', $110 = 'SGDJPY', $111 = 'USDCNH', $112 = 'USDCZK', $113 = 'USDDKK', $114 = 'USDHKD', $115 = 'USDHUF', $116 = 'USDMXN', $117 = 'USDNOK', $118 = 'USDPLN', $119 = 'USDRUB', $120 = 'USDSEK', $121 = 'USDTHB', $122 = 'USDTRY', $123 = 'USDZAR', $124 = 'AUDUSD', $125 = 'EURUSD', $126 = 'GBPUSD', $127 = 'USDCAD', $128 = 'USDCHF', $129 = 'USDJPY', $130 = 'AUDCAD', $131 = 'AUDCHF', $132 = 'AUDJPY', $133 = 'AUDNZD', $134 = 'CADCHF', $135 = 'CADJPY', $136 = 'CHFJPY', $137 = 'EURAUD', $138 = 'EURCAD', $139 = 'EURCHF', $140 = 'EURGBP', $141 = 'EURJPY', $142 = 'EURNZD', $143 = 'GBPAUD', $144 = 'GBPCAD', $145 = 'GBPCHF', $146 = 'GBPJPY', $147 = 'GBPNZD', $148 = 'NZDCAD', $149 = 'NZDCHF', $150 = 'NZDJPY', $151 = 'NZDUSD', $152 = 'USDSGD', $153 = 'AUS200', $154 = 'DE30', $155 = 'ES35', $156 = 'F40', $157 = 'HK50', $158 = 'IT40', $159 = 'JP225', $160 = 'STOXX50', $161 = 'UK100', $162 = 'US2000', $163 = 'US30', $164 = 'US500', $165 = 'CHINA50', $166 = 'USTEC', $167 = 'XAGEUR', $168 = 'XAGUSD', $169 = 'XAUUSD', $170 = 'XAUEUR', $171 = 'XPDUSD', $172 = 'XPTUSD', $173 = '700', $174 = '700', $175 = 't', $176 = '10', $177 = '10'
2 13s647ms 24,672 0ms 23ms 0ms SELECT ;Times Reported Time consuming bind #2
Day Hour Count Duration Avg duration 01 24,672 13s647ms 0ms -
SELECT ;
Date: 2026-01-12 01:00:04 Duration: 23ms Database: postgres parameters: $1 = '958', $2 = '958', $3 = '515840243129828300'
-
SELECT ;
Date: 2026-01-12 01:00:04 Duration: 22ms Database: postgres parameters: $1 = '958', $2 = '958', $3 = '515840217487435300'
-
SELECT ;
Date: 2026-01-12 01:11:33 Duration: 22ms Database: postgres parameters: $1 = '958', $2 = '958', $3 = '515840248167614300'
3 2s557ms 1,046 0ms 3ms 2ms SELECT symbolid, ;Times Reported Time consuming bind #3
Day Hour Count Duration Avg duration 01 1,046 2s557ms 2ms -
SELECT symbolid, ;
Date: 2026-01-12 01:31:49 Duration: 3ms Database: postgres parameters: $1 = 'GLOBALGTMT5', $2 = '15', $3 = 'NZDUSD', $4 = 'USOil', $5 = 'SHBUSD', $6 = 'ETHJPY', $7 = 'MATUSD', $8 = '#AMZN', $9 = 'BTCUSD', $10 = 'ETHUSD', $11 = 'UKOil', $12 = 'DOGUSD', $13 = 'USDJPY', $14 = 'EURUSD', $15 = 'LTCJPY', $16 = 'JP225', $17 = 'USDZAR', $18 = 'US30', $19 = 'BTCJPY', $20 = 'GBPUSD', $21 = 'ADAUSD', $22 = 'DOTBTC', $23 = 'EURGBP', $24 = 'EURJPY', $25 = 'BTCXAU', $26 = 'LTCUSD', $27 = 'BTCZAR', $28 = 'EURHUF', $29 = 'USDCHF', $30 = 'SOLUSD', $31 = 'BCHUSD', $32 = 'GTi12', $33 = 'AUDUSD', $34 = 'GBPJPY', $35 = 'BTCAPL', $36 = 'USDCZK', $37 = '#AAPL', $38 = 'AUDJPY', $39 = 'EURAUD', $40 = 'BCHJPY', $41 = 'USDCAD', $42 = 'BTCUSO'
-
SELECT symbolid, ;
Date: 2026-01-12 01:01:48 Duration: 3ms Database: postgres parameters: $1 = 'BDSWISS', $2 = '1440', $3 = 'GBPZAR', $4 = 'GBPUSD', $5 = 'HKDJPY'
-
SELECT symbolid, ;
Date: 2026-01-12 01:30:39 Duration: 3ms Database: postgres parameters: $1 = 'BDSWISS', $2 = '15', $3 = 'CADCHF', $4 = 'CADJPY', $5 = 'BTCUSD', $6 = 'BTCGBP', $7 = 'AUS_200', $8 = 'BTCEUR', $9 = 'CHFJPY', $10 = 'AUDUSD'
4 717ms 105 4ms 18ms 6ms WITH last_candle AS ( ;Times Reported Time consuming bind #4
Day Hour Count Duration Avg duration 01 105 717ms 6ms -
WITH last_candle AS ( ;
Date: 2026-01-12 01:48:00 Duration: 18ms Database: postgres parameters: $1 = '538', $2 = '538'
-
WITH last_candle AS ( ;
Date: 2026-01-12 01:16:03 Duration: 17ms Database: postgres parameters: $1 = '621', $2 = '621'
-
WITH last_candle AS ( ;
Date: 2026-01-12 01:36:00 Duration: 11ms Database: postgres parameters: $1 = '538', $2 = '538'
5 615ms 29,021 0ms 6ms 0ms select 1;Times Reported Time consuming bind #5
Day Hour Count Duration Avg duration 01 29,021 615ms 0ms -
select 1;
Date: 2026-01-12 01:02:00 Duration: 6ms Database: postgres
-
select 1;
Date: 2026-01-12 01:00:04 Duration: 3ms Database: postgres
-
select 1;
Date: 2026-01-12 01:35:40 Duration: 2ms Database: postgres
6 582ms 358 1ms 2ms 1ms SELECT s.symbolid, dss.downloadfrequency, dss.downloadersymbol;Times Reported Time consuming bind #6
Day Hour Count Duration Avg duration 01 358 582ms 1ms -
SELECT s.symbolid, dss.downloadfrequency, dss.downloadersymbol;
Date: 2026-01-12 01:30:20 Duration: 2ms Database: postgres parameters: $1 = 'PEPPERSTONEMT5'
-
SELECT s.symbolid, dss.downloadfrequency, dss.downloadersymbol;
Date: 2026-01-12 01:45:45 Duration: 2ms Database: postgres parameters: $1 = 'AXIORY'
-
SELECT s.symbolid, dss.downloadfrequency, dss.downloadersymbol;
Date: 2026-01-12 01:30:39 Duration: 2ms Database: postgres parameters: $1 = 'BDSWISS'
7 305ms 22 0ms 42ms 13ms with wh_patitioned as ( ;Times Reported Time consuming bind #7
Day Hour Count Duration Avg duration 01 22 305ms 13ms -
with wh_patitioned as ( ;
Date: 2026-01-12 01:16:21 Duration: 42ms Database: postgres parameters: $1 = '627', $2 = '627', $3 = '627', $4 = '627', $5 = '627', $6 = '627', $7 = '627', $8 = '627', $9 = '627'
-
with wh_patitioned as ( ;
Date: 2026-01-12 01:50:17 Duration: 26ms Database: postgres parameters: $1 = '627', $2 = '627', $3 = '627', $4 = '627', $5 = '627', $6 = '627', $7 = '627', $8 = '627', $9 = '627'
-
with wh_patitioned as ( ;
Date: 2026-01-12 01:30:02 Duration: 25ms Database: postgres parameters: $1 = '627', $2 = '627', $3 = '627', $4 = '627', $5 = '627', $6 = '627', $7 = '627', $8 = '627', $9 = '627'
8 285ms 4,054 0ms 0ms 0ms INSERT INTO T30 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;Times Reported Time consuming bind #8
Day Hour Count Duration Avg duration 01 4,054 285ms 0ms -
INSERT INTO T30 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;
Date: 2026-01-12 01:01:03 Duration: 0ms Database: postgres parameters: $1 = '2026-01-12 00:30:00', $2 = '9.20022', $3 = '9.2049', $4 = '9.198', $5 = '9.20097', $6 = '919', $7 = '515840217493571300', $8 = '0', $9 = '2026-01-12 01:01:03.344', $10 = '2026-01-12 01:01:03.235', $11 = '9.20022', $12 = '9.2049', $13 = '9.198', $14 = '9.20097', $15 = '919', $16 = '0', $17 = '2026-01-12 01:01:03.344', $18 = '2026-01-12 01:01:03.235'
-
INSERT INTO T30 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;
Date: 2026-01-12 01:45:59 Duration: 0ms Database: postgres parameters: $1 = '2026-01-09 22:30:00', $2 = '2.339045', $3 = '2.33913', $4 = '2.33803', $5 = '2.3383', $6 = '1836', $7 = '515840230500992300', $8 = '0', $9 = '2026-01-12 01:45:59.758', $10 = '2026-01-12 01:45:59.75', $11 = '2.339045', $12 = '2.33913', $13 = '2.33803', $14 = '2.3383', $15 = '1836', $16 = '0', $17 = '2026-01-12 01:45:59.758', $18 = '2026-01-12 01:45:59.75'
-
INSERT INTO T30 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;
Date: 2026-01-12 01:02:00 Duration: 0ms Database: postgres parameters: $1 = '2026-01-09 23:30:00', $2 = '6961.13', $3 = '6963.13', $4 = '6959.88', $5 = '6962.13', $6 = '157', $7 = '515840238284097300', $8 = '0', $9 = '2026-01-12 01:02:00.873', $10 = '2026-01-12 01:02:00.434', $11 = '6961.13', $12 = '6963.13', $13 = '6959.88', $14 = '6962.13', $15 = '157', $16 = '0', $17 = '2026-01-12 01:02:00.873', $18 = '2026-01-12 01:02:00.434'
9 276ms 3,447 0ms 0ms 0ms INSERT INTO T60 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;Times Reported Time consuming bind #9
Day Hour Count Duration Avg duration 01 3,447 276ms 0ms -
INSERT INTO T60 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;
Date: 2026-01-12 01:01:44 Duration: 0ms Database: postgres parameters: $1 = '2026-01-09 22:00:00', $2 = '1.16348', $3 = '1.16379', $4 = '1.16342', $5 = '1.16347', $6 = '913', $7 = '515840230404720300', $8 = '0', $9 = '2026-01-12 01:01:44.024', $10 = '2026-01-12 01:01:44.023', $11 = '1.16348', $12 = '1.16379', $13 = '1.16342', $14 = '1.16347', $15 = '913', $16 = '0', $17 = '2026-01-12 01:01:44.024', $18 = '2026-01-12 01:01:44.023'
-
INSERT INTO T60 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;
Date: 2026-01-12 01:01:51 Duration: 0ms Database: postgres parameters: $1 = '2026-01-10 00:00:00', $2 = '0.80078', $3 = '0.8013', $4 = '0.80067', $5 = '0.801195', $6 = '1302', $7 = '515840249378432300', $8 = '0', $9 = '2026-01-12 01:01:51.835', $10 = '2026-01-12 01:01:51.834', $11 = '0.80078', $12 = '0.8013', $13 = '0.80067', $14 = '0.801195', $15 = '1302', $16 = '0', $17 = '2026-01-12 01:01:51.835', $18 = '2026-01-12 01:01:51.834'
-
INSERT INTO T60 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;
Date: 2026-01-12 01:01:03 Duration: 0ms Database: postgres parameters: $1 = '2026-01-09 23:00:00', $2 = '3874.1', $3 = '3878.62', $4 = '3871.4', $5 = '3877.72', $6 = '6450', $7 = '515840217485108300', $8 = '0', $9 = '2026-01-12 01:01:03.763', $10 = '2026-01-12 01:01:03.345', $11 = '3874.1', $12 = '3878.62', $13 = '3871.4', $14 = '3877.72', $15 = '6450', $16 = '0', $17 = '2026-01-12 01:01:03.763', $18 = '2026-01-12 01:01:03.345'
10 222ms 5,618 0ms 0ms 0ms INSERT INTO T15 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;Times Reported Time consuming bind #10
Day Hour Count Duration Avg duration 01 5,618 222ms 0ms -
INSERT INTO T15 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;
Date: 2026-01-12 01:26:52 Duration: 0ms Database: postgres parameters: $1 = '2026-01-12 01:00:00', $2 = '49411.9', $3 = '49451.9', $4 = '49401.9', $5 = '49426.4', $6 = '2715', $7 = '515840248000537300', $8 = '0', $9 = '2026-01-12 01:26:52.359', $10 = '2026-01-12 01:26:52.295', $11 = '49411.9', $12 = '49451.9', $13 = '49401.9', $14 = '49426.4', $15 = '2715', $16 = '0', $17 = '2026-01-12 01:26:52.359', $18 = '2026-01-12 01:26:52.295'
-
INSERT INTO T15 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;
Date: 2026-01-12 01:56:52 Duration: 0ms Database: postgres parameters: $1 = '2026-01-12 01:30:00', $2 = '49440.15', $3 = '49444.4', $4 = '49420.15', $5 = '49425.6', $6 = '1470', $7 = '515840248000537300', $8 = '0', $9 = '2026-01-12 01:56:52.977', $10 = '2026-01-12 01:56:52.909', $11 = '49440.15', $12 = '49444.4', $13 = '49420.15', $14 = '49425.6', $15 = '1470', $16 = '0', $17 = '2026-01-12 01:56:52.977', $18 = '2026-01-12 01:56:52.909'
-
INSERT INTO T15 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;
Date: 2026-01-12 01:55:52 Duration: 0ms Database: postgres parameters: $1 = '2026-01-12 01:15:00', $2 = '49425.9', $3 = '49449.4', $4 = '49425.9', $5 = '49439.9', $6 = '1441', $7 = '515840248000537300', $8 = '0', $9 = '2026-01-12 01:55:52.681', $10 = '2026-01-12 01:55:52.613', $11 = '49425.9', $12 = '49449.4', $13 = '49425.9', $14 = '49439.9', $15 = '1441', $16 = '0', $17 = '2026-01-12 01:55:52.681', $18 = '2026-01-12 01:55:52.613'
11 108ms 23 0ms 11ms 4ms WITH /*Latest.JapSticks*/ all_results AS ( SELECT ;Times Reported Time consuming bind #11
Day Hour Count Duration Avg duration 01 23 108ms 4ms -
WITH /*Latest.JapSticks*/ all_results AS ( SELECT ;
Date: 2026-01-12 01:02:54 Duration: 11ms Database: postgres parameters: $1 = '667', $2 = '0', $3 = '0', $4 = '0', $5 = '', $6 = '0', $7 = '', $8 = '0', $9 = '', $10 = '0', $11 = '0'
-
WITH /*Latest.JapSticks*/ all_results AS ( SELECT ;
Date: 2026-01-12 01:43:41 Duration: 10ms Database: postgres parameters: $1 = '667', $2 = '0', $3 = '0', $4 = '0', $5 = '', $6 = '0', $7 = '', $8 = '0', $9 = '', $10 = '0', $11 = '0'
-
WITH /*Latest.JapSticks*/ all_results AS ( SELECT ;
Date: 2026-01-12 01:18:39 Duration: 10ms Database: postgres parameters: $1 = '667', $2 = '0', $3 = '0', $4 = '0', $5 = '', $6 = '0', $7 = '', $8 = '0', $9 = '', $10 = '0', $11 = '0'
12 79ms 849 0ms 0ms 0ms INSERT INTO T1440_underlying (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;Times Reported Time consuming bind #12
Day Hour Count Duration Avg duration 01 849 79ms 0ms -
INSERT INTO T1440_underlying (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;
Date: 2026-01-12 01:01:37 Duration: 0ms Database: postgres parameters: $1 = '2026-01-09 00:00:00', $2 = '2609.42', $3 = '2635.4', $4 = '2599.35', $5 = '2622.21', $6 = '119923', $7 = '515840233921455300', $8 = '0', $9 = '2026-01-12 01:01:37.432', $10 = '2026-01-12 01:01:37.432', $11 = '2609.42', $12 = '2635.4', $13 = '2599.35', $14 = '2622.21', $15 = '119923', $16 = '0', $17 = '2026-01-12 01:01:37.432', $18 = '2026-01-12 01:01:37.432'
-
INSERT INTO T1440_underlying (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;
Date: 2026-01-12 01:01:51 Duration: 0ms Database: postgres parameters: $1 = '2026-01-09 00:00:00', $2 = '689.8', $3 = '689.8', $4 = '685.4', $5 = '686.2', $6 = '2861', $7 = '515840248217339300', $8 = '0', $9 = '2026-01-12 01:01:51.154', $10 = '2026-01-12 01:01:51.154', $11 = '689.8', $12 = '689.8', $13 = '685.4', $14 = '686.2', $15 = '2861', $16 = '0', $17 = '2026-01-12 01:01:51.154', $18 = '2026-01-12 01:01:51.154'
-
INSERT INTO T1440_underlying (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;
Date: 2026-01-12 01:02:03 Duration: 0ms Database: postgres parameters: $1 = '2026-01-09 00:00:00', $2 = '3843.71', $3 = '3881.92', $4 = '3820.2', $5 = '3877.72', $6 = '219359', $7 = '515840216686339300', $8 = '0', $9 = '2026-01-12 01:02:03.149', $10 = '2026-01-12 01:02:03.149', $11 = '3843.71', $12 = '3881.92', $13 = '3820.2', $14 = '3877.72', $15 = '219359', $16 = '0', $17 = '2026-01-12 01:02:03.149', $18 = '2026-01-12 01:02:03.149'
13 60ms 659 0ms 0ms 0ms INSERT INTO T240 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;Times Reported Time consuming bind #13
Day Hour Count Duration Avg duration 01 659 60ms 0ms -
INSERT INTO T240 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;
Date: 2026-01-12 01:46:07 Duration: 0ms Database: postgres parameters: $1 = '2026-01-09 16:00:00', $2 = '52305', $3 = '53850', $4 = '52250', $5 = '53580', $6 = '5402', $7 = '515840230561612300', $8 = '0', $9 = '2026-01-12 01:46:07.907', $10 = '2026-01-12 01:46:07.906', $11 = '52305', $12 = '53850', $13 = '52250', $14 = '53580', $15 = '5402', $16 = '0', $17 = '2026-01-12 01:46:07.907', $18 = '2026-01-12 01:46:07.906'
-
INSERT INTO T240 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;
Date: 2026-01-12 01:30:50 Duration: 0ms Database: postgres parameters: $1 = '2026-01-09 16:00:00', $2 = '52305', $3 = '53850', $4 = '52250', $5 = '53580', $6 = '5402', $7 = '515840230561612300', $8 = '0', $9 = '2026-01-12 01:30:50.753', $10 = '2026-01-12 01:30:50.752', $11 = '52305', $12 = '53850', $13 = '52250', $14 = '53580', $15 = '5402', $16 = '0', $17 = '2026-01-12 01:30:50.753', $18 = '2026-01-12 01:30:50.752'
-
INSERT INTO T240 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;
Date: 2026-01-12 01:16:34 Duration: 0ms Database: postgres parameters: $1 = '2026-01-09 20:00:00', $2 = '25767.55', $3 = '25831.55', $4 = '25726.05', $5 = '25778.9', $6 = '50006', $7 = '500991628277892200', $8 = '0', $9 = '2026-01-12 01:16:34.148', $10 = '2026-01-12 01:16:33.997', $11 = '25767.55', $12 = '25831.55', $13 = '25726.05', $14 = '25778.9', $15 = '50006', $16 = '0', $17 = '2026-01-12 01:16:34.148', $18 = '2026-01-12 01:16:33.997'
14 43ms 10 3ms 5ms 4ms SELECT DISTINCT ON (basegroupname, symbol) ;Times Reported Time consuming bind #14
Day Hour Count Duration Avg duration 01 10 43ms 4ms -
SELECT DISTINCT ON (basegroupname, symbol) ;
Date: 2026-01-12 01:01:00 Duration: 5ms Database: postgres parameters: $1 = '627', $2 = '627'
-
SELECT DISTINCT ON (basegroupname, symbol) ;
Date: 2026-01-12 01:41:06 Duration: 5ms Database: postgres parameters: $1 = '627', $2 = '627'
-
SELECT DISTINCT ON (basegroupname, symbol) ;
Date: 2026-01-12 01:27:21 Duration: 4ms Database: postgres parameters: $1 = '689', $2 = '689'
15 42ms 51 0ms 1ms 0ms SELECT timegranularity FROM brokersymbollist bsl INNER JOIN symbols s ON bsl.symbolid = s.symbolid INNER JOIN downloadersymbolsettings dss on s.symbolid = dss.symbolid LEFT OUTER JOIN brokerinstrumentmapping bdfi ON bdfi.brokerid = $1 AND dss.datafeedinstrumentid = bdfi.datafeedinstrumentid WHERE s.nonliquid = 0 and s.deleted = 0 and dss.enabled = 1 AND s.symbol ILIKE $2 AND bsl.brokerid = $3 AND timegranularity >= 15 ORDER BY timegranularity LIMIT 1;Times Reported Time consuming bind #15
Day Hour Count Duration Avg duration 01 51 42ms 0ms -
SELECT timegranularity FROM brokersymbollist bsl INNER JOIN symbols s ON bsl.symbolid = s.symbolid INNER JOIN downloadersymbolsettings dss on s.symbolid = dss.symbolid LEFT OUTER JOIN brokerinstrumentmapping bdfi ON bdfi.brokerid = $1 AND dss.datafeedinstrumentid = bdfi.datafeedinstrumentid WHERE s.nonliquid = 0 and s.deleted = 0 and dss.enabled = 1 AND s.symbol ILIKE $2 AND bsl.brokerid = $3 AND timegranularity >= 15 ORDER BY timegranularity LIMIT 1;
Date: 2026-01-12 01:45:43 Duration: 1ms Database: postgres parameters: $1 = '558', $2 = 'USDSEK', $3 = '558'
-
SELECT timegranularity FROM brokersymbollist bsl INNER JOIN symbols s ON bsl.symbolid = s.symbolid INNER JOIN downloadersymbolsettings dss on s.symbolid = dss.symbolid LEFT OUTER JOIN brokerinstrumentmapping bdfi ON bdfi.brokerid = $1 AND dss.datafeedinstrumentid = bdfi.datafeedinstrumentid WHERE s.nonliquid = 0 and s.deleted = 0 and dss.enabled = 1 AND s.symbol ILIKE $2 AND bsl.brokerid = $3 AND timegranularity >= 15 ORDER BY timegranularity LIMIT 1;
Date: 2026-01-12 01:18:09 Duration: 1ms Database: postgres parameters: $1 = '538', $2 = 'EURGBPb', $3 = '538'
-
SELECT timegranularity FROM brokersymbollist bsl INNER JOIN symbols s ON bsl.symbolid = s.symbolid INNER JOIN downloadersymbolsettings dss on s.symbolid = dss.symbolid LEFT OUTER JOIN brokerinstrumentmapping bdfi ON bdfi.brokerid = $1 AND dss.datafeedinstrumentid = bdfi.datafeedinstrumentid WHERE s.nonliquid = 0 and s.deleted = 0 and dss.enabled = 1 AND s.symbol ILIKE $2 AND bsl.brokerid = $3 AND timegranularity >= 15 ORDER BY timegranularity LIMIT 1;
Date: 2026-01-12 01:17:27 Duration: 1ms Database: postgres parameters: $1 = '558', $2 = 'AUDUSD', $3 = '558'
16 38ms 8 2ms 9ms 4ms WITH pre_symbols AS ( /* find relevant symbols */ ;Times Reported Time consuming bind #16
Day Hour Count Duration Avg duration 01 8 38ms 4ms -
WITH pre_symbols AS ( /* find relevant symbols */ ;
Date: 2026-01-12 01:01:08 Duration: 9ms Database: postgres parameters: $1 = '619', $2 = 'AXIORY', $3 = 'EURUSD', $4 = 'USDJPY', $5 = 'GBPUSD', $6 = 'AUDUSD', $7 = 'USDCHF', $8 = 'USDCAD', $9 = 'NZDUSD', $10 = 'GBPJPY', $11 = 'EURJPY', $12 = 'EURCHF', $13 = 'GBPCHF', $14 = 'EURCAD', $15 = 'GBPCAD', $16 = 'EURNZD', $17 = 'GBPNZD', $18 = 'CADJPY', $19 = 'CADCHF', $20 = 'CHFJPY', $21 = 'NZDJPY', $22 = 'XAUUSD', $23 = 'XAGUSD', $24 = 'EURUSD', $25 = 'USDJPY', $26 = 'GBPUSD', $27 = 'AUDUSD', $28 = 'USDCHF', $29 = 'USDCAD', $30 = 'NZDUSD', $31 = 'GBPJPY', $32 = 'EURJPY', $33 = 'EURCHF', $34 = 'GBPCHF', $35 = 'EURCAD', $36 = 'GBPCAD', $37 = 'EURNZD', $38 = 'GBPNZD', $39 = 'CADJPY', $40 = 'CADCHF', $41 = 'CHFJPY', $42 = 'NZDJPY', $43 = 'XAUUSD', $44 = 'XAGUSD', $45 = '5'
-
WITH pre_symbols AS ( /* find relevant symbols */ ;
Date: 2026-01-12 01:14:26 Duration: 7ms Database: postgres parameters: $1 = '1018', $2 = 'ICMARKETS-AU-MT5', $3 = 'AAPL.NAS', $4 = 'ABBV.NYSE', $5 = 'AMCR.NYSE', $6 = 'AMP.NYSE', $7 = 'AMZN.NAS', $8 = 'ANZ.ASX', $9 = 'AUDJPY', $10 = 'AUDUSD', $11 = 'AUS200', $12 = 'BABA.NYSE', $13 = 'BIIB.NAS', $14 = 'BXB.ASX', $15 = 'CBA.ASX', $16 = 'CHINA50', $17 = 'CSL.ASX', $18 = 'DE30', $19 = 'ES35', $20 = 'EURCHF', $21 = 'EURGBP', $22 = 'EURUSD', $23 = 'F40', $24 = 'FMG.ASX', $25 = 'GBPJPY', $26 = 'GBPUSD', $27 = 'GOOG.NAS', $28 = 'HK50', $29 = 'IT40', $30 = 'JP225', $31 = 'KO.NYSE', $32 = 'MQG.ASX', $33 = 'MSFT.NAS', $34 = 'NAB.ASX', $35 = 'NFLX.NAS', $36 = 'PYPL.NAS', $37 = 'QBE.ASX', $38 = 'STOXX50', $39 = 'SUN.ASX', $40 = 'TCL.ASX', $41 = 'TLS.ASX', $42 = 'TSLA.NAS', $43 = 'UK100', $44 = 'UNH.NYSE', $45 = 'US2000', $46 = 'US30', $47 = 'US500', $48 = 'USDCAD', $49 = 'USDCHF', $50 = 'USDCNH', $51 = 'USDJPY', $52 = 'USTEC', $53 = 'WBC.ASX', $54 = 'WES.ASX', $55 = 'WOW.ASX', $56 = 'WPL.ASX', $57 = 'XAUEUR', $58 = 'XAUUSD', $59 = 'XBRUSD', $60 = 'XTIUSD', $61 = 'AAPL.NAS', $62 = 'ABBV.NYSE', $63 = 'AMCR.NYSE', $64 = 'AMP.NYSE', $65 = 'AMZN.NAS', $66 = 'ANZ.ASX', $67 = 'AUDJPY', $68 = 'AUDUSD', $69 = 'AUS200', $70 = 'BABA.NYSE', $71 = 'BIIB.NAS', $72 = 'BXB.ASX', $73 = 'CBA.ASX', $74 = 'CHINA50', $75 = 'CSL.ASX', $76 = 'DE30', $77 = 'ES35', $78 = 'EURCHF', $79 = 'EURGBP', $80 = 'EURUSD', $81 = 'F40', $82 = 'FMG.ASX', $83 = 'GBPJPY', $84 = 'GBPUSD', $85 = 'GOOG.NAS', $86 = 'HK50', $87 = 'IT40', $88 = 'JP225', $89 = 'KO.NYSE', $90 = 'MQG.ASX', $91 = 'MSFT.NAS', $92 = 'NAB.ASX', $93 = 'NFLX.NAS', $94 = 'PYPL.NAS', $95 = 'QBE.ASX', $96 = 'STOXX50', $97 = 'SUN.ASX', $98 = 'TCL.ASX', $99 = 'TLS.ASX', $100 = 'TSLA.NAS', $101 = 'UK100', $102 = 'UNH.NYSE', $103 = 'US2000', $104 = 'US30', $105 = 'US500', $106 = 'USDCAD', $107 = 'USDCHF', $108 = 'USDCNH', $109 = 'USDJPY', $110 = 'USTEC', $111 = 'WBC.ASX', $112 = 'WES.ASX', $113 = 'WOW.ASX', $114 = 'WPL.ASX', $115 = 'XAUEUR', $116 = 'XAUUSD', $117 = 'XBRUSD', $118 = 'XTIUSD', $119 = '5'
-
WITH pre_symbols AS ( /* find relevant symbols */ ;
Date: 2026-01-12 01:14:26 Duration: 5ms Database: postgres parameters: $1 = '1018', $2 = 'ICMARKETS-AU-MT5', $3 = 'AAPL.NAS', $4 = 'ABBV.NYSE', $5 = 'AMCR.NYSE', $6 = 'AMP.NYSE', $7 = 'AMZN.NAS', $8 = 'ANZ.ASX', $9 = 'AUDJPY', $10 = 'AUDUSD', $11 = 'AUS200', $12 = 'BABA.NYSE', $13 = 'BIIB.NAS', $14 = 'BXB.ASX', $15 = 'CBA.ASX', $16 = 'CHINA50', $17 = 'CSL.ASX', $18 = 'DE30', $19 = 'ES35', $20 = 'EURCHF', $21 = 'EURGBP', $22 = 'EURUSD', $23 = 'F40', $24 = 'FMG.ASX', $25 = 'GBPJPY', $26 = 'GBPUSD', $27 = 'GOOG.NAS', $28 = 'HK50', $29 = 'IT40', $30 = 'JP225', $31 = 'KO.NYSE', $32 = 'MQG.ASX', $33 = 'MSFT.NAS', $34 = 'NAB.ASX', $35 = 'NFLX.NAS', $36 = 'PYPL.NAS', $37 = 'QBE.ASX', $38 = 'STOXX50', $39 = 'SUN.ASX', $40 = 'TCL.ASX', $41 = 'TLS.ASX', $42 = 'TSLA.NAS', $43 = 'UK100', $44 = 'UNH.NYSE', $45 = 'US2000', $46 = 'US30', $47 = 'US500', $48 = 'USDCAD', $49 = 'USDCHF', $50 = 'USDCNH', $51 = 'USDJPY', $52 = 'USTEC', $53 = 'WBC.ASX', $54 = 'WES.ASX', $55 = 'WOW.ASX', $56 = 'WPL.ASX', $57 = 'XAUEUR', $58 = 'XAUUSD', $59 = 'XBRUSD', $60 = 'XTIUSD', $61 = 'AAPL.NAS', $62 = 'ABBV.NYSE', $63 = 'AMCR.NYSE', $64 = 'AMP.NYSE', $65 = 'AMZN.NAS', $66 = 'ANZ.ASX', $67 = 'AUDJPY', $68 = 'AUDUSD', $69 = 'AUS200', $70 = 'BABA.NYSE', $71 = 'BIIB.NAS', $72 = 'BXB.ASX', $73 = 'CBA.ASX', $74 = 'CHINA50', $75 = 'CSL.ASX', $76 = 'DE30', $77 = 'ES35', $78 = 'EURCHF', $79 = 'EURGBP', $80 = 'EURUSD', $81 = 'F40', $82 = 'FMG.ASX', $83 = 'GBPJPY', $84 = 'GBPUSD', $85 = 'GOOG.NAS', $86 = 'HK50', $87 = 'IT40', $88 = 'JP225', $89 = 'KO.NYSE', $90 = 'MQG.ASX', $91 = 'MSFT.NAS', $92 = 'NAB.ASX', $93 = 'NFLX.NAS', $94 = 'PYPL.NAS', $95 = 'QBE.ASX', $96 = 'STOXX50', $97 = 'SUN.ASX', $98 = 'TCL.ASX', $99 = 'TLS.ASX', $100 = 'TSLA.NAS', $101 = 'UK100', $102 = 'UNH.NYSE', $103 = 'US2000', $104 = 'US30', $105 = 'US500', $106 = 'USDCAD', $107 = 'USDCHF', $108 = 'USDCNH', $109 = 'USDJPY', $110 = 'USTEC', $111 = 'WBC.ASX', $112 = 'WES.ASX', $113 = 'WOW.ASX', $114 = 'WPL.ASX', $115 = 'XAUEUR', $116 = 'XAUUSD', $117 = 'XBRUSD', $118 = 'XTIUSD'
17 28ms 37 0ms 1ms 0ms WITH tr_max AS ( ;Times Reported Time consuming bind #17
Day Hour Count Duration Avg duration 01 37 28ms 0ms -
WITH tr_max AS ( ;
Date: 2026-01-12 01:14:28 Duration: 1ms Database: postgres parameters: $1 = '605679104904728300', $2 = '605679104904728300', $3 = '605679104904728300', $4 = '5', $5 = '5', $6 = '1320', $7 = 't', $8 = '1320'
-
WITH tr_max AS ( ;
Date: 2026-01-12 01:14:29 Duration: 1ms Database: postgres parameters: $1 = '606715251039980300', $2 = '606715251039980300', $3 = '606715251039980300', $4 = '5', $5 = '5', $6 = '1320', $7 = 't', $8 = '1320'
-
WITH tr_max AS ( ;
Date: 2026-01-12 01:14:28 Duration: 1ms Database: postgres parameters: $1 = '605679104881405300', $2 = '605679104881405300', $3 = '605679104881405300', $4 = '1', $5 = '1', $6 = '0', $7 = 't', $8 = '0'
18 24ms 6 3ms 5ms 4ms with rankedmt4 as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors ), last_feed_entry as ( select * from rankedmt4 where r = 1 ), ok_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where status = 'OK' ), earliest_entry_after_ok as ( select m.datafeedname, min(m.eventtimestamp) as eventtimestamp from mt4datafeederrors m left outer join ( select datafeedname, eventtimestamp from ok_entries where r = 1) oo on m.datafeedname = oo.datafeedname where m.eventtimestamp > coalesce(oo.eventtimestamp, '1900-01-01'::timestamp without time zone) group by m.datafeedname ), notified_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where notified is not null and notified <> '' ), broker as ( select *, row_number() over (partition by feedname order by brokerid) r from ( select distinct b.brokerid, b.name as brokername, dss.classname as feedname from downloadersymbolsettings dss inner join brokersymbollist bsl on dss.symbolid = bsl.symbolid inner join broker b on bsl.brokerid = b.brokerid where dss.enabled = 1) a ) select last.id, last.datafeedname, last.eventtimestamp, last.status, last.errordescription, last.serveraddress, last.username, note.notified, note.eventtimestamp, broker.brokername from last_feed_entry last inner join earliest_entry_after_ok after_ok on last.datafeedname = after_ok.datafeedname inner join broker on last.datafeedname = broker.feedname left outer join ok_entries ok on ok.datafeedname = last.datafeedname left outer join notified_entries note on note.datafeedname = last.datafeedname and note.r = 1 where (ok.r is null or ok.r = 1) and last.datafeedname not in ( select distinct datafeedname from last_feed_entry where status = 'OK') and extract(epoch from (last.eventtimestamp - after_ok.eventtimestamp)) > 60 * 60 and last.eventtimestamp > current_timestamp - interval '1 day' and (note.eventtimestamp is null or note.eventtimestamp < current_timestamp - interval '10 hours') and last.eventtimestamp > current_timestamp - interval '1 hour' and broker.r = 1;Times Reported Time consuming bind #18
Day Hour Count Duration Avg duration 01 6 24ms 4ms -
with rankedmt4 as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors ), last_feed_entry as ( select * from rankedmt4 where r = 1 ), ok_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where status = 'OK' ), earliest_entry_after_ok as ( select m.datafeedname, min(m.eventtimestamp) as eventtimestamp from mt4datafeederrors m left outer join ( select datafeedname, eventtimestamp from ok_entries where r = 1) oo on m.datafeedname = oo.datafeedname where m.eventtimestamp > coalesce(oo.eventtimestamp, '1900-01-01'::timestamp without time zone) group by m.datafeedname ), notified_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where notified is not null and notified <> '' ), broker as ( select *, row_number() over (partition by feedname order by brokerid) r from ( select distinct b.brokerid, b.name as brokername, dss.classname as feedname from downloadersymbolsettings dss inner join brokersymbollist bsl on dss.symbolid = bsl.symbolid inner join broker b on bsl.brokerid = b.brokerid where dss.enabled = 1) a ) select last.id, last.datafeedname, last.eventtimestamp, last.status, last.errordescription, last.serveraddress, last.username, note.notified, note.eventtimestamp, broker.brokername from last_feed_entry last inner join earliest_entry_after_ok after_ok on last.datafeedname = after_ok.datafeedname inner join broker on last.datafeedname = broker.feedname left outer join ok_entries ok on ok.datafeedname = last.datafeedname left outer join notified_entries note on note.datafeedname = last.datafeedname and note.r = 1 where (ok.r is null or ok.r = 1) and last.datafeedname not in ( select distinct datafeedname from last_feed_entry where status = 'OK') and extract(epoch from (last.eventtimestamp - after_ok.eventtimestamp)) > 60 * 60 and last.eventtimestamp > current_timestamp - interval '1 day' and (note.eventtimestamp is null or note.eventtimestamp < current_timestamp - interval '10 hours') and last.eventtimestamp > current_timestamp - interval '1 hour' and broker.r = 1;
Date: 2026-01-12 01:00:02 Duration: 5ms Database: postgres
-
with rankedmt4 as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors ), last_feed_entry as ( select * from rankedmt4 where r = 1 ), ok_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where status = 'OK' ), earliest_entry_after_ok as ( select m.datafeedname, min(m.eventtimestamp) as eventtimestamp from mt4datafeederrors m left outer join ( select datafeedname, eventtimestamp from ok_entries where r = 1) oo on m.datafeedname = oo.datafeedname where m.eventtimestamp > coalesce(oo.eventtimestamp, '1900-01-01'::timestamp without time zone) group by m.datafeedname ), notified_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where notified is not null and notified <> '' ), broker as ( select *, row_number() over (partition by feedname order by brokerid) r from ( select distinct b.brokerid, b.name as brokername, dss.classname as feedname from downloadersymbolsettings dss inner join brokersymbollist bsl on dss.symbolid = bsl.symbolid inner join broker b on bsl.brokerid = b.brokerid where dss.enabled = 1) a ) select last.id, last.datafeedname, last.eventtimestamp, last.status, last.errordescription, last.serveraddress, last.username, note.notified, note.eventtimestamp, broker.brokername from last_feed_entry last inner join earliest_entry_after_ok after_ok on last.datafeedname = after_ok.datafeedname inner join broker on last.datafeedname = broker.feedname left outer join ok_entries ok on ok.datafeedname = last.datafeedname left outer join notified_entries note on note.datafeedname = last.datafeedname and note.r = 1 where (ok.r is null or ok.r = 1) and last.datafeedname not in ( select distinct datafeedname from last_feed_entry where status = 'OK') and extract(epoch from (last.eventtimestamp - after_ok.eventtimestamp)) > 60 * 60 and last.eventtimestamp > current_timestamp - interval '1 day' and (note.eventtimestamp is null or note.eventtimestamp < current_timestamp - interval '10 hours') and last.eventtimestamp > current_timestamp - interval '1 hour' and broker.r = 1;
Date: 2026-01-12 01:10:02 Duration: 4ms Database: postgres
-
with rankedmt4 as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors ), last_feed_entry as ( select * from rankedmt4 where r = 1 ), ok_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where status = 'OK' ), earliest_entry_after_ok as ( select m.datafeedname, min(m.eventtimestamp) as eventtimestamp from mt4datafeederrors m left outer join ( select datafeedname, eventtimestamp from ok_entries where r = 1) oo on m.datafeedname = oo.datafeedname where m.eventtimestamp > coalesce(oo.eventtimestamp, '1900-01-01'::timestamp without time zone) group by m.datafeedname ), notified_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where notified is not null and notified <> '' ), broker as ( select *, row_number() over (partition by feedname order by brokerid) r from ( select distinct b.brokerid, b.name as brokername, dss.classname as feedname from downloadersymbolsettings dss inner join brokersymbollist bsl on dss.symbolid = bsl.symbolid inner join broker b on bsl.brokerid = b.brokerid where dss.enabled = 1) a ) select last.id, last.datafeedname, last.eventtimestamp, last.status, last.errordescription, last.serveraddress, last.username, note.notified, note.eventtimestamp, broker.brokername from last_feed_entry last inner join earliest_entry_after_ok after_ok on last.datafeedname = after_ok.datafeedname inner join broker on last.datafeedname = broker.feedname left outer join ok_entries ok on ok.datafeedname = last.datafeedname left outer join notified_entries note on note.datafeedname = last.datafeedname and note.r = 1 where (ok.r is null or ok.r = 1) and last.datafeedname not in ( select distinct datafeedname from last_feed_entry where status = 'OK') and extract(epoch from (last.eventtimestamp - after_ok.eventtimestamp)) > 60 * 60 and last.eventtimestamp > current_timestamp - interval '1 day' and (note.eventtimestamp is null or note.eventtimestamp < current_timestamp - interval '10 hours') and last.eventtimestamp > current_timestamp - interval '1 hour' and broker.r = 1;
Date: 2026-01-12 01:20:02 Duration: 3ms Database: postgres
19 21ms 4,488 0ms 3ms 0ms SET application_name = 'PostgreSQL JDBC Driver';Times Reported Time consuming bind #19
Day Hour Count Duration Avg duration 01 4,488 21ms 0ms -
SET application_name = 'PostgreSQL JDBC Driver';
Date: 2026-01-12 01:55:15 Duration: 3ms Database: postgres
-
SET application_name = 'PostgreSQL JDBC Driver';
Date: 2026-01-12 01:00:30 Duration: 0ms Database: postgres
-
SET application_name = 'PostgreSQL JDBC Driver';
Date: 2026-01-12 01:16:34 Duration: 0ms Database: postgres
20 19ms 4,506 0ms 0ms 0ms SET extra_float_digits = 3;Times Reported Time consuming bind #20
Day Hour Count Duration Avg duration 01 4,506 19ms 0ms -
SET extra_float_digits = 3;
Date: 2026-01-12 01:14:33 Duration: 0ms Database: postgres
-
SET extra_float_digits = 3;
Date: 2026-01-12 01:36:40 Duration: 0ms Database: postgres
-
SET extra_float_digits = 3;
Date: 2026-01-12 01:23:06 Duration: 0ms Database: postgres
-
Events
Log levels
Key values
- 482,517 Log entries
Events distribution
Key values
- 0 PANIC entries
- 0 FATAL entries
- 0 ERROR entries
- 0 WARNING entries
Errors per 5 minutes
NO DATASET
Most Frequent Errors/Events
Key values
- 0 Max number of times the same event was reported
- 0 Total events found
Rank Times reported Error NO DATASET