-
Global information
- Generated on Tue Feb 10 04:59:50 2026
- Log file: /home/postgres/pg_data/data/pg_log/postgresql-2026-02-10_060000.log
- Parsed 2,209,396 log entries in 48s
- Log start from 2026-02-10 06:00:00 to 2026-02-10 06:59:48
-
Overview
Global Stats
- 1,016 Number of unique normalized queries
- 232,040 Number of queries
- 3h45m56s Total query duration
- 2026-02-10 06:00:00 First query
- 2026-02-10 06:59:48 Last query
- 5,268 queries/s at 2026-02-10 06:05:02 Query peak
- 3h45m56s Total query duration
- 50s851ms Prepare/parse total duration
- 1m8s Bind total duration
- 3h43m57s Execute total duration
- 8 Number of events
- 2 Number of unique normalized events
- 5 Max number of times the same event was reported
- 0 Number of cancellation
- 38 Total number of automatic vacuums
- 55 Total number of automatic analyzes
- 774 Number temporary file
- 164.75 MiB Max size of temporary file
- 7.16 MiB Average size of temporary file
- 3,390 Total number of sessions
- 12 sessions at 2026-02-10 06:55:24 Session peak
- 3d4h40m39s Total duration of sessions
- 1m21s Average duration of sessions
- 68 Average queries per session
- 3s999ms Average queries duration per session
- 1m17s Average idle time per session
- 3,389 Total number of connections
- 27 connections/s at 2026-02-10 06:33:48 Connection peak
- 6 Total number of databases
SQL Traffic
Key values
- 5,268 queries/s Query Peak
- 2026-02-10 06:05:02 Date
SELECT Traffic
Key values
- 2,616 queries/s Query Peak
- 2026-02-10 06:05:02 Date
INSERT/UPDATE/DELETE Traffic
Key values
- 171 queries/s Query Peak
- 2026-02-10 06:00:54 Date
Queries duration
Key values
- 3h45m56s Total query duration
Prepared queries ratio
Key values
- 0.00 Ratio of bind vs prepare
- 0.00 % Ratio between prepared and "usual" statements
General Activity
↑ Back to the top of the General Activity tableDay Hour Count Min duration Max duration Avg duration Latency Percentile(90) Latency Percentile(95) Latency Percentile(99) Feb 10 06 232,040 0ms 14m42s 57ms 6m22s 7m40s 25m26s Day Hour SELECT COPY TO Average Duration Latency Percentile(90) Latency Percentile(95) Latency Percentile(99) Feb 10 06 73,010 601 0ms 0ms 0ms 0ms Day Hour INSERT UPDATE DELETE COPY FROM Average Duration Latency Percentile(90) Latency Percentile(95) Latency Percentile(99) Feb 10 06 25,851 1,685 13 78 0ms 0ms 0ms 0ms Day Hour Prepare Bind Bind/Prepare Percentage of prepare Feb 10 06 25,807 82,237 3.19 23.47% Day Hour Count Average / Second Feb 10 06 3,389 0.94/s Day Hour Count Average Duration Average idle time Feb 10 06 3,390 1m21s 1m17s -
Connections
Established Connections
Key values
- 27 connections Connection Peak
- 2026-02-10 06:33:48 Date
Connections per database
Key values
- acaweb_fx Main Database
- 3,389 connections Total
Connections per user
Key values
- postgres Main User
- 3,389 connections Total
Connections per host
Key values
- 192.168.4.142 Main host with 1121 connections
- 3,389 Total connections
Host Count 127.0.0.1 114 192.168.0.114 3 192.168.0.216 103 192.168.0.74 318 192.168.1.145 282 192.168.1.15 315 192.168.1.20 315 192.168.1.231 20 192.168.1.239 13 192.168.1.90 41 192.168.2.126 38 192.168.2.182 12 192.168.3.199 36 192.168.4.142 1,121 192.168.4.150 10 192.168.4.177 7 192.168.4.192 4 192.168.4.233 1 192.168.4.238 12 192.168.4.33 86 192.168.4.98 330 [local] 208 -
Sessions
Simultaneous sessions
Key values
- 12 sessions Session Peak
- 2026-02-10 06:55:24 Date
Histogram of session times
Key values
- 2,537 0-500ms duration
Sessions per database
Key values
- acaweb_fx Main Database
- 3,390 sessions Total
Sessions per user
Key values
- postgres Main User
- 3,390 sessions Total
Sessions per host
Key values
- 192.168.4.142 Main Host
- 3,390 sessions Total
Host Count Total Duration Average Duration 127.0.0.1 114 37m26s 19s708ms 192.168.0.114 3 15m56s 5m18s 192.168.0.216 103 1m12s 706ms 192.168.0.74 318 8h8m 1m32s 192.168.1.145 282 4h15m46s 54s421ms 192.168.1.15 316 11h30m37s 2m11s 192.168.1.20 315 21h21m7s 4m4s 192.168.1.231 20 9h51m59s 29m35s 192.168.1.239 13 106ms 8ms 192.168.1.90 41 1m56s 2s841ms 192.168.2.126 38 6s40ms 158ms 192.168.2.182 12 1s1ms 83ms 192.168.3.199 36 1s358ms 37ms 192.168.4.142 1,121 8m29s 454ms 192.168.4.150 10 20h1m13s 2h7s 192.168.4.177 7 42s885ms 6s126ms 192.168.4.192 4 44ms 11ms 192.168.4.233 1 280ms 280ms 192.168.4.238 12 17s336ms 1s444ms 192.168.4.33 86 1m37s 1s134ms 192.168.4.98 330 37s26ms 112ms [local] 208 23m27s 6s766ms -
Checkpoints / Restartpoints
Checkpoints Buffers
Key values
- 10,289 buffers Checkpoint Peak
- 2026-02-10 06:05:38 Date
- 209.944 seconds Highest write time
- 0.067 seconds Sync time
Checkpoints Wal files
Key values
- 4 files Wal files usage Peak
- 2026-02-10 06:05:38 Date
Checkpoints distance
Key values
- 123.19 Mo Distance Peak
- 2026-02-10 06:05:38 Date
Checkpoints Activity
↑ Back to the top of the Checkpoint Activity tableDay Hour Written buffers Write time Sync time Total time Feb 10 06 40,963 2,216.146s 0.185s 2,216.725s Day Hour Added Removed Recycled Synced files Longest sync Average sync Feb 10 06 0 0 22 1,971 0.051s 0s Day Hour Count Avg time (sec) Feb 10 06 0 0s Day Hour Mean distance Mean estimate Feb 10 06 29,514.42 kB 48,956.42 kB -
Temporary Files
Size of temporary files
Key values
- 184.24 MiB Temp Files size Peak
- 2026-02-10 06:10:07 Date
Number of temporary files
Key values
- 30 per second Temp Files Peak
- 2026-02-10 06:17:30 Date
Temporary Files Activity
↑ Back to the top of the Temporary Files Activity tableDay Hour Count Total size Average size Feb 10 06 774 5.41 GiB 7.16 MiB Queries generating the most temporary files (N)
Rank Count Total size Min size Max size Avg size Query 1 62 278.55 MiB 4.36 MiB 4.60 MiB 4.49 MiB select resultuid from relevance_fibonacci_results order by resultuid desc limit ?), fr as ( select a.*, rr.age, rr.relevant from fibonacci_results a left outer join relevance_fibonacci_results rr on a.resultuid = rr.resultuid where case when false = ? then true else a.resultuid > ( select min(resultuid) from relevance_fibonacci_results) end), all_results as ( select fr.resultuid as resultuid, fr.direction as direction, s.exchange as exchange, s.symbolid as symbolid, coalesce(bim.code, s.symbol) as symbol_code, s.longname as symbol_name, s.timegranularity as interval, fr.pattern as pattern_name, fr.timed as timed, fr.patternendtime as identified, dtt.timezone as timezone, fr.patternlengthbars as length, g.basegroupname, newlevels.filtered, case when fr.age is not null then fr.age when fr.resultuid <= rm.resultuid then ? else ? end as age, case when fr.relevant is not null then fr.relevant when fr.resultuid <= rm.resultuid then ? else ? end as relevant, cps.pip from fr inner join brokersymbollist bsl on bsl.brokerid = ? and bsl.symbolid = fr.symbolid inner join symbols s on fr.symbolid = s.symbolid and s.nonliquid = ? inner join symbolgroup sg on fr.symbolid = sg.symbolid inner join groups g on sg.groupid = g.groupid inner join brokergroups bg on g.groupid = bg.groupid and bsl.brokerid = bg.brokerid inner join downloadersymbolsettings dss on fr.symbolid = dss.symbolid inner join datafeedstimetable dtt on dss.classname = dtt.classname and dtt.dayofweek = ? inner join rar_max rm on ? = ? left join lateral calc_fib_signal_filter (fr.resultuid) newlevels on true left join currencypips cps on cps.symbol = s.symbol left outer join brokerinstrumentmap bim on dss.datafeedinstrumentid = bim.datafeedinstrumentid and bim.brokerid = bsl.brokerid and bim.type = ? where fr.gmttimefound > now() - interval ? and dss.enabled = ? and s.deleted = ? and (fr.simulation = ? or fr.simulation is null) and (? = ? or s.timegranularity in (...)) and (? = ? or s.exchange in (...)) and (? = ? or coalesce(bim.code, s.symbol) in (...)) and (? = ? or fr.pattern in (...)) and (? = ? or fr.patternlengthbars <= ?) and (? = ? or (? = ? and fr.timed > cast(? as timestamp)) or (? = ? and fr.timed < cast(? as timestamp)))), results as ( select distinct on (symbolid) * from all_results where (false = ? or relevant = ?) and (? = ? or age <= ?) order by symbolid, resultuid ) select * from results order by identified desc, length desc;-
SELECT resultuid FROM relevance_fibonacci_results ORDER BY resultuid DESC LIMIT 1), fr AS ( SELECT a.*, rr.age, rr.relevant from fibonacci_results a LEFT OUTER JOIN relevance_fibonacci_results rr on a.resultuid = rr.resultuid WHERE CASE WHEN FALSE = $1 THEN true ELSE a.resultuid > ( select min(resultuid) from relevance_fibonacci_results) END), all_results AS ( SELECT fr.resultuid AS resultuid, fr.direction AS direction, s.exchange AS exchange, s.symbolid AS symbolid, coalesce(bim.code, s.symbol) AS symbol_code, s.longname AS symbol_name, s.timegranularity AS interval, fr.pattern AS pattern_name, fr.timed AS timed, fr.patternendtime AS identified, dtt.timezone AS timezone, fr.patternlengthbars AS length, g.basegroupname, newLevels.filtered, CASE WHEN fr.age IS NOT NULL THEN fr.age WHEN fr.resultuid <= rm.resultuid THEN 11 ELSE 0 END as age, CASE WHEN fr.relevant IS NOT NULL THEN fr.relevant WHEN fr.resultuid <= rm.resultuid THEN 0 ELSE 1 END as relevant, cps.pip FROM fr INNER JOIN brokersymbollist bsl ON bsl.brokerid = $2 AND bsl.symbolid = fr.symbolid INNER JOIN symbols s ON fr.symbolid = s.symbolid AND s.nonliquid = 0 INNER JOIN symbolgroup sg on fr.symbolid = sg.symbolid INNER JOIN groups g ON sg.groupid = g.groupid INNER JOIN brokergroups bg on g.groupid = bg.groupid AND bsl.brokerid = bg.brokerid INNER JOIN downloadersymbolsettings dss ON fr.symbolid = dss.symbolid INNER JOIN datafeedstimetable dtt ON dss.classname = dtt.classname AND dtt.dayofweek = 3 INNER JOIN rar_max rm ON 1 = 1 LEFT JOIN LATERAL calc_fib_signal_filter (fr.resultuid) newLevels on true LEFT JOIN currencypips cps on cps.symbol = s.symbol LEFT OUTER JOIN brokerinstrumentmap bim ON dss.datafeedinstrumentid = bim.datafeedinstrumentid AND bim.brokerid = bsl.brokerid AND bim.TYPE = 'OUTBOUND' WHERE fr.gmttimefound > now() - INTERVAL '7 DAYS' AND dss.enabled = 1 AND s.deleted = 0 AND (fr.simulation = 0 OR fr.simulation IS NULL) AND ($3 = 0 OR s.timegranularity in ($4, $5, $6, $7, $8, $9, $10)) AND ($11 = 0 OR s.exchange in ($12)) AND ($13 = 0 OR coalesce(bim.code, s.symbol) in ($14, $15, $16, $17, $18, $19, $20, $21, $22, $23, $24, $25, $26, $27, $28, $29, $30, $31, $32, $33, $34, $35, $36, $37, $38, $39, $40, $41, $42, $43, $44, $45, $46, $47, $48, $49, $50, $51, $52, $53, $54, $55, $56, $57, $58, $59, $60, $61, $62, $63, $64, $65, $66, $67, $68, $69, $70, $71, $72, $73, $74, $75, $76, $77, $78, $79, $80, $81, $82, $83, $84, $85, $86, $87, $88, $89, $90, $91, $92, $93, $94, $95, $96, $97, $98, $99, $100, $101, $102, $103, $104, $105, $106, $107, $108, $109, $110, $111, $112, $113, $114, $115, $116, $117, $118, $119, $120, $121, $122, $123, $124, $125, $126, $127, $128, $129, $130, $131, $132, $133, $134, $135, $136, $137, $138, $139, $140, $141, $142, $143, $144, $145, $146, $147, $148, $149, $150, $151, $152, $153, $154, $155, $156, $157, $158, $159, $160, $161, $162, $163, $164, $165, $166, $167, $168, $169, $170, $171, $172, $173, $174, $175, $176, $177, $178, $179, $180, $181, $182, $183, $184, $185, $186, $187, $188, $189, $190, $191, $192, $193, $194, $195, $196, $197, $198, $199, $200, $201, $202, $203, $204, $205, $206, $207, $208, $209, $210, $211, $212, $213, $214, $215, $216, $217, $218, $219, $220, $221, $222, $223, $224, $225, $226, $227, $228, $229, $230, $231, $232, $233, $234, $235, $236, $237, $238, $239, $240, $241, $242, $243, $244, $245, $246, $247, $248, $249, $250, $251, $252, $253, $254, $255, $256, $257, $258, $259, $260, $261, $262, $263, $264, $265, $266, $267, $268, $269, $270, $271, $272, $273, $274, $275, $276, $277, $278, $279, $280, $281, $282, $283, $284, $285, $286, $287, $288, $289, $290, $291, $292, $293, $294, $295, $296, $297, $298, $299, $300, $301, $302, $303, $304, $305, $306, $307, $308, $309, $310, $311, $312, $313, $314, $315, $316, $317, $318, $319, $320, $321, $322, $323)) AND ($324 = 0 OR fr.pattern in ($325)) AND ($326 = 0 OR fr.patternlengthbars <= $327) AND ($328 = 0 OR ($329 = 1 AND fr.timed > cast('1970-01-01' as timestamp)) OR ($330 = 2 AND fr.timed < cast('1970-01-01' as timestamp)))), results AS ( SELECT DISTINCT ON (symbolid) * FROM all_results WHERE (FALSE = $331 OR relevant = 1) AND ($332 = 0 OR age <= $333) ORDER BY symbolid, resultuid ) SELECT * from results ORDER BY identified DESC, length DESC;
Date: 2026-02-10 06:00:46 Duration: 0ms
2 29 1.66 GiB 5.07 MiB 164.75 MiB 58.55 MiB with rankedmt4 as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors ), last_feed_entry as ( select * from rankedmt4 where r = ? ), ok_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where status = ? ), earliest_entry_after_ok as ( select m.datafeedname, min(m.eventtimestamp) as eventtimestamp from mt4datafeederrors m left outer join ( select datafeedname, eventtimestamp from ok_entries where r = ?) oo on m.datafeedname = oo.datafeedname where m.eventtimestamp > coalesce(oo.eventtimestamp, ?::timestamp without time zone) group by m.datafeedname ), notified_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where notified is not null and notified <> ? ), broker as ( select *, row_number() over (partition by feedname order by brokerid) r from ( select distinct b.brokerid, b.name as brokername, dss.classname as feedname from downloadersymbolsettings dss inner join brokersymbollist bsl on dss.symbolid = bsl.symbolid inner join broker b on bsl.brokerid = b.brokerid where dss.enabled = ?) a ) select last.id, last.datafeedname, last.eventtimestamp, last.status, last.errordescription, last.serveraddress, last.username, note.notified, note.eventtimestamp, broker.brokername from last_feed_entry last inner join earliest_entry_after_ok after_ok on last.datafeedname = after_ok.datafeedname inner join broker on last.datafeedname = broker.feedname left outer join ok_entries ok on ok.datafeedname = last.datafeedname left outer join notified_entries note on note.datafeedname = last.datafeedname and note.r = ? where (ok.r is null or ok.r = ?) and last.datafeedname not in ( select distinct datafeedname from last_feed_entry where status = ?) and extract(epoch from (last.eventtimestamp - after_ok.eventtimestamp)) > ? * ? and last.eventtimestamp > current_timestamp - interval ? and (note.eventtimestamp is null or note.eventtimestamp < current_timestamp - interval ?) and last.eventtimestamp > current_timestamp - interval ? and broker.r = ?;-
with rankedmt4 as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors ), last_feed_entry as ( select * from rankedmt4 where r = 1 ), ok_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where status = 'OK' ), earliest_entry_after_ok as ( select m.datafeedname, min(m.eventtimestamp) as eventtimestamp from mt4datafeederrors m left outer join ( select datafeedname, eventtimestamp from ok_entries where r = 1) oo on m.datafeedname = oo.datafeedname where m.eventtimestamp > coalesce(oo.eventtimestamp, '1900-01-01'::timestamp without time zone) group by m.datafeedname ), notified_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where notified is not null and notified <> '' ), broker as ( select *, row_number() over (partition by feedname order by brokerid) r from ( select distinct b.brokerid, b.name as brokername, dss.classname as feedname from downloadersymbolsettings dss inner join brokersymbollist bsl on dss.symbolid = bsl.symbolid inner join broker b on bsl.brokerid = b.brokerid where dss.enabled = 1) a ) select last.id, last.datafeedname, last.eventtimestamp, last.status, last.errordescription, last.serveraddress, last.username, note.notified, note.eventtimestamp, broker.brokername from last_feed_entry last inner join earliest_entry_after_ok after_ok on last.datafeedname = after_ok.datafeedname inner join broker on last.datafeedname = broker.feedname left outer join ok_entries ok on ok.datafeedname = last.datafeedname left outer join notified_entries note on note.datafeedname = last.datafeedname and note.r = 1 where (ok.r is null or ok.r = 1) and last.datafeedname not in ( select distinct datafeedname from last_feed_entry where status = 'OK') and extract(epoch from (last.eventtimestamp - after_ok.eventtimestamp)) > 60 * 60 and last.eventtimestamp > current_timestamp - interval '1 day' and (note.eventtimestamp is null or note.eventtimestamp < current_timestamp - interval '10 hours') and last.eventtimestamp > current_timestamp - interval '1 hour' and broker.r = 1;
Date: 2026-02-10 06:00:06 Duration: 0ms
3 28 84.90 MiB 3.00 MiB 3.15 MiB 3.03 MiB select resultuid from relevance_consecutivecandles_results order by resultuid desc limit ?), all_results as ( select ccr.resultuid as resultuid, ccr.direction as direction, s.exchange as exchange, s.symbolid as symbolid, coalesce(bim.code, s.symbol) as symbol_code, s.longname as symbol_name, s.timegranularity as interval, ccr.patternendtime as identified, dtt.timezone as timezone, ccr.qtyconsecutivecandles as length, g.basegroupname, case when rcr.age is not null then rcr.age when ccr.resultuid <= rm.resultuid then ? else ? end as age, case when rcr.relevant is not null then rcr.relevant when ccr.resultuid <= rm.resultuid then ? else ? end as relevant, cps.pip, newlevels.filtered from consecutivecandles_results ccr inner join brokersymbollist bsl on bsl.brokerid = ? and bsl.symbolid = ccr.symbolid inner join symbols s on ccr.symbolid = s.symbolid and s.nonliquid = ? inner join downloadersymbolsettings dss on ccr.symbolid = dss.symbolid inner join datafeedstimetable dtt on dss.classname = dtt.classname and dtt.dayofweek = ? inner join symbolgroup sg on ccr.symbolid = sg.symbolid inner join groups g on sg.groupid = g.groupid inner join brokergroups bg on g.groupid = bg.groupid and bsl.brokerid = bg.brokerid inner join rar_max rm on ? = ? left outer join relevance_consecutivecandles_results rcr on rcr.resultuid = ccr.resultuid left join currencypips cps on cps.symbol = s.symbol left outer join brokerinstrumentmap bim on dss.datafeedinstrumentid = bim.datafeedinstrumentid and bim.brokerid = bsl.brokerid and bim.type = ? left join lateral calc_cc_signal_filter (ccr.resultuid) newlevels on true where ccr.gmttimefound > now() - interval ? and s.deleted = ? and (ccr.simulation = ? or ccr.simulation is null) and (? = ? or s.timegranularity in (...)) and (? = ? or s.exchange in (...)) and (? = ? or coalesce(bim.code, s.symbol) in (...)) and (? = ? or ccr.patternlengthbars <= ?)), results as ( select distinct on (symbolid) * from all_results where (false = ? or relevant = ?) and (? = ? or age <= ?) order by symbolid, resultuid ) select * from results order by identified desc, length desc;-
SELECT resultuid FROM relevance_consecutivecandles_results ORDER BY resultuid DESC LIMIT 1), all_results AS ( SELECT ccr.resultuid AS resultuid, ccr.direction AS direction, s.exchange AS exchange, s.symbolid AS symbolid, coalesce(bim.code, s.symbol) AS symbol_code, s.longname AS symbol_name, s.timegranularity AS interval, ccr.patternendtime AS identified, dtt.timezone AS timezone, ccr.qtyconsecutivecandles AS length, g.basegroupname, CASE WHEN rcr.age IS NOT NULL THEN rcr.age WHEN ccr.resultuid <= rm.resultuid THEN 1 ELSE 0 END as age, CASE WHEN rcr.relevant IS NOT NULL THEN rcr.relevant WHEN ccr.resultuid <= rm.resultuid THEN 0 ELSE 1 END as relevant, cps.pip, newLevels.filtered FROM consecutivecandles_results ccr INNER JOIN brokersymbollist bsl ON bsl.brokerid = $1 AND bsl.symbolid = ccr.symbolid INNER JOIN symbols s ON ccr.symbolid = s.symbolid AND s.nonliquid = 0 INNER JOIN downloadersymbolsettings dss ON ccr.symbolid = dss.symbolid INNER JOIN datafeedstimetable dtt ON dss.classname = dtt.classname AND dtt.dayofweek = 3 INNER JOIN symbolgroup sg on ccr.symbolid = sg.symbolid INNER JOIN groups g ON sg.groupid = g.groupid INNER JOIN brokergroups bg on g.groupid = bg.groupid AND bsl.brokerid = bg.brokerid INNER JOIN rar_max rm ON 1 = 1 LEFT OUTER JOIN relevance_consecutivecandles_results rcr ON rcr.resultuid = ccr.resultuid LEFT JOIN currencypips cps on cps.symbol = s.symbol LEFT OUTER JOIN brokerinstrumentmap bim ON dss.datafeedinstrumentid = bim.datafeedinstrumentid AND bim.brokerid = bsl.brokerid AND bim.TYPE = 'OUTBOUND' LEFT JOIN LATERAL calc_cc_signal_filter (ccr.resultuid) newLevels on true WHERE ccr.gmttimefound > now() - INTERVAL '7 DAYS' AND s.deleted = 0 AND (ccr.simulation = 0 OR ccr.simulation IS NULL) AND ($2 = 0 OR s.timegranularity in ($3, $4, $5, $6, $7, $8, $9)) AND ($10 = 0 OR s.exchange in ($11)) AND ($12 = 0 OR coalesce(bim.code, s.symbol) in ($13, $14, $15, $16, $17, $18, $19, $20, $21, $22, $23, $24, $25, $26, $27, $28, $29, $30, $31, $32, $33, $34, $35, $36, $37, $38, $39, $40, $41, $42, $43, $44, $45, $46, $47, $48, $49, $50, $51, $52, $53, $54, $55, $56, $57, $58, $59, $60, $61, $62, $63, $64, $65, $66, $67, $68, $69, $70, $71, $72, $73, $74, $75, $76, $77, $78, $79, $80, $81, $82, $83, $84, $85, $86, $87, $88, $89, $90, $91, $92, $93, $94, $95, $96, $97, $98, $99, $100, $101, $102, $103, $104, $105, $106, $107, $108, $109, $110, $111, $112, $113, $114, $115, $116, $117, $118, $119, $120, $121, $122, $123, $124, $125, $126, $127, $128, $129, $130, $131, $132, $133, $134, $135, $136, $137, $138, $139, $140, $141, $142, $143, $144, $145, $146, $147, $148, $149, $150, $151, $152, $153, $154, $155, $156, $157, $158, $159, $160, $161, $162, $163, $164, $165, $166, $167, $168, $169, $170, $171, $172, $173, $174, $175, $176, $177, $178, $179, $180, $181, $182, $183, $184, $185, $186, $187, $188, $189, $190, $191, $192, $193, $194, $195, $196, $197, $198, $199, $200, $201, $202, $203, $204, $205, $206, $207, $208, $209, $210, $211, $212, $213, $214, $215, $216, $217, $218, $219, $220, $221, $222, $223, $224, $225, $226, $227, $228, $229, $230, $231, $232, $233, $234, $235, $236, $237, $238, $239, $240, $241, $242, $243, $244, $245, $246, $247, $248, $249, $250, $251, $252, $253, $254, $255, $256, $257, $258, $259, $260, $261, $262, $263, $264, $265, $266, $267, $268, $269, $270, $271, $272, $273, $274, $275, $276, $277, $278, $279, $280, $281, $282, $283, $284, $285, $286, $287, $288, $289, $290, $291, $292, $293, $294, $295, $296, $297, $298, $299, $300, $301, $302, $303, $304, $305, $306, $307, $308, $309, $310, $311, $312, $313, $314, $315, $316, $317, $318, $319, $320, $321, $322)) AND ($323 = 0 OR ccr.patternlengthbars <= $324)), results AS ( SELECT DISTINCT ON (symbolid) * FROM all_results WHERE (FALSE = $325 OR relevant = 1) AND ($326 = 0 OR age <= $327) ORDER BY symbolid, resultuid ) SELECT * from results ORDER BY identified DESC, length DESC;
Date: 2026-02-10 06:00:24 Duration: 0ms
4 20 61.58 MiB 3.08 MiB 3.08 MiB 3.08 MiB select resultuid from relevance_autochartist_results order by resultuid desc limit ?), ar as ( select a.*, rr.age, rr.relevant from autochartist_results a left outer join relevance_autochartist_results rr on a.resultuid = rr.resultuid where case when false = ? then true else a.resultuid > ( select min(resultuid) from relevance_autochartist_results) end), all_results as ( select ar.resultuid as resultuid, ar.direction as direction, ar.predictiontimeto as predictiontimeto, ar.predictionpricefrom as predictionpricefrom, ar.predictionpriceto as predictionpriceto, cp.pip as pip, s.exchange as exchange, s.symbolid as symbolid, coalesce(bim.code, s.symbol) as symbol_code, s.longname as symbol_name, s.timegranularity as interval, ar.pattern as pattern_name, ar.breakout as breakout, ar.patternendtime as identified, dtt.timezone as timezone, ar.patternlengthbars as length, g.basegroupname, newlevels.profit, newlevels.stop, newlevels.filtered, case when ar.age is not null then ar.age when ar.resultuid <= rm.resultuid then ? else ? end as age, case when ar.relevant is not null then ar.relevant when ar.resultuid <= rm.resultuid then ? else ? end as relevant from ar inner join symbols s on ar.symbolid = s.symbolid and s.nonliquid = ? inner join brokersymbollist bsl on bsl.brokerid = ? and bsl.symbolid = s.symbolid inner join symbolgroup sg on bsl.symbolid = sg.symbolid inner join groups g on sg.groupid = g.groupid inner join brokergroups bg on g.groupid = bg.groupid and bsl.brokerid = bg.brokerid inner join downloadersymbolsettings dss on sg.symbolid = dss.symbolid inner join datafeedstimetable dtt on dss.classname = dtt.classname and dtt.dayofweek = ? inner join rar_max rm on ? = ? left outer join autochartist_symbolupdates au on dss.symbolid = au.symbolid left outer join currencypips cp on s.symbol = cp.symbol left join lateral calc_cp_signal (ar.resultuid) newlevels on true left outer join brokerinstrumentmap bim on dss.datafeedinstrumentid = bim.datafeedinstrumentid and bim.brokerid = bsl.brokerid and bim.type = ? where ar.gmttimefound > now() - interval ? and dss.enabled = ? and s.deleted = ? and (ar.simulation = ? or ar.simulation is null) and (? = ? or s.timegranularity in (...)) and (? = ? or s.exchange in (...)) and (? = ? or coalesce(bim.code, s.symbol) in (...)) and (? = ? or ar.pattern in (...)) and (? = ? or (? = ? and ar.breakout >= ?) or (? = ? and ar.breakout < ?)) and (? = ? or ar.patternlengthbars <= ?) and newlevels.filtered = false and ar.patternstarttime >= coalesce(au.earliestpricedatetime, ?::timestamp without time zone) -- to make sure patternstarttime is in our t-tables ), results as ( select distinct on (symbolid) * from all_results where (false = ? or relevant = ?) and (? = ? or age <= ?) order by symbolid, resultuid ) select * from results order by identified desc, length desc ;-
SELECT resultuid FROM relevance_autochartist_results ORDER BY resultuid DESC LIMIT 1), ar AS ( SELECT a.*, rr.age, rr.relevant from autochartist_results a LEFT OUTER JOIN relevance_autochartist_results rr on a.resultuid = rr.resultuid WHERE CASE WHEN FALSE = $1 THEN true ELSE a.resultuid > ( select min(resultuid) from relevance_autochartist_results) END), all_results AS ( SELECT ar.resultuid AS resultuid, ar.direction AS direction, ar.predictiontimeto AS predictiontimeto, ar.predictionpricefrom AS predictionpricefrom, ar.predictionpriceto AS predictionpriceto, cp.pip AS pip, s.exchange AS exchange, s.symbolid AS symbolid, coalesce(bim.code, s.symbol) AS symbol_code, s.longname AS symbol_name, s.timegranularity AS interval, ar.pattern AS pattern_name, ar.breakout AS breakout, ar.patternendtime AS identified, dtt.timezone AS timezone, ar.patternlengthbars AS length, g.basegroupname, newLevels.profit, newLevels.stop, newLevels.filtered, CASE WHEN ar.age IS NOT NULL THEN ar.age WHEN ar.resultuid <= rm.resultuid THEN 11 ELSE 0 END as age, CASE WHEN ar.relevant IS NOT NULL THEN ar.relevant WHEN ar.resultuid <= rm.resultuid THEN 0 ELSE 1 END as relevant FROM ar INNER JOIN symbols s ON ar.symbolid = s.symbolid AND s.nonliquid = 0 INNER JOIN brokersymbollist bsl ON bsl.brokerid = $2 AND bsl.symbolid = s.symbolid INNER JOIN symbolgroup sg on bsl.symbolid = sg.symbolid INNER JOIN groups g ON sg.groupid = g.groupid INNER JOIN brokergroups bg on g.groupid = bg.groupid AND bsl.brokerid = bg.brokerid INNER JOIN downloadersymbolsettings dss ON sg.symbolid = dss.symbolid INNER JOIN datafeedstimetable dtt ON dss.classname = dtt.classname AND dtt.dayofweek = 3 INNER JOIN rar_max rm ON 1 = 1 LEFT OUTER JOIN autochartist_symbolupdates au on dss.symbolid = au.symbolid LEFT OUTER JOIN currencypips cp ON s.symbol = cp.symbol LEFT JOIN LATERAL calc_cp_signal (ar.resultuid) newLevels on true LEFT OUTER JOIN brokerinstrumentmap bim ON dss.datafeedinstrumentid = bim.datafeedinstrumentid AND bim.brokerid = bsl.brokerid AND bim.TYPE = 'OUTBOUND' WHERE ar.gmttimefound > now() - INTERVAL '7 DAYS' AND dss.enabled = 1 AND s.deleted = 0 AND (ar.simulation = 0 OR ar.simulation IS NULL) AND ($3 = 0 OR s.timegranularity in ($4)) AND ($5 = 0 OR s.exchange in ($6)) AND ($7 = 0 OR coalesce(bim.code, s.symbol) in ($8)) AND ($9 = 0 OR ar.pattern in ($10)) AND ($11 = 0 OR ($12 = 1 AND ar.breakout >= 0) OR ($13 = 2 AND ar.breakout < 0)) AND ($14 = 0 OR ar.patternlengthbars <= $15) and newLevels.filtered = false AND ar.patternstarttime >= coalesce(au.earliestpricedatetime, '1900-01-01'::timestamp without time zone) -- To make sure patternstarttime is in our t-tables ), results AS ( SELECT DISTINCT ON (symbolid) * FROM all_results WHERE (FALSE = $16 OR relevant = 1) AND ($17 = 0 OR age <= $18) ORDER BY symbolid, resultuid ) SELECT * from results ORDER BY identified DESC, length DESC;
Date: 2026-02-10 06:01:54 Duration: 0ms
5 13 503.85 MiB 38.76 MiB 38.76 MiB 38.76 MiB update solr_relevance_old set new_hod_correct = sub.hod_correct, new_hod_percent = sub.hod_percent, new_hod_total = sub.hod_total, new_pattern_correct = sub.pattern_correct, new_pattern_percent = sub.pattern_percent, new_pattern_total = sub.pattern_total, new_percent = sub.percent, new_symbol_correct = sub.symbol_correct, new_symbol_percent = sub.symbol_percent, new_symbol_total = sub.symbol_total from ( select distinct resultuid, hod_correct, hod_percent, hod_total, hod, pattern_correct, pattern_percent, pattern_total, percent, symbol_correct, symbol_percent, symbol_total from whatshot_probability where type = ?) sub where result_uid = sub.resultuid;-
UPDATE solr_relevance_old SET new_hod_correct = sub.hod_correct, new_hod_percent = sub.hod_percent, new_hod_total = sub.hod_total, new_pattern_correct = sub.pattern_correct, new_pattern_percent = sub.pattern_percent, new_pattern_total = sub.pattern_total, new_percent = sub.percent, new_symbol_correct = sub.symbol_correct, new_symbol_percent = sub.symbol_percent, new_symbol_total = sub.symbol_total FROM ( select distinct resultuid, hod_correct, hod_percent, hod_total, hod, pattern_correct, pattern_percent, pattern_total, percent, symbol_correct, symbol_percent, symbol_total FROM whatshot_probability WHERE type = 'cp') sub WHERE result_uid = sub.resultuid;
Date: 2026-02-10 06:01:13 Duration: 0ms
6 13 928.88 MiB 71.15 MiB 71.51 MiB 71.45 MiB with max_ra as ( select resultuid from relevance_keylevels_results order by resultuid desc limit ?) update solr_relevance_old set newrelevant = sub.relevant, newage = sub.age from ( select so.uuid, case when ra.relevant is not null then ra.relevant when so.result_uid < max_ra.resultuid then ? else ? end as relevant, case when ra.age is not null then ra.age when so.result_uid < max_ra.resultuid then ? else ? end as age, so.result_uid from max_ra, solr_relevance_old so inner join keylevels_results k on so.result_uid = k.resultuid and so.uuid ilike ? inner join downloadersymbolsettings dss on k.symbolid = dss.symbolid left outer join relevance_keylevels_results ra on so.result_uid = ra.resultuid and so.uuid ilike ?) sub where solr_relevance_old.result_uid = sub.result_uid and solr_relevance_old.uuid ilike ?; update solr_relevance_old set newrelevant = ? where result_uid in ( select result_uid from solr_relevance_old s left outer join keylevels_results a on a.resultuid = s.result_uid where s.uuid ilike ? and a.resultuid is null); update solr_relevance_old set new_hod_correct = sub.hod_correct, new_hod_percent = sub.hod_percent, new_hod_total = sub.hod_total, new_pattern_correct = sub.pattern_correct, new_pattern_percent = sub.pattern_percent, new_pattern_total = sub.pattern_total, new_percent = sub.percent, new_symbol_correct = sub.symbol_correct, new_symbol_percent = sub.symbol_percent, new_symbol_total = sub.symbol_total from ( select distinct resultuid, hod_correct, hod_percent, hod_total, hod, pattern_correct, pattern_percent, pattern_total, percent, symbol_correct, symbol_percent, symbol_total from whatshot_probability where type in (...)) sub where result_uid = sub.resultuid;-
with max_ra as ( select resultuid from relevance_keylevels_results order by resultuid desc limit 1) update solr_relevance_old set newrelevant = sub.relevant, newage = sub.age from ( select so.uuid, case when ra.relevant is not null then ra.relevant when so.result_uid < max_ra.resultuid then 0 else 1 end as relevant, case when ra.age is not null then ra.age when so.result_uid < max_ra.resultuid then 11 else 0 end as age, so.result_uid from max_ra, solr_relevance_old so inner join keylevels_results k on so.result_uid = k.resultuid and so.uuid ilike 'kl_%' inner join downloadersymbolsettings dss on k.symbolid = dss.symbolid left outer join relevance_keylevels_results ra on so.result_uid = ra.resultuid and so.uuid ilike 'kl_%') sub where solr_relevance_old.result_uid = sub.result_uid and solr_relevance_old.uuid ilike 'kl_%'; update solr_relevance_old set newrelevant = 0 where result_uid in ( select result_uid from solr_relevance_old s left outer join keylevels_results a on a.resultuid = s.result_uid where s.uuid ilike 'kl_%' and a.resultuid is null); UPDATE solr_relevance_old SET new_hod_correct = sub.hod_correct, new_hod_percent = sub.hod_percent, new_hod_total = sub.hod_total, new_pattern_correct = sub.pattern_correct, new_pattern_percent = sub.pattern_percent, new_pattern_total = sub.pattern_total, new_percent = sub.percent, new_symbol_correct = sub.symbol_correct, new_symbol_percent = sub.symbol_percent, new_symbol_total = sub.symbol_total FROM ( select distinct resultuid, hod_correct, hod_percent, hod_total, hod, pattern_correct, pattern_percent, pattern_total, percent, symbol_correct, symbol_percent, symbol_total FROM whatshot_probability WHERE type in ('kl', 'ekl')) sub WHERE result_uid = sub.resultuid;
Date: 2026-02-10 06:01:16 Duration: 0ms
7 11 88.60 MiB 8.05 MiB 8.06 MiB 8.05 MiB jr.resultuid as resultuid, jr.direction as direction, jr.patternendtime as identified, jr.patternlengthbars as length, jr.patternstarttime as patternstarttime, case when jr.trendchangeid = ? then ? else ? end as trendchange, s.exchange as exchange, s.symbolid as symbolid, coalesce(bim.code, s.symbol) as symbol_code, s.longname as symbol_name, s.timegranularity as interval, jp.patternname as pattern_name, dtt.timezone as timezone, ? as age, cps.pip, g.basegroupname from japsticks_results jr inner join brokersymbollist bsl on bsl.brokerid = ? and bsl.symbolid = jr.symbolid inner join relevance_japsticks_results rar on rar.resultuid = jr.resultuid inner join symbols s on jr.symbolid = s.symbolid and s.nonliquid = ? inner join japsticks_patterns jp on jr.patternid = jp.id inner join downloadersymbolsettings dss on jr.symbolid = dss.symbolid inner join datafeedstimetable dtt on dss.classname = dtt.classname and dtt.dayofweek = ? inner join symbolgroup sg on s.symbolid = sg.symbolid inner join groups g on sg.groupid = g.groupid inner join brokergroups bg on g.groupid = bg.groupid and bsl.brokerid = bg.brokerid left join currencypips cps on cps.symbol = s.symbol left outer join brokerinstrumentmap bim on dss.datafeedinstrumentid = bim.datafeedinstrumentid and bim.brokerid = bsl.brokerid and bim.type = ? where jr.gmttimefound > now() - interval ? and s.deleted = ? and (jr.simulation = ? or jr.simulation is null) and (rar.relevant = ?) --and (semicolon_age = ? or rar.age <= semicolon_age) and (? = ? or s.timegranularity in (...)) and (? = ? or s.exchange in (...)) and (? = ? or coalesce(bim.code, s.symbol) in (...)) and (? = ? or jp.patternname in (...)) and (? = ? or jr.patternlengthbars <= ?) ), results as ( select distinct on (symbolid) * from all_results order by symbolid, resultuid ) select * from results order by identified desc, length desc ;-
jr.resultuid AS resultuid, jr.direction AS direction, jr.patternendtime AS identified, jr.patternlengthbars AS length, jr.patternstarttime AS patternstarttime, case when jr.trendchangeid = 1 then 'Continuation' else 'Reversal' end AS trendchange, s.exchange AS exchange, s.symbolid AS symbolid, coalesce(bim.code, s.symbol) AS symbol_code, s.longname AS symbol_name, s.timegranularity AS interval, jp.patternname AS pattern_name, dtt.timezone AS timezone, 0 AS age, cps.pip, g.basegroupname FROM japsticks_results jr INNER JOIN brokersymbollist bsl ON bsl.brokerid = $1 AND bsl.symbolid = jr.symbolid INNER JOIN relevance_japsticks_results rar ON rar.resultuid = jr.resultuid INNER JOIN symbols s ON jr.symbolid = s.symbolid AND s.nonliquid = 0 INNER JOIN japsticks_patterns jp ON jr.patternid = jp.id INNER JOIN downloadersymbolsettings dss ON jr.symbolid = dss.symbolid INNER JOIN datafeedstimetable dtt ON dss.classname = dtt.classname AND dtt.dayofweek = 3 INNER JOIN symbolgroup sg on s.symbolid = sg.symbolid INNER JOIN groups g ON sg.groupid = g.groupid INNER JOIN brokergroups bg on g.groupid = bg.groupid AND bsl.brokerid = bg.brokerid LEFT JOIN currencypips cps on cps.symbol = s.symbol LEFT OUTER JOIN brokerinstrumentmap bim ON dss.datafeedinstrumentid = bim.datafeedinstrumentid AND bim.brokerid = bsl.brokerid AND bim.TYPE = 'OUTBOUND' WHERE jr.gmttimefound > now() - INTERVAL '7 DAYS' AND s.deleted = 0 AND (jr.simulation = 0 OR jr.simulation IS NULL) AND (rar.relevant = 1) --AND (semicolon_age = 0 OR rar.age <= semicolon_age) AND ($2 = 0 OR s.timegranularity in ($3)) AND ($4 = 0 OR s.exchange in ($5)) AND ($6 = 0 OR coalesce(bim.code, s.symbol) in ($7)) AND ($8 = 0 OR jp.patternname in ($9)) AND ($10 = 0 OR jr.patternlengthbars <= $11)), results AS ( SELECT DISTINCT ON (symbolid) * FROM all_results ORDER BY symbolid, resultuid ) SELECT * from results ORDER BY identified DESC, length DESC;
Date: 2026-02-10 06:06:49 Duration: 0ms
8 8 1.06 GiB 135.96 MiB 136.00 MiB 135.98 MiB select updateresultsmaterializedview ();-
select updateresultsmaterializedview ();
Date: 2026-02-10 06:02:16 Duration: 0ms
9 4 341.41 MiB 85.31 MiB 85.41 MiB 85.35 MiB select updateageforrelevantresults ();-
select updateageforrelevantresults ();
Date: 2026-02-10 06:02:05 Duration: 0ms
10 3 9.24 MiB 3.08 MiB 3.08 MiB 3.08 MiB select resultuid from relevance_keylevels_results order by resultuid desc limit ? ) select case when a.old_resultuid = ? then a.old_resultuid else a.resultuid end as ruid, s.symbolid as sid, s.symbol as sym, longname, shortname, exchange as e, timegranularity as tg, a.patternid as pid, a.direction as d, a.patternprice as pp, atbaridentified as pet, case when (x9 != ?) then x9 when (x8 != ?) then x8 when (x7 != ?) then x7 when (x6 != ?) then x6 when (x5 != ?) then x5 when (x4 != ?) then x4 when (x3 != ?) then x3 when (x2 != ?) then x2 end as pst, patternprice as patp, x0, x1, x2, case when (x3 != ?) then x3 else ? end as x3, case when (x4 != ?) then x4 else ? end as x4, case when (x5 != ?) then x5 else ? end as x5, case when (x6 != ?) then x6 else ? end as x6, case when (x7 != ?) then x7 else ? end as x7, case when (x8 != ?) then x8 else ? end as x8, errormargin as erm, breakoutprice as pe, breakoutbars as be, breakout, atbaridentified as atbar, atpriceidentified as atprice, patternlengthbars as l, bandwidth as bw, qtytp as qtp, p.patternname as patternname, dtt.absolutetimezoneoffset as tzos, dtt.timezone as timezone, approachingtimestamp as apt, approachingregion as apr, predictionpricefrom as ppf, predictionpriceto as ppt, predictiontimefrom as ptf, predictiontimebars as ptb, furthestprice as fp, newlevels.filtered, a.uniquepointsvalue as upv, case when rar.age is not null then rar.age when a.resultuid <= rm.resultuid then ? else ? end as age, case when rar.relevant is not null then rar.relevant when a.resultuid <= rm.resultuid then ? else ? end as relevant, cps.pip from keylevels_results a inner join downloadersymbolsettings dss on a.symbolid = dss.symbolid inner join datafeedstimetable dtt on dss.classname = dtt.classname inner join symbols s on a.symbolid = s.symbolid inner join hrspatterns p on a.patternid = p.patternid inner join rar_max rm on ? = ? left outer join relevance_keylevels_results rar on a.resultuid = rar.resultuid left join lateral calc_kl_signal_filter (a.resultuid) newlevels on true left join currencypips cps on cps.symbol = s.symbol where (a.old_resultuid = ? or a.resultuid = ?) and dtt.dayofweek = ?;-
SELECT resultuid FROM relevance_keylevels_results ORDER BY resultuid DESC LIMIT 1 ) SELECT CASE WHEN a.old_resultuid = $1 THEN a.old_resultuid ELSE a.resultuid END AS ruid, s.symbolid AS sid, s.symbol AS sym, longname, shortname, Exchange AS e, timegranularity AS tg, a.PatternID AS pid, a.direction AS d, a.patternprice AS pp, atbaridentified AS pet, CASE WHEN (x9 != '') THEN x9 WHEN (x8 != '') THEN x8 WHEN (x7 != '') THEN x7 WHEN (x6 != '') THEN x6 WHEN (x5 != '') THEN x5 WHEN (x4 != '') THEN x4 WHEN (x3 != '') THEN x3 WHEN (x2 != '') THEN x2 END AS pst, PatternPrice AS patp, x0, x1, x2, CASE WHEN (x3 != '') THEN x3 ELSE '0' END AS x3, CASE WHEN (x4 != '') THEN x4 ELSE '0' END AS x4, CASE WHEN (x5 != '') THEN x5 ELSE '0' END AS x5, CASE WHEN (x6 != '') THEN x6 ELSE '0' END AS x6, CASE WHEN (x7 != '') THEN x7 ELSE '0' END AS x7, CASE WHEN (x8 != '') THEN x8 ELSE '0' END AS x8, errorMargin AS erm, breakoutprice AS pE, breakoutbars AS be, breakout, atbaridentified AS atBar, atpriceidentified AS atPrice, PatternLengthBars AS l, Bandwidth AS bw, QtyTP AS qtp, p.patternname AS patternname, dtt.absolutetimezoneoffset AS tzOs, dtt.timezone AS timezone, approachingtimestamp AS apt, approachingregion AS apr, predictionpricefrom AS ppf, predictionpriceto AS ppt, predictiontimefrom AS ptf, predictiontimebars AS ptb, furthestprice AS fp, newLevels.filtered, a.uniquepointsvalue AS upv, CASE WHEN rar.age IS NOT NULL THEN rar.age WHEN a.resultuid <= rm.resultuid THEN 11 ELSE 0 END as age, CASE WHEN rar.relevant IS NOT NULL THEN rar.relevant WHEN a.resultuid <= rm.resultuid THEN 0 ELSE 1 END as relevant, cps.pip FROM keylevels_results a INNER JOIN downloadersymbolsettings dss ON a.symbolid = dss.symbolid INNER JOIN datafeedstimetable dtt ON dss.classname = dtt.classname INNER JOIN symbols s ON a.symbolid = s.symbolid INNER JOIN hrspatterns p ON a.patternid = p.patternid INNER JOIN rar_max rm ON 1 = 1 LEFT OUTER JOIN relevance_keylevels_results rar ON a.resultuid = rar.resultuid LEFT JOIN LATERAL calc_kl_signal_filter (a.resultuid) newLevels on true LEFT JOIN currencypips cps on cps.symbol = s.symbol WHERE (a.old_resultuid = $2 OR a.resultuid = $3) AND dtt.dayofweek = 3;
Date: 2026-02-10 06:26:19 Duration: 0ms
Queries generating the largest temporary files
Rank Size Query 1 164.75 MiB with rankedmt4 as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors ), last_feed_entry as ( select * from rankedmt4 where r = 1 ), ok_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where status = 'OK' ), earliest_entry_after_ok as ( select m.datafeedname, min(m.eventtimestamp) as eventtimestamp from mt4datafeederrors m left outer join ( select datafeedname, eventtimestamp from ok_entries where r = 1) oo on m.datafeedname = oo.datafeedname where m.eventtimestamp > coalesce(oo.eventtimestamp, '1900-01-01'::timestamp without time zone) group by m.datafeedname ), notified_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where notified is not null and notified <> '' ), broker as ( select *, row_number() over (partition by feedname order by brokerid) r from ( select distinct b.brokerid, b.name as brokername, dss.classname as feedname from downloadersymbolsettings dss inner join brokersymbollist bsl on dss.symbolid = bsl.symbolid inner join broker b on bsl.brokerid = b.brokerid where dss.enabled = 1) a ) select last.id, last.datafeedname, last.eventtimestamp, last.status, last.errordescription, last.serveraddress, last.username, note.notified, note.eventtimestamp, broker.brokername from last_feed_entry last inner join earliest_entry_after_ok after_ok on last.datafeedname = after_ok.datafeedname inner join broker on last.datafeedname = broker.feedname left outer join ok_entries ok on ok.datafeedname = last.datafeedname left outer join notified_entries note on note.datafeedname = last.datafeedname and note.r = 1 where (ok.r is null or ok.r = 1) and last.datafeedname not in ( select distinct datafeedname from last_feed_entry where status = 'OK') and extract(epoch from (last.eventtimestamp - after_ok.eventtimestamp)) > 60 * 60 and last.eventtimestamp > current_timestamp - interval '1 day' and (note.eventtimestamp is null or note.eventtimestamp < current_timestamp - interval '10 hours') and last.eventtimestamp > current_timestamp - interval '1 hour' and broker.r = 1;[ Date: 2026-02-10 06:20:09 ]
2 163.68 MiB with rankedmt4 as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors ), last_feed_entry as ( select * from rankedmt4 where r = 1 ), ok_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where status = 'OK' ), earliest_entry_after_ok as ( select m.datafeedname, min(m.eventtimestamp) as eventtimestamp from mt4datafeederrors m left outer join ( select datafeedname, eventtimestamp from ok_entries where r = 1) oo on m.datafeedname = oo.datafeedname where m.eventtimestamp > coalesce(oo.eventtimestamp, '1900-01-01'::timestamp without time zone) group by m.datafeedname ), notified_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where notified is not null and notified <> '' ), broker as ( select *, row_number() over (partition by feedname order by brokerid) r from ( select distinct b.brokerid, b.name as brokername, dss.classname as feedname from downloadersymbolsettings dss inner join brokersymbollist bsl on dss.symbolid = bsl.symbolid inner join broker b on bsl.brokerid = b.brokerid where dss.enabled = 1) a ) select last.id, last.datafeedname, last.eventtimestamp, last.status, last.errordescription, last.serveraddress, last.username, note.notified, note.eventtimestamp, broker.brokername from last_feed_entry last inner join earliest_entry_after_ok after_ok on last.datafeedname = after_ok.datafeedname inner join broker on last.datafeedname = broker.feedname left outer join ok_entries ok on ok.datafeedname = last.datafeedname left outer join notified_entries note on note.datafeedname = last.datafeedname and note.r = 1 where (ok.r is null or ok.r = 1) and last.datafeedname not in ( select distinct datafeedname from last_feed_entry where status = 'OK') and extract(epoch from (last.eventtimestamp - after_ok.eventtimestamp)) > 60 * 60 and last.eventtimestamp > current_timestamp - interval '1 day' and (note.eventtimestamp is null or note.eventtimestamp < current_timestamp - interval '10 hours') and last.eventtimestamp > current_timestamp - interval '1 hour' and broker.r = 1;[ Date: 2026-02-10 06:30:10 ]
3 157.52 MiB with rankedmt4 as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors ), last_feed_entry as ( select * from rankedmt4 where r = 1 ), ok_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where status = 'OK' ), earliest_entry_after_ok as ( select m.datafeedname, min(m.eventtimestamp) as eventtimestamp from mt4datafeederrors m left outer join ( select datafeedname, eventtimestamp from ok_entries where r = 1) oo on m.datafeedname = oo.datafeedname where m.eventtimestamp > coalesce(oo.eventtimestamp, '1900-01-01'::timestamp without time zone) group by m.datafeedname ), notified_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where notified is not null and notified <> '' ), broker as ( select *, row_number() over (partition by feedname order by brokerid) r from ( select distinct b.brokerid, b.name as brokername, dss.classname as feedname from downloadersymbolsettings dss inner join brokersymbollist bsl on dss.symbolid = bsl.symbolid inner join broker b on bsl.brokerid = b.brokerid where dss.enabled = 1) a ) select last.id, last.datafeedname, last.eventtimestamp, last.status, last.errordescription, last.serveraddress, last.username, note.notified, note.eventtimestamp, broker.brokername from last_feed_entry last inner join earliest_entry_after_ok after_ok on last.datafeedname = after_ok.datafeedname inner join broker on last.datafeedname = broker.feedname left outer join ok_entries ok on ok.datafeedname = last.datafeedname left outer join notified_entries note on note.datafeedname = last.datafeedname and note.r = 1 where (ok.r is null or ok.r = 1) and last.datafeedname not in ( select distinct datafeedname from last_feed_entry where status = 'OK') and extract(epoch from (last.eventtimestamp - after_ok.eventtimestamp)) > 60 * 60 and last.eventtimestamp > current_timestamp - interval '1 day' and (note.eventtimestamp is null or note.eventtimestamp < current_timestamp - interval '10 hours') and last.eventtimestamp > current_timestamp - interval '1 hour' and broker.r = 1;[ Date: 2026-02-10 06:40:32 ]
4 136.00 MiB select updateresultsmaterializedview ();[ Date: 2026-02-10 06:32:40 ]
5 135.99 MiB select updateresultsmaterializedview ();[ Date: 2026-02-10 06:47:36 ]
6 135.99 MiB select updateresultsmaterializedview ();[ Date: 2026-02-10 06:17:38 ]
7 135.98 MiB select updateresultsmaterializedview ();[ Date: 2026-02-10 06:35:41 ]
8 135.98 MiB select updateresultsmaterializedview ();[ Date: 2026-02-10 06:02:16 ]
9 135.98 MiB select updateresultsmaterializedview ();[ Date: 2026-02-10 06:50:36 ]
10 135.97 MiB select updateresultsmaterializedview ();[ Date: 2026-02-10 06:20:40 ]
11 135.96 MiB select updateresultsmaterializedview ();[ Date: 2026-02-10 06:05:33 ]
12 128.86 MiB with rankedmt4 as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors ), last_feed_entry as ( select * from rankedmt4 where r = 1 ), ok_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where status = 'OK' ), earliest_entry_after_ok as ( select m.datafeedname, min(m.eventtimestamp) as eventtimestamp from mt4datafeederrors m left outer join ( select datafeedname, eventtimestamp from ok_entries where r = 1) oo on m.datafeedname = oo.datafeedname where m.eventtimestamp > coalesce(oo.eventtimestamp, '1900-01-01'::timestamp without time zone) group by m.datafeedname ), notified_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where notified is not null and notified <> '' ), broker as ( select *, row_number() over (partition by feedname order by brokerid) r from ( select distinct b.brokerid, b.name as brokername, dss.classname as feedname from downloadersymbolsettings dss inner join brokersymbollist bsl on dss.symbolid = bsl.symbolid inner join broker b on bsl.brokerid = b.brokerid where dss.enabled = 1) a ) select last.id, last.datafeedname, last.eventtimestamp, last.status, last.errordescription, last.serveraddress, last.username, note.notified, note.eventtimestamp, broker.brokername from last_feed_entry last inner join earliest_entry_after_ok after_ok on last.datafeedname = after_ok.datafeedname inner join broker on last.datafeedname = broker.feedname left outer join ok_entries ok on ok.datafeedname = last.datafeedname left outer join notified_entries note on note.datafeedname = last.datafeedname and note.r = 1 where (ok.r is null or ok.r = 1) and last.datafeedname not in ( select distinct datafeedname from last_feed_entry where status = 'OK') and extract(epoch from (last.eventtimestamp - after_ok.eventtimestamp)) > 60 * 60 and last.eventtimestamp > current_timestamp - interval '1 day' and (note.eventtimestamp is null or note.eventtimestamp < current_timestamp - interval '10 hours') and last.eventtimestamp > current_timestamp - interval '1 hour' and broker.r = 1;[ Date: 2026-02-10 06:00:05 ]
13 121.12 MiB with rankedmt4 as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors ), last_feed_entry as ( select * from rankedmt4 where r = 1 ), ok_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where status = 'OK' ), earliest_entry_after_ok as ( select m.datafeedname, min(m.eventtimestamp) as eventtimestamp from mt4datafeederrors m left outer join ( select datafeedname, eventtimestamp from ok_entries where r = 1) oo on m.datafeedname = oo.datafeedname where m.eventtimestamp > coalesce(oo.eventtimestamp, '1900-01-01'::timestamp without time zone) group by m.datafeedname ), notified_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where notified is not null and notified <> '' ), broker as ( select *, row_number() over (partition by feedname order by brokerid) r from ( select distinct b.brokerid, b.name as brokername, dss.classname as feedname from downloadersymbolsettings dss inner join brokersymbollist bsl on dss.symbolid = bsl.symbolid inner join broker b on bsl.brokerid = b.brokerid where dss.enabled = 1) a ) select last.id, last.datafeedname, last.eventtimestamp, last.status, last.errordescription, last.serveraddress, last.username, note.notified, note.eventtimestamp, broker.brokername from last_feed_entry last inner join earliest_entry_after_ok after_ok on last.datafeedname = after_ok.datafeedname inner join broker on last.datafeedname = broker.feedname left outer join ok_entries ok on ok.datafeedname = last.datafeedname left outer join notified_entries note on note.datafeedname = last.datafeedname and note.r = 1 where (ok.r is null or ok.r = 1) and last.datafeedname not in ( select distinct datafeedname from last_feed_entry where status = 'OK') and extract(epoch from (last.eventtimestamp - after_ok.eventtimestamp)) > 60 * 60 and last.eventtimestamp > current_timestamp - interval '1 day' and (note.eventtimestamp is null or note.eventtimestamp < current_timestamp - interval '10 hours') and last.eventtimestamp > current_timestamp - interval '1 hour' and broker.r = 1;[ Date: 2026-02-10 06:10:04 ]
14 97.34 MiB with rankedmt4 as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors ), last_feed_entry as ( select * from rankedmt4 where r = 1 ), ok_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where status = 'OK' ), earliest_entry_after_ok as ( select m.datafeedname, min(m.eventtimestamp) as eventtimestamp from mt4datafeederrors m left outer join ( select datafeedname, eventtimestamp from ok_entries where r = 1) oo on m.datafeedname = oo.datafeedname where m.eventtimestamp > coalesce(oo.eventtimestamp, '1900-01-01'::timestamp without time zone) group by m.datafeedname ), notified_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where notified is not null and notified <> '' ), broker as ( select *, row_number() over (partition by feedname order by brokerid) r from ( select distinct b.brokerid, b.name as brokername, dss.classname as feedname from downloadersymbolsettings dss inner join brokersymbollist bsl on dss.symbolid = bsl.symbolid inner join broker b on bsl.brokerid = b.brokerid where dss.enabled = 1) a ) select last.id, last.datafeedname, last.eventtimestamp, last.status, last.errordescription, last.serveraddress, last.username, note.notified, note.eventtimestamp, broker.brokername from last_feed_entry last inner join earliest_entry_after_ok after_ok on last.datafeedname = after_ok.datafeedname inner join broker on last.datafeedname = broker.feedname left outer join ok_entries ok on ok.datafeedname = last.datafeedname left outer join notified_entries note on note.datafeedname = last.datafeedname and note.r = 1 where (ok.r is null or ok.r = 1) and last.datafeedname not in ( select distinct datafeedname from last_feed_entry where status = 'OK') and extract(epoch from (last.eventtimestamp - after_ok.eventtimestamp)) > 60 * 60 and last.eventtimestamp > current_timestamp - interval '1 day' and (note.eventtimestamp is null or note.eventtimestamp < current_timestamp - interval '10 hours') and last.eventtimestamp > current_timestamp - interval '1 hour' and broker.r = 1;[ Date: 2026-02-10 06:50:12 ]
15 90.96 MiB with rankedmt4 as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors ), last_feed_entry as ( select * from rankedmt4 where r = 1 ), ok_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where status = 'OK' ), earliest_entry_after_ok as ( select m.datafeedname, min(m.eventtimestamp) as eventtimestamp from mt4datafeederrors m left outer join ( select datafeedname, eventtimestamp from ok_entries where r = 1) oo on m.datafeedname = oo.datafeedname where m.eventtimestamp > coalesce(oo.eventtimestamp, '1900-01-01'::timestamp without time zone) group by m.datafeedname ), notified_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where notified is not null and notified <> '' ), broker as ( select *, row_number() over (partition by feedname order by brokerid) r from ( select distinct b.brokerid, b.name as brokername, dss.classname as feedname from downloadersymbolsettings dss inner join brokersymbollist bsl on dss.symbolid = bsl.symbolid inner join broker b on bsl.brokerid = b.brokerid where dss.enabled = 1) a ) select last.id, last.datafeedname, last.eventtimestamp, last.status, last.errordescription, last.serveraddress, last.username, note.notified, note.eventtimestamp, broker.brokername from last_feed_entry last inner join earliest_entry_after_ok after_ok on last.datafeedname = after_ok.datafeedname inner join broker on last.datafeedname = broker.feedname left outer join ok_entries ok on ok.datafeedname = last.datafeedname left outer join notified_entries note on note.datafeedname = last.datafeedname and note.r = 1 where (ok.r is null or ok.r = 1) and last.datafeedname not in ( select distinct datafeedname from last_feed_entry where status = 'OK') and extract(epoch from (last.eventtimestamp - after_ok.eventtimestamp)) > 60 * 60 and last.eventtimestamp > current_timestamp - interval '1 day' and (note.eventtimestamp is null or note.eventtimestamp < current_timestamp - interval '10 hours') and last.eventtimestamp > current_timestamp - interval '1 hour' and broker.r = 1;[ Date: 2026-02-10 06:10:07 ]
16 85.41 MiB select updateageforrelevantresults ();[ Date: 2026-02-10 06:02:05 ]
17 85.37 MiB select updateageforrelevantresults ();[ Date: 2026-02-10 06:32:24 ]
18 85.32 MiB select updateageforrelevantresults ();[ Date: 2026-02-10 06:17:21 ]
19 85.31 MiB select updateageforrelevantresults ();[ Date: 2026-02-10 06:47:24 ]
20 78.41 MiB with rankedmt4 as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors ), last_feed_entry as ( select * from rankedmt4 where r = 1 ), ok_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where status = 'OK' ), earliest_entry_after_ok as ( select m.datafeedname, min(m.eventtimestamp) as eventtimestamp from mt4datafeederrors m left outer join ( select datafeedname, eventtimestamp from ok_entries where r = 1) oo on m.datafeedname = oo.datafeedname where m.eventtimestamp > coalesce(oo.eventtimestamp, '1900-01-01'::timestamp without time zone) group by m.datafeedname ), notified_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where notified is not null and notified <> '' ), broker as ( select *, row_number() over (partition by feedname order by brokerid) r from ( select distinct b.brokerid, b.name as brokername, dss.classname as feedname from downloadersymbolsettings dss inner join brokersymbollist bsl on dss.symbolid = bsl.symbolid inner join broker b on bsl.brokerid = b.brokerid where dss.enabled = 1) a ) select last.id, last.datafeedname, last.eventtimestamp, last.status, last.errordescription, last.serveraddress, last.username, note.notified, note.eventtimestamp, broker.brokername from last_feed_entry last inner join earliest_entry_after_ok after_ok on last.datafeedname = after_ok.datafeedname inner join broker on last.datafeedname = broker.feedname left outer join ok_entries ok on ok.datafeedname = last.datafeedname left outer join notified_entries note on note.datafeedname = last.datafeedname and note.r = 1 where (ok.r is null or ok.r = 1) and last.datafeedname not in ( select distinct datafeedname from last_feed_entry where status = 'OK') and extract(epoch from (last.eventtimestamp - after_ok.eventtimestamp)) > 60 * 60 and last.eventtimestamp > current_timestamp - interval '1 day' and (note.eventtimestamp is null or note.eventtimestamp < current_timestamp - interval '10 hours') and last.eventtimestamp > current_timestamp - interval '1 hour' and broker.r = 1;[ Date: 2026-02-10 06:50:09 ]
-
Vacuums
Vacuums / Analyzes Distribution
Key values
- 0 sec Highest CPU-cost vacuum
Table
Database - Date
- 0 sec Highest CPU-cost analyze
Table
Database - Date
Analyzes per table
Key values
- public.solr_relevance_old (13) Main table analyzed (database acaweb_fx)
- 55 analyzes Total
Table Number of analyzes acaweb_fx.public.solr_relevance_old 13 acaweb_fx.pg_catalog.pg_attribute 8 acaweb_fx.pg_catalog.pg_class 6 acaweb_fx.pg_catalog.pg_type 4 acaweb_fx.public.relevance_keylevels_results 4 acaweb_fx.public.relevance_autochartist_results 4 acaweb_fx.public.relevance_fibonacci_results 4 acaweb_fx.public.datafeeds_latestrun 3 acaweb_fx.public.latest_t15_candle_view 2 acaweb_fx.pg_catalog.pg_depend 2 acaweb_fx.pg_catalog.pg_index 1 acaweb_fx.public.autochartist_symbolupdates 1 acaweb_fx.public.relevance_consecutivecandles_results 1 acaweb_fx.public.symbollatestupdatetime 1 acaweb_fx.public.latest_candle_datetime_per_receng 1 Total 55 Vacuums per table
Key values
- public.solr_relevance_old (13) Main table vacuumed on database acaweb_fx
- 38 vacuums Total
Index Buffer usage Skipped WAL usage Table Vacuums scans hits misses dirtied pins frozen records full page bytes acaweb_fx.public.solr_relevance_old 13 13 10,202 0 49 0 57 7,287 98 1,629,323 acaweb_fx.pg_catalog.pg_attribute 5 3 3,533 0 258 0 335 1,197 219 1,402,543 acaweb_fx.pg_catalog.pg_type 3 3 503 0 71 0 0 197 46 241,987 acaweb_fx.public.datafeeds_latestrun 3 0 360 0 7 0 0 33 5 34,481 acaweb_fx.pg_toast.pg_toast_2619 2 2 267 0 64 0 0 207 55 216,922 acaweb_fx.public.relevance_keylevels_results 2 2 7,638 0 245 4 201 1,804 207 771,639 acaweb_fx.pg_catalog.pg_class 2 2 924 0 76 0 0 320 77 417,883 acaweb_fx.public.relevance_autochartist_results 2 2 6,463 0 153 2 484 1,251 118 421,570 acaweb_fx.public.relevance_fibonacci_results 2 2 2,353 0 76 2 100 319 60 222,757 acaweb_fx.public.autochartist_symbolupdates 1 1 21,257 0 2,994 1 38,420 5,403 3,003 1,518,287 acaweb_fx.pg_catalog.pg_statistic 1 1 987 0 217 0 582 464 210 774,162 acaweb_fx.public.latest_t15_candle_view 1 1 66 0 1 0 0 6 1 9,037 acaweb_fx.pg_catalog.pg_depend 1 1 368 0 66 0 59 168 59 302,927 Total 38 33 54,921 49,121 4,277 9 40,238 18,656 4,158 7,963,518 Tuples removed per table
Key values
- public.solr_relevance_old (48802) Main table with removed tuples on database acaweb_fx
- 68981 tuples Total removed
Index Tuples Pages Table Vacuums scans removed remain not yet removable removed remain acaweb_fx.public.solr_relevance_old 13 13 48,802 76,292 0 0 2,623 acaweb_fx.pg_catalog.pg_attribute 5 3 6,464 59,636 6,565 27 1,335 acaweb_fx.public.autochartist_symbolupdates 1 1 5,174 48,733 14 0 40,691 acaweb_fx.public.relevance_keylevels_results 2 2 2,951 24,397 0 0 558 acaweb_fx.public.relevance_autochartist_results 2 2 1,752 17,262 0 0 760 acaweb_fx.pg_catalog.pg_depend 1 1 1,197 14,410 0 0 140 acaweb_fx.pg_catalog.pg_type 3 3 781 4,514 176 0 131 acaweb_fx.pg_catalog.pg_statistic 1 1 671 3,701 0 0 1,194 acaweb_fx.pg_catalog.pg_class 2 2 472 3,298 0 0 300 acaweb_fx.public.relevance_fibonacci_results 2 2 369 3,303 147 0 204 acaweb_fx.pg_toast.pg_toast_2619 2 2 156 346 10 0 108 acaweb_fx.public.datafeeds_latestrun 3 0 136 70 28 0 48 acaweb_fx.public.latest_t15_candle_view 1 1 56 14 0 0 1 Total 38 33 68,981 255,976 6,940 27 48,093 Pages removed per table
Key values
- pg_catalog.pg_attribute (27) Main table with removed pages on database acaweb_fx
- 27 pages Total removed
Table Number of vacuums Index scans Tuples removed Pages removed acaweb_fx.pg_catalog.pg_attribute 5 3 6464 27 acaweb_fx.pg_toast.pg_toast_2619 2 2 156 0 acaweb_fx.pg_catalog.pg_type 3 3 781 0 acaweb_fx.public.autochartist_symbolupdates 1 1 5174 0 acaweb_fx.public.datafeeds_latestrun 3 0 136 0 acaweb_fx.pg_catalog.pg_statistic 1 1 671 0 acaweb_fx.public.latest_t15_candle_view 1 1 56 0 acaweb_fx.pg_catalog.pg_depend 1 1 1197 0 acaweb_fx.public.relevance_keylevels_results 2 2 2951 0 acaweb_fx.pg_catalog.pg_class 2 2 472 0 acaweb_fx.public.solr_relevance_old 13 13 48802 0 acaweb_fx.public.relevance_autochartist_results 2 2 1752 0 acaweb_fx.public.relevance_fibonacci_results 2 2 369 0 Total 38 33 68,981 27 Autovacuum Activity
↑ Back to the top of the Autovacuum Activity tableDay Hour VACUUMs ANALYZEs Feb 10 06 38 55 - 0 sec Highest CPU-cost vacuum
-
Locks
Locks by types
Key values
- AccessShareLock Main Lock Type
- 5 locks Total
Most frequent waiting queries (N)
Rank Count Total time Min time Max time Avg duration Query 1 1 14m42s 14m42s 14m42s 14m42s truncate table solr_relevance_old;-
TRUNCATE TABLE solr_relevance_old;
Date: 2026-02-10 06:19:54
2 2 5m6s 2m25s 2m41s 2m33s select datetimeupdate from latest_candle_datetime_per_receng where recognitionengine ilike ?;-
SELECT datetimeupdate FROM latest_candle_datetime_per_receng WHERE recognitionengine ILIKE 'IQFEED_FX - 1';
Date: 2026-02-10 06:19:55
3 1 2m52s 2m52s 2m52s 2m52s refresh materialized view latest_candle_datetime_per_receng;-
refresh materialized view latest_candle_datetime_per_receng;
Date: 2026-02-10 06:19:54
4 1 40s282ms 40s282ms 40s282ms 40s282ms select recognitionengine, to_char(datetimeupdate, ?) from latest_candle_datetime_per_receng;-
select recognitionengine, to_char(datetimeupdate, 'yyyy-mm-dd HH24:MI') from latest_candle_datetime_per_receng;
Date: 2026-02-10 06:19:55
Queries that waited the most
Rank Wait time Query 1 14m42s TRUNCATE TABLE solr_relevance_old;[ Date: 2026-02-10 06:19:54 ]
2 2m52s refresh materialized view latest_candle_datetime_per_receng;[ Date: 2026-02-10 06:19:54 ]
3 2m41s SELECT datetimeupdate FROM latest_candle_datetime_per_receng WHERE recognitionengine ILIKE 'IQFEED_FX - 1';[ Date: 2026-02-10 06:19:55 ]
4 2m25s SELECT datetimeupdate FROM latest_candle_datetime_per_receng WHERE recognitionengine ILIKE 'PEPPERSTONE - 1';[ Date: 2026-02-10 06:19:55 ]
5 40s282ms select recognitionengine, to_char(datetimeupdate, 'yyyy-mm-dd HH24:MI') from latest_candle_datetime_per_receng;[ Date: 2026-02-10 06:19:55 ]
-
Queries
Queries by type
Key values
- 73,010 Total read queries
- 36,924 Total write queries
Queries by database
Key values
- unknown Main database
- 231,135 Requests
- 3h43m57s (unknown)
- Main time consuming database
Database Request type Count Duration acaweb_fx Total 811 0ms copy from 65 0ms copy to 26 0ms cte 85 0ms ddl 13 0ms delete 13 0ms others 168 0ms select 73 0ms tcl 331 0ms update 37 0ms acaweb_fx_integer Total 1 0ms select 1 0ms postgres Total 3 0ms select 3 0ms socialmedia Total 89 0ms select 87 0ms tcl 2 0ms translations Total 1 0ms select 1 0ms unknown Total 231,135 3h43m57s copy from 13 0ms copy to 575 0ms cte 7,897 0ms insert 25,851 0ms others 5,918 0ms select 72,845 0ms tcl 368 0ms update 1,648 0ms Queries by user
Key values
- unknown Main user
- 231,135 Requests
User Request type Count Duration postgres Total 905 0ms copy from 65 0ms copy to 26 0ms cte 85 0ms ddl 13 0ms delete 13 0ms others 168 0ms select 165 0ms tcl 333 0ms update 37 0ms unknown Total 231,135 3h43m57s copy from 13 0ms copy to 575 0ms cte 7,897 0ms insert 25,851 0ms others 5,918 0ms select 72,845 0ms tcl 368 0ms update 1,648 0ms Duration by user
Key values
- 3h43m57s (unknown) Main time consuming user
User Request type Count Duration postgres Total 905 0ms copy from 65 0ms copy to 26 0ms cte 85 0ms ddl 13 0ms delete 13 0ms others 168 0ms select 165 0ms tcl 333 0ms update 37 0ms unknown Total 231,135 3h43m57s copy from 13 0ms copy to 575 0ms cte 7,897 0ms insert 25,851 0ms others 5,918 0ms select 72,845 0ms tcl 368 0ms update 1,648 0ms Queries by host
Key values
- unknown Main host
- 232,040 Requests
- 3h43m57s (unknown)
- Main time consuming host
Queries by application
Key values
- unknown Main application
- 231,718 Requests
- 3h43m57s (unknown)
- Main time consuming application
Application Request type Count Duration pg_dump Total 5 0ms select 5 0ms psql Total 317 0ms copy from 65 0ms copy to 26 0ms cte 85 0ms ddl 13 0ms delete 13 0ms others 4 0ms select 74 0ms update 37 0ms unknown Total 231,718 3h43m57s copy from 13 0ms copy to 575 0ms cte 7,897 0ms insert 25,851 0ms others 6,082 0ms select 72,931 0ms tcl 701 0ms update 1,648 0ms Number of cancelled queries
Key values
- 0 per second Cancelled query Peak
- 2026-02-10 06:49:12 Date
Number of cancelled queries (5 minutes period)
NO DATASET
-
Top Queries
Histogram of query times
Key values
- 74,265 0-1ms duration
Slowest individual queries
Rank Duration Query NO DATASET
Time consuming queries
Rank Total duration Times executed Min duration Max duration Avg duration Query 1 0ms 1 0ms 0ms 0ms lock table public.stats_hrs_summary_by_groups in access share mode;Times Reported Time consuming queries #1
Day Hour Count Duration Avg duration Feb 10 06 1 0ms 0ms 2 0ms 2 0ms 0ms 0ms copy public.powerstats_trumpet (symbolid, dayofweek, fromtime, enddate, stddev_15, ave_15, stddev_30, ave_30, stddev_60, ave_60, stddev_240, ave_240, stddev_1440, ave_1440) to stdout;Times Reported Time consuming queries #2
Day Hour Count Duration Avg duration Feb 10 06 2 0ms 0ms 3 0ms 2 0ms 0ms 0ms select last_value, is_called from public.stats_by_groups_statsid_seq;Times Reported Time consuming queries #3
Day Hour Count Duration Avg duration Feb 10 06 2 0ms 0ms 4 0ms 2 0ms 0ms 0ms lock table public.timezones_korean in access share mode;Times Reported Time consuming queries #4
Day Hour Count Duration Avg duration Feb 10 06 2 0ms 0ms 5 0ms 27 0ms 0ms 0ms select key, value from datasources ds inner join datasourceparams dsp on ds.id = dsp.datasourceid where ds.name = ?;Times Reported Time consuming queries #5
Day Hour Count Duration Avg duration Feb 10 06 27 0ms 0ms 6 0ms 2 0ms 0ms 0ms copy public.t30 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived, pricesadjusted, relevanceprocessed, recengprocessedadjusted, historical) to stdout;Times Reported Time consuming queries #6
Day Hour Count Duration Avg duration Feb 10 06 2 0ms 0ms 7 0ms 1 0ms 0ms 0ms copy public.processresults (id, processid, resultdate, shorttext, longtext, title, test) to stdout;Times Reported Time consuming queries #7
Day Hour Count Duration Avg duration Feb 10 06 1 0ms 0ms 8 0ms 2,172 0ms 0ms 0ms insert into t60 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) values (?, ?, ?, ?, ?, ?, ?, ?, ?, ?) on conflict (pricedatetime, symbolid) do update set open = ?, high = ?, low = ?, close = ?, volume = ?, bsf = ?, sastdatetimewritten = ?, sastdatetimereceived = ?;Times Reported Time consuming queries #8
Day Hour Count Duration Avg duration Feb 10 06 2,172 0ms 0ms 9 0ms 2 0ms 0ms 0ms select last_value, is_called from public.keylevels_results_resultuid_seq19;Times Reported Time consuming queries #9
Day Hour Count Duration Avg duration Feb 10 06 2 0ms 0ms 10 0ms 2 0ms 0ms 0ms copy archive.whatshot (whid, brokerid, whdatetime, jobcompletiontime) to stdout;Times Reported Time consuming queries #10
Day Hour Count Duration Avg duration Feb 10 06 2 0ms 0ms 11 0ms 5 0ms 0ms 0ms set synchronize_seqscans to off;Times Reported Time consuming queries #11
Day Hour Count Duration Avg duration Feb 10 06 5 0ms 0ms 12 0ms 1 0ms 0ms 0ms select last_value, is_called from public."ProcessLocaleParameters_id_seq";Times Reported Time consuming queries #12
Day Hour Count Duration Avg duration Feb 10 06 1 0ms 0ms 13 0ms 975 0ms 0ms 0ms select oid, tableoid, pol.polname, pol.polcmd, pol.polpermissive, case when pol.polroles = ? then null else pg_catalog.array_to_string(array ( select pg_catalog.quote_ident(rolname) from pg_catalog.pg_roles where oid = any (pol.polroles)), ?) end as polroles, pg_catalog.pg_get_expr(pol.polqual, pol.polrelid) as polqual, pg_catalog.pg_get_expr(pol.polwithcheck, pol.polrelid) as polwithcheck from pg_catalog.pg_policy pol where polrelid = ?;Times Reported Time consuming queries #13
Day Hour Count Duration Avg duration Feb 10 06 975 0ms 0ms 14 0ms 2 0ms 0ms 0ms lock table public.timezones_turkish in access share mode;Times Reported Time consuming queries #14
Day Hour Count Duration Avg duration Feb 10 06 2 0ms 0ms 15 0ms 2 0ms 0ms 0ms copy public.timezones_greek (timezoneid, timezone) to stdout;Times Reported Time consuming queries #15
Day Hour Count Duration Avg duration Feb 10 06 2 0ms 0ms 16 0ms 2 0ms 0ms 0ms lock table public.futures_symbols in access share mode;Times Reported Time consuming queries #16
Day Hour Count Duration Avg duration Feb 10 06 2 0ms 0ms 17 0ms 2 0ms 0ms 0ms copy archive.autochartist_results (resultid, bandwidth, pattern, gmttimefound, direction, initialtrend, breakout, volumeincrease, noise, symmetry, patternstarttime, patternendtime, patternstartprice, patternendprice, resx0, resx1, supportx0, supportx1, resy0, resy1, supporty0, supporty1, supportgradient, resgradient, riskreward, qtytp, predictionpricefrom, predictionpriceto, predictiontimefrom, predictiontimeto, patternquality, trendchange, maxmovementafterbreakout, latestbaratbreakouttime, latestbaratbreakoutprice, patternlengthbars, temporarypattern, symbolid, resultuid, relevancestartdistance, simulation, writtendatetime, old_resultuid) to stdout;Times Reported Time consuming queries #17
Day Hour Count Duration Avg duration Feb 10 06 2 0ms 0ms 18 0ms 2 0ms 0ms 0ms copy public.satelliteauthentication (id, satelliteserverid, brokerid, authenticationstatus, maketradeurl, notifications, validusername, validpassword, enabled) to stdout;Times Reported Time consuming queries #18
Day Hour Count Duration Avg duration Feb 10 06 2 0ms 0ms 19 0ms 2 0ms 0ms 0ms lock table archive.patternresultsage in access share mode;Times Reported Time consuming queries #19
Day Hour Count Duration Avg duration Feb 10 06 2 0ms 0ms 20 0ms 1 0ms 0ms 0ms copy public.processes (id, locale, region, schedule, enabled, live, lastmodified, lastrun, contenttypeid, brokerid, uuid) to stdout;Times Reported Time consuming queries #20
Day Hour Count Duration Avg duration Feb 10 06 1 0ms 0ms Most frequent queries (N)
Rank Times executed Total duration Min duration Max duration Avg duration Query 1 28,451 0ms 0ms 0ms 0ms select ?;Times Reported Time consuming queries #1
Day Hour Count Duration Avg duration Feb 10 06 28,451 0ms 0ms 2 18,130 0ms 0ms 0ms 0ms select distinct on (coalesce(bim.code, s.symbol) , s.exchange, s.timegranularity, df.timezone) s.symbolid as id, coalesce(bim.code, s.symbol) as name, s.symbol as symbol, dss.downloadersymbol as ticker, s.exchange as exchange, s.timegranularity as interval, df.timezone as timezone from symbols s inner join downloadersymbolsettings dss on dss.symbolid = s.symbolid inner join datafeedstimetable df on df.classname ilike dss.classname left join brokersymbollist bsl on bsl.brokerid = ? and bsl.symbolid = s.symbolid left outer join brokerinstrumentmap bim on dss.datafeedinstrumentid = bim.datafeedinstrumentid and bim.brokerid = ? and bim.type = ? where s.symbolid = ?;Times Reported Time consuming queries #2
Day Hour Count Duration Avg duration Feb 10 06 18,130 0ms 0ms 3 6,485 0ms 0ms 0ms 0ms insert into executionlogs (executionid, status, message, details, detailtype) values (null, ?, ?, null, null);Times Reported Time consuming queries #3
Day Hour Count Duration Avg duration Feb 10 06 6,485 0ms 0ms 4 6,377 0ms 0ms 0ms 0ms select s.symbolid as id, s.symbol as name, s.exchange as exchange, s.timegranularity as interval, dtt.timezone as timezone from symbols s inner join downloadersymbolsettings dss on dss.symbolid = s.symbolid inner join datafeedstimetable dtt on dss.classname = dtt.classname and dtt.dayofweek = ? inner join brokersymbollist bsl on bsl.symbolid = s.symbolid where bsl.brokerid = ? and (? = ? or s.timegranularity = ?) and (s.symbol = ? or dss.downloadersymbol = ?) and dss.enabled = ?;Times Reported Time consuming queries #4
Day Hour Count Duration Avg duration Feb 10 06 6,377 0ms 0ms 5 5,001 0ms 0ms 0ms 0ms insert into t15 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) values (?, ?, ?, ?, ?, ?, ?, ?, ?, ?) on conflict (pricedatetime, symbolid) do update set open = ?, high = ?, low = ?, close = ?, volume = ?, bsf = ?, sastdatetimewritten = ?, sastdatetimereceived = ?;Times Reported Time consuming queries #5
Day Hour Count Duration Avg duration Feb 10 06 5,001 0ms 0ms 6 4,219 0ms 0ms 0ms 0ms insert into autochartist_results (resultid, symbolid, bandwidth, pattern, qtytp, gmttimefound, direction, initialtrend, breakout, volumeincrease, noise, symmetry, predictionpricefrom, predictionpriceto, predictiontimefrom, predictiontimeto, patternstarttime, patternendtime, patternstartprice, patternendprice, resx0, resx1, supportx0, supportx1, resy0, resy1, supporty0, supporty1, supportgradient, resgradient, riskreward, patternquality, trendchange, maxmovementafterbreakout, latestbaratbreakouttime, latestbaratbreakoutprice, patternlengthbars, temporarypattern, relevancestartdistance, simulation, writtendatetime) values (?, ?, ?.?, ?, ?, ?::timestamp without time zone, ?, ?.?, ?.?, ?.?, ?.?, ?.?, ?.?, ?.?, ?::timestamp without time zone, ?::timestamp without time zone, ?::timestamp without time zone, ?::timestamp without time zone, ?.?, ?.?, ?::timestamp without time zone, ?::timestamp without time zone, ?::timestamp without time zone, ?::timestamp without time zone, ?.?, ?.?, ?.?, ?.?, ?.?, ?.?, ?.?, ?.?, ?, ?.?, ?::timestamp without time zone, ?.?, ?, ?, ?.?, ?, current_timestamp::timestamp without time zone) on conflict do nothing;Times Reported Time consuming queries #6
Day Hour Count Duration Avg duration Feb 10 06 4,219 0ms 0ms 7 3,087 0ms 0ms 0ms 0ms with rar_max as ( select resultuid from relevance_autochartist_results order by resultuid desc limit ? ) select a.symbolid, pattern, patternid, resy0, resy1, resx0, resx1, supporty0, supporty1, supportx0, supportx1, predictiontimeto, patternstarttime, timegranularity, patternendtime, direction, trendchange, patternlengthbars, patternquality, case when a.old_resultuid = ? then a.old_resultuid else a.resultuid end as uid, breakout, initialtrend, volumeincrease, symmetry as uniformity, predictionpricefrom, predictionpriceto, noise, s.exchange, s.symbol, s.longname, s.shortname, breakout, dtt.timezone, patternstartprice, patternendprice, qtytp, newlevels.profit, newlevels.stop, newlevels.filtered, case when rar.age is not null then rar.age when a.resultuid <= rm.resultuid then ? else ? end as age, case when rar.relevant is not null then rar.relevant when a.resultuid <= rm.resultuid then ? else ? end as relevant, cps.pip from autochartist_results a inner join downloadersymbolsettings dss on a.symbolid = dss.symbolid inner join datafeedstimetable dtt on dss.classname = dtt.classname inner join symbols s on a.symbolid = s.symbolid inner join patterns p on p.patternname = a.pattern inner join rar_max rm on ? = ? left outer join relevance_autochartist_results rar on rar.resultuid = a.resultuid left join lateral calc_cp_signal (a.resultuid) newlevels on true left join currencypips cps on cps.symbol = s.symbol where (a.old_resultuid = ? or a.resultuid = ?) and dtt.dayofweek = ?;Times Reported Time consuming queries #7
Day Hour Count Duration Avg duration Feb 10 06 3,087 0ms 0ms 8 2,991 0ms 0ms 0ms 0ms insert into t30 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) values (?, ?, ?, ?, ?, ?, ?, ?, ?, ?) on conflict (pricedatetime, symbolid) do update set open = ?, high = ?, low = ?, close = ?, volume = ?, bsf = ?, sastdatetimewritten = ?, sastdatetimereceived = ?;Times Reported Time consuming queries #8
Day Hour Count Duration Avg duration Feb 10 06 2,991 0ms 0ms 9 2,444 0ms 0ms 0ms 0ms set extra_float_digits = ?;Times Reported Time consuming queries #9
Day Hour Count Duration Avg duration Feb 10 06 2,444 0ms 0ms 10 2,418 0ms 0ms 0ms 0ms set application_name = ?;Times Reported Time consuming queries #10
Day Hour Count Duration Avg duration Feb 10 06 2,418 0ms 0ms 11 2,204 0ms 0ms 0ms 0ms insert into fibonacci_results (bandwidth, pattern, gmttimefound, direction, patternstarttime, patternendtime, patternstartprice, patternendprice, qtytp, pricex, timex, pricea, timea, priceb, timeb, pricec, timec, priced, timed, averagequality, timequality, errormargin, patternlengthbars, target10, target06, target16, target07, target12, target05, target03, symbolid, noise, ratiosfound, temporarypattern, uniqueindex, completed, simulation, writtendatetime) values (?.?, ?, ?::timestamp without time zone, ?, ?::timestamp without time zone, ?::timestamp without time zone, ?.?, ?.?, ?, ?.?, ?::timestamp without time zone, ?.?, ?::timestamp without time zone, ?.?, ?::timestamp without time zone, ?.?, ?::timestamp without time zone, ?.?, ?::timestamp without time zone, ?.?, ?.?, ?.?, ?, ?.?, ?.?, ?.?, ?.?, ?.?, ?.?, ?.?, ?, ?.?, ?, ?, ?, ?, ?, current_timestamp::timestamp without time zone) on conflict do nothing;Times Reported Time consuming queries #11
Day Hour Count Duration Avg duration Feb 10 06 2,204 0ms 0ms 12 2,172 0ms 0ms 0ms 0ms insert into t60 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) values (?, ?, ?, ?, ?, ?, ?, ?, ?, ?) on conflict (pricedatetime, symbolid) do update set open = ?, high = ?, low = ?, close = ?, volume = ?, bsf = ?, sastdatetimewritten = ?, sastdatetimereceived = ?;Times Reported Time consuming queries #12
Day Hour Count Duration Avg duration Feb 10 06 2,172 0ms 0ms 13 1,663 0ms 0ms 0ms 0ms select pg_catalog.format_type(?::pg_catalog.oid, null);Times Reported Time consuming queries #13
Day Hour Count Duration Avg duration Feb 10 06 1,663 0ms 0ms 14 1,578 0ms 0ms 0ms 0ms with rar_max as ( select resultuid from relevance_keylevels_results order by resultuid desc limit ? ) select case when a.old_resultuid = ? then a.old_resultuid else a.resultuid end as ruid, s.symbolid as sid, s.symbol as sym, longname, shortname, exchange as e, timegranularity as tg, a.patternid as pid, a.direction as d, a.patternprice as pp, atbaridentified as pet, case when (x9 != ?) then x9 when (x8 != ?) then x8 when (x7 != ?) then x7 when (x6 != ?) then x6 when (x5 != ?) then x5 when (x4 != ?) then x4 when (x3 != ?) then x3 when (x2 != ?) then x2 end as pst, patternprice as patp, x0, x1, x2, case when (x3 != ?) then x3 else ? end as x3, case when (x4 != ?) then x4 else ? end as x4, case when (x5 != ?) then x5 else ? end as x5, case when (x6 != ?) then x6 else ? end as x6, case when (x7 != ?) then x7 else ? end as x7, case when (x8 != ?) then x8 else ? end as x8, errormargin as erm, breakoutprice as pe, breakoutbars as be, breakout, atbaridentified as atbar, atpriceidentified as atprice, patternlengthbars as l, bandwidth as bw, qtytp as qtp, p.patternname as patternname, dtt.absolutetimezoneoffset as tzos, dtt.timezone as timezone, approachingtimestamp as apt, approachingregion as apr, predictionpricefrom as ppf, predictionpriceto as ppt, predictiontimefrom as ptf, predictiontimebars as ptb, furthestprice as fp, newlevels.filtered, a.uniquepointsvalue as upv, case when rar.age is not null then rar.age when a.resultuid <= rm.resultuid then ? else ? end as age, case when rar.relevant is not null then rar.relevant when a.resultuid <= rm.resultuid then ? else ? end as relevant, cps.pip from keylevels_results a inner join downloadersymbolsettings dss on a.symbolid = dss.symbolid inner join datafeedstimetable dtt on dss.classname = dtt.classname inner join symbols s on a.symbolid = s.symbolid inner join hrspatterns p on a.patternid = p.patternid inner join rar_max rm on ? = ? left outer join relevance_keylevels_results rar on a.resultuid = rar.resultuid left join lateral calc_kl_signal_filter (a.resultuid) newlevels on true left join currencypips cps on cps.symbol = s.symbol where (a.old_resultuid = ? or a.resultuid = ?) and dtt.dayofweek = ?;Times Reported Time consuming queries #14
Day Hour Count Duration Avg duration Feb 10 06 1,578 0ms 0ms 15 1,500 0ms 0ms 0ms 0ms insert into keylevels_results (bandwidth, breakout, patternid, gmttimefound, approachingtimestamp, approachingregion, qtytp, patternlengthbars, patternprice, x0, x1, x2, x3, x4, x5, x6, x7, x8, x9, breakoutbars, breakoutprice, patternendtime, atbaridentified, atpriceidentified, errormargin, direction, symbolid, predictionpricefrom, predictionpriceto, predictiontimefrom, predictiontimebars, uniquepointsvalue, furthestprice, relevancestartdistance, patternclassid, patternstarttime, stoplosslevel, simulation, writtendatetime) values (?.?, ?, ?, ?::timestamp without time zone, ?, ?.?, ?, ?, ?.?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?.?, ?::timestamp without time zone, ?, ?.?, ?.?, ?, ?, ?.?, ?.?, ?::timestamp without time zone, ?, ?, ?.?, ?.?, ?, ?, ?.?, ?, current_timestamp::timestamp without time zone) on conflict do nothing;Times Reported Time consuming queries #15
Day Hour Count Duration Avg duration Feb 10 06 1,500 0ms 0ms 16 1,207 0ms 0ms 0ms 0ms update patternresultsrelevance set relevant = ?, saxo_relevant = ?, notrelevantpricedatetime = ?, reason = ? where uniqueindex = ? and relevant = ?;Times Reported Time consuming queries #16
Day Hour Count Duration Avg duration Feb 10 06 1,207 0ms 0ms 17 988 0ms 0ms 0ms 0ms select symbolid, pricedatetime, classname, downloadfrequency, downloadersymbol, open, high, low, close, volume, bsf, sastdatetimereceived from ( select pricedatetime, dss.classname, dss.downloadfrequency, dss.symbolid, dss.downloadersymbol, t.open, t.high, t.low, t.close, t.volume, t.bsf, t.sastdatetimereceived, row_number() over (partition by t.symbolid order by t.pricedatetime desc) as rn from t15 t, downloadersymbolsettings dss, symbols s where dss.classname = ? and dss.downloadfrequency = ? and dss.symbolid = t.symbolid and s.symbolid = dss.symbolid and dss.enabled = ? and s.deleted = ? and dss.downloadersymbol in (...) and t.pricedatetime > now() - interval ?) as ranked_candles_table where rn = ?;Times Reported Time consuming queries #17
Day Hour Count Duration Avg duration Feb 10 06 988 0ms 0ms 18 975 0ms 0ms 0ms 0ms select oid, tableoid, pol.polname, pol.polcmd, pol.polpermissive, case when pol.polroles = ? then null else pg_catalog.array_to_string(array ( select pg_catalog.quote_ident(rolname) from pg_catalog.pg_roles where oid = any (pol.polroles)), ?) end as polroles, pg_catalog.pg_get_expr(pol.polqual, pol.polrelid) as polqual, pg_catalog.pg_get_expr(pol.polwithcheck, pol.polrelid) as polwithcheck from pg_catalog.pg_policy pol where polrelid = ?;Times Reported Time consuming queries #18
Day Hour Count Duration Avg duration Feb 10 06 975 0ms 0ms 19 785 0ms 0ms 0ms 0ms select a.attnum, a.attname, a.atttypmod, a.attstattarget, a.attstorage, t.typstorage, a.attnotnull, a.atthasdef, a.attisdropped, a.attlen, a.attalign, a.attislocal, pg_catalog.format_type(t.oid, a.atttypmod) as atttypname, a.attgenerated, case when a.atthasmissing and not a.attisdropped then a.attmissingval else null end as attmissingval, a.attidentity, pg_catalog.array_to_string(array ( select pg_catalog.quote_ident(option_name) || ? || pg_catalog.quote_literal(option_value) from pg_catalog.pg_options_to_table(attfdwoptions) order by option_name), e ?) as attfdwoptions, case when a.attcollation <> t.typcollation then a.attcollation else ? end as attcollation, array_to_string(a.attoptions, ?) as attoptions from pg_catalog.pg_attribute a left join pg_catalog.pg_type t on a.atttypid = t.oid where a.attrelid = ?::pg_catalog.oid and a.attnum > ?::pg_catalog.int2 order by a.attnum;Times Reported Time consuming queries #19
Day Hour Count Duration Avg duration Feb 10 06 785 0ms 0ms 20 677 0ms 0ms 0ms 0ms select at.attname, ( select pg_catalog.array_agg(acl order by row_n) from ( select acl, row_n from pg_catalog.unnest(coalesce(at.attacl, pg_catalog.acldefault(?, c.relowner))) with ordinality as perm (acl, row_n) where not exists ( select ? from pg_catalog.unnest(coalesce(pip.initprivs, pg_catalog.acldefault(?, c.relowner))) as init (init_acl) where acl = init_acl)) as foo) as attacl, ( select pg_catalog.array_agg(acl order by row_n) from ( select acl, row_n from pg_catalog.unnest(coalesce(pip.initprivs, pg_catalog.acldefault(?, c.relowner))) with ordinality as initp (acl, row_n) where not exists ( select ? from pg_catalog.unnest(coalesce(at.attacl, pg_catalog.acldefault(?, c.relowner))) as permp (orig_acl) where acl = orig_acl)) as foo) as rattacl, null as initattacl, null as initrattacl from pg_catalog.pg_attribute at join pg_catalog.pg_class c on (at.attrelid = c.oid) left join pg_catalog.pg_init_privs pip on (at.attrelid = pip.objoid and pip.classoid = ?::pg_catalog.regclass and at.attnum = pip.objsubid) where at.attrelid = ?::pg_catalog.oid and not at.attisdropped and (( select pg_catalog.array_agg(acl order by row_n) from ( select acl, row_n from pg_catalog.unnest(coalesce(at.attacl, pg_catalog.acldefault(?, c.relowner))) with ordinality as perm (acl, row_n) where not exists ( select ? from pg_catalog.unnest(coalesce(pip.initprivs, pg_catalog.acldefault(?, c.relowner))) as init (init_acl) where acl = init_acl)) as foo) is not null or ( select pg_catalog.array_agg(acl order by row_n) from ( select acl, row_n from pg_catalog.unnest(coalesce(pip.initprivs, pg_catalog.acldefault(?, c.relowner))) with ordinality as initp (acl, row_n) where not exists ( select ? from pg_catalog.unnest(coalesce(at.attacl, pg_catalog.acldefault(?, c.relowner))) as permp (orig_acl) where acl = orig_acl)) as foo) is not null or null is not null or null is not null) order by at.attnum;Times Reported Time consuming queries #20
Day Hour Count Duration Avg duration Feb 10 06 677 0ms 0ms Normalized slowest queries (N)
Rank Min duration Max duration Avg duration Times executed Total duration Query 1 0ms 0ms 0ms 1 0ms lock table public.stats_hrs_summary_by_groups in access share mode;Times Reported Time consuming queries #1
Day Hour Count Duration Avg duration Feb 10 06 1 0ms 0ms 2 0ms 0ms 0ms 2 0ms copy public.powerstats_trumpet (symbolid, dayofweek, fromtime, enddate, stddev_15, ave_15, stddev_30, ave_30, stddev_60, ave_60, stddev_240, ave_240, stddev_1440, ave_1440) to stdout;Times Reported Time consuming queries #2
Day Hour Count Duration Avg duration Feb 10 06 2 0ms 0ms 3 0ms 0ms 0ms 2 0ms select last_value, is_called from public.stats_by_groups_statsid_seq;Times Reported Time consuming queries #3
Day Hour Count Duration Avg duration Feb 10 06 2 0ms 0ms 4 0ms 0ms 0ms 2 0ms lock table public.timezones_korean in access share mode;Times Reported Time consuming queries #4
Day Hour Count Duration Avg duration Feb 10 06 2 0ms 0ms 5 0ms 0ms 0ms 27 0ms select key, value from datasources ds inner join datasourceparams dsp on ds.id = dsp.datasourceid where ds.name = ?;Times Reported Time consuming queries #5
Day Hour Count Duration Avg duration Feb 10 06 27 0ms 0ms 6 0ms 0ms 0ms 2 0ms copy public.t30 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived, pricesadjusted, relevanceprocessed, recengprocessedadjusted, historical) to stdout;Times Reported Time consuming queries #6
Day Hour Count Duration Avg duration Feb 10 06 2 0ms 0ms 7 0ms 0ms 0ms 1 0ms copy public.processresults (id, processid, resultdate, shorttext, longtext, title, test) to stdout;Times Reported Time consuming queries #7
Day Hour Count Duration Avg duration Feb 10 06 1 0ms 0ms 8 0ms 0ms 0ms 2,172 0ms insert into t60 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) values (?, ?, ?, ?, ?, ?, ?, ?, ?, ?) on conflict (pricedatetime, symbolid) do update set open = ?, high = ?, low = ?, close = ?, volume = ?, bsf = ?, sastdatetimewritten = ?, sastdatetimereceived = ?;Times Reported Time consuming queries #8
Day Hour Count Duration Avg duration Feb 10 06 2,172 0ms 0ms 9 0ms 0ms 0ms 2 0ms select last_value, is_called from public.keylevels_results_resultuid_seq19;Times Reported Time consuming queries #9
Day Hour Count Duration Avg duration Feb 10 06 2 0ms 0ms 10 0ms 0ms 0ms 2 0ms copy archive.whatshot (whid, brokerid, whdatetime, jobcompletiontime) to stdout;Times Reported Time consuming queries #10
Day Hour Count Duration Avg duration Feb 10 06 2 0ms 0ms 11 0ms 0ms 0ms 5 0ms set synchronize_seqscans to off;Times Reported Time consuming queries #11
Day Hour Count Duration Avg duration Feb 10 06 5 0ms 0ms 12 0ms 0ms 0ms 1 0ms select last_value, is_called from public."ProcessLocaleParameters_id_seq";Times Reported Time consuming queries #12
Day Hour Count Duration Avg duration Feb 10 06 1 0ms 0ms 13 0ms 0ms 0ms 975 0ms select oid, tableoid, pol.polname, pol.polcmd, pol.polpermissive, case when pol.polroles = ? then null else pg_catalog.array_to_string(array ( select pg_catalog.quote_ident(rolname) from pg_catalog.pg_roles where oid = any (pol.polroles)), ?) end as polroles, pg_catalog.pg_get_expr(pol.polqual, pol.polrelid) as polqual, pg_catalog.pg_get_expr(pol.polwithcheck, pol.polrelid) as polwithcheck from pg_catalog.pg_policy pol where polrelid = ?;Times Reported Time consuming queries #13
Day Hour Count Duration Avg duration Feb 10 06 975 0ms 0ms 14 0ms 0ms 0ms 2 0ms lock table public.timezones_turkish in access share mode;Times Reported Time consuming queries #14
Day Hour Count Duration Avg duration Feb 10 06 2 0ms 0ms 15 0ms 0ms 0ms 2 0ms copy public.timezones_greek (timezoneid, timezone) to stdout;Times Reported Time consuming queries #15
Day Hour Count Duration Avg duration Feb 10 06 2 0ms 0ms 16 0ms 0ms 0ms 2 0ms lock table public.futures_symbols in access share mode;Times Reported Time consuming queries #16
Day Hour Count Duration Avg duration Feb 10 06 2 0ms 0ms 17 0ms 0ms 0ms 2 0ms copy archive.autochartist_results (resultid, bandwidth, pattern, gmttimefound, direction, initialtrend, breakout, volumeincrease, noise, symmetry, patternstarttime, patternendtime, patternstartprice, patternendprice, resx0, resx1, supportx0, supportx1, resy0, resy1, supporty0, supporty1, supportgradient, resgradient, riskreward, qtytp, predictionpricefrom, predictionpriceto, predictiontimefrom, predictiontimeto, patternquality, trendchange, maxmovementafterbreakout, latestbaratbreakouttime, latestbaratbreakoutprice, patternlengthbars, temporarypattern, symbolid, resultuid, relevancestartdistance, simulation, writtendatetime, old_resultuid) to stdout;Times Reported Time consuming queries #17
Day Hour Count Duration Avg duration Feb 10 06 2 0ms 0ms 18 0ms 0ms 0ms 2 0ms copy public.satelliteauthentication (id, satelliteserverid, brokerid, authenticationstatus, maketradeurl, notifications, validusername, validpassword, enabled) to stdout;Times Reported Time consuming queries #18
Day Hour Count Duration Avg duration Feb 10 06 2 0ms 0ms 19 0ms 0ms 0ms 2 0ms lock table archive.patternresultsage in access share mode;Times Reported Time consuming queries #19
Day Hour Count Duration Avg duration Feb 10 06 2 0ms 0ms 20 0ms 0ms 0ms 1 0ms copy public.processes (id, locale, region, schedule, enabled, live, lastmodified, lastrun, contenttypeid, brokerid, uuid) to stdout;Times Reported Time consuming queries #20
Day Hour Count Duration Avg duration Feb 10 06 1 0ms 0ms Time consuming prepare
Rank Total duration Times executed Min duration Max duration Avg duration Query 1 40s286ms 13 0ms 40s283ms 3s98ms select recognitionengine, to_char(datetimeupdate, 'yyyy-mm-dd HH24:MI') from latest_candle_datetime_per_receng;Times Reported Time consuming prepare #1
Day Hour Count Duration Avg duration Feb 10 06 13 40s286ms 3s98ms -
select recognitionengine, to_char(datetimeupdate, 'yyyy-mm-dd HH24:MI') from latest_candle_datetime_per_receng;
Date: 2026-02-10 06:19:55 Duration: 40s283ms Database: postgres
-
select recognitionengine, to_char(datetimeupdate, 'yyyy-mm-dd HH24:MI') from latest_candle_datetime_per_receng;
Date: 2026-02-10 06:15:05 Duration: 0ms Database: postgres
-
select recognitionengine, to_char(datetimeupdate, 'yyyy-mm-dd HH24:MI') from latest_candle_datetime_per_receng;
Date: 2026-02-10 06:10:04 Duration: 0ms Database: postgres
2 4s880ms 4,400 0ms 11ms 1ms WITH rar_max as ( ;Times Reported Time consuming prepare #2
Day Hour Count Duration Avg duration 06 4,400 4s880ms 1ms -
WITH rar_max as ( ;
Date: 2026-02-10 06:15:57 Duration: 11ms Database: postgres
-
WITH rar_max as ( ;
Date: 2026-02-10 06:16:33 Duration: 10ms Database: postgres
-
WITH rar_max as ( ;
Date: 2026-02-10 06:01:37 Duration: 9ms Database: postgres
3 1s775ms 4,615 0ms 17ms 0ms SELECT ;Times Reported Time consuming prepare #3
Day Hour Count Duration Avg duration 06 4,615 1s775ms 0ms -
SELECT ;
Date: 2026-02-10 06:42:08 Duration: 17ms Database: postgres
-
SELECT ;
Date: 2026-02-10 06:16:04 Duration: 14ms Database: postgres
-
SELECT ;
Date: 2026-02-10 06:15:54 Duration: 5ms Database: postgres
4 1s620ms 1,093 0ms 8ms 1ms SELECT symbolid, ;Times Reported Time consuming prepare #4
Day Hour Count Duration Avg duration 06 1,093 1s620ms 1ms -
SELECT symbolid, ;
Date: 2026-02-10 06:31:10 Duration: 8ms Database: postgres
-
SELECT symbolid, ;
Date: 2026-02-10 06:15:04 Duration: 4ms Database: postgres
-
SELECT symbolid, ;
Date: 2026-02-10 06:30:42 Duration: 3ms Database: postgres
5 499ms 417 0ms 6ms 1ms SELECT s.symbolid, dss.downloadfrequency, dss.downloadersymbol;Times Reported Time consuming prepare #5
Day Hour Count Duration Avg duration 06 417 499ms 1ms -
SELECT s.symbolid, dss.downloadfrequency, dss.downloadersymbol;
Date: 2026-02-10 06:16:00 Duration: 6ms Database: postgres
-
SELECT s.symbolid, dss.downloadfrequency, dss.downloadersymbol;
Date: 2026-02-10 06:30:54 Duration: 4ms Database: postgres
-
SELECT s.symbolid, dss.downloadfrequency, dss.downloadersymbol;
Date: 2026-02-10 06:32:33 Duration: 2ms Database: postgres
6 380ms 2,444 0ms 2ms 0ms SET extra_float_digits = 3;Times Reported Time consuming prepare #6
Day Hour Count Duration Avg duration 06 2,444 380ms 0ms -
SET extra_float_digits = 3;
Date: 2026-02-10 06:16:08 Duration: 2ms Database: postgres
-
SET extra_float_digits = 3;
Date: 2026-02-10 06:45:59 Duration: 2ms Database: postgres
-
SET extra_float_digits = 3;
Date: 2026-02-10 06:46:39 Duration: 2ms Database: postgres
7 272ms 2,851 0ms 2ms 0ms INSERT INTO T30 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;Times Reported Time consuming prepare #7
Day Hour Count Duration Avg duration 06 2,851 272ms 0ms -
INSERT INTO T30 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;
Date: 2026-02-10 06:32:07 Duration: 2ms Database: postgres
-
INSERT INTO T30 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;
Date: 2026-02-10 06:01:02 Duration: 0ms Database: postgres
-
INSERT INTO T30 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;
Date: 2026-02-10 06:11:37 Duration: 0ms Database: postgres
8 217ms 2,016 0ms 0ms 0ms INSERT INTO T60 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;Times Reported Time consuming prepare #8
Day Hour Count Duration Avg duration 06 2,016 217ms 0ms -
INSERT INTO T60 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;
Date: 2026-02-10 06:00:51 Duration: 0ms Database: postgres
-
INSERT INTO T60 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;
Date: 2026-02-10 06:01:02 Duration: 0ms Database: postgres
-
INSERT INTO T60 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;
Date: 2026-02-10 06:10:53 Duration: 0ms Database: postgres
9 161ms 1,012 0ms 0ms 0ms INSERT INTO T15 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;Times Reported Time consuming prepare #9
Day Hour Count Duration Avg duration 06 1,012 161ms 0ms -
INSERT INTO T15 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;
Date: 2026-02-10 06:41:52 Duration: 0ms Database: postgres
-
INSERT INTO T15 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;
Date: 2026-02-10 06:15:16 Duration: 0ms Database: postgres
-
INSERT INTO T15 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;
Date: 2026-02-10 06:45:55 Duration: 0ms Database: postgres
10 159ms 3,336 0ms 5ms 0ms select 1;Times Reported Time consuming prepare #10
Day Hour Count Duration Avg duration 06 3,336 159ms 0ms -
select 1;
Date: 2026-02-10 06:34:24 Duration: 5ms Database: postgres
-
select 1;
Date: 2026-02-10 06:15:06 Duration: 2ms Database: postgres
-
select 1;
Date: 2026-02-10 06:51:17 Duration: 1ms Database: postgres
11 93ms 52 0ms 10ms 1ms WITH last_candle AS ( ;Times Reported Time consuming prepare #11
Day Hour Count Duration Avg duration 06 52 93ms 1ms -
WITH last_candle AS ( ;
Date: 2026-02-10 06:36:00 Duration: 10ms Database: postgres
-
WITH last_candle AS ( ;
Date: 2026-02-10 06:24:00 Duration: 6ms Database: postgres
-
WITH last_candle AS ( ;
Date: 2026-02-10 06:24:00 Duration: 6ms Database: postgres
12 86ms 12 4ms 14ms 7ms with sym_info as ( ;Times Reported Time consuming prepare #12
Day Hour Count Duration Avg duration 06 12 86ms 7ms -
with sym_info as ( ;
Date: 2026-02-10 06:36:51 Duration: 14ms Database: postgres
-
with sym_info as ( ;
Date: 2026-02-10 06:36:41 Duration: 8ms Database: postgres
-
with sym_info as ( ;
Date: 2026-02-10 06:06:42 Duration: 7ms Database: postgres
13 64ms 300 0ms 0ms 0ms SELECT NULL AS TABLE_CAT, n.nspname AS TABLE_SCHEM, c.relname AS TABLE_NAME, CASE n.nspname ~ '^pg_' OR n.nspname = 'information_schema' WHEN true THEN CASE WHEN n.nspname = 'pg_catalog' OR n.nspname = 'information_schema' THEN CASE c.relkind WHEN 'r' THEN 'SYSTEM TABLE' WHEN 'v' THEN 'SYSTEM VIEW' WHEN 'i' THEN 'SYSTEM INDEX' ELSE NULL END WHEN n.nspname = 'pg_toast' THEN CASE c.relkind WHEN 'r' THEN 'SYSTEM TOAST TABLE' WHEN 'i' THEN 'SYSTEM TOAST INDEX' ELSE NULL END ELSE CASE c.relkind WHEN 'r' THEN 'TEMPORARY TABLE' WHEN 'p' THEN 'TEMPORARY TABLE' WHEN 'i' THEN 'TEMPORARY INDEX' WHEN 'S' THEN 'TEMPORARY SEQUENCE' WHEN 'v' THEN 'TEMPORARY VIEW' ELSE NULL END END WHEN false THEN CASE c.relkind WHEN 'r' THEN 'TABLE' WHEN 'p' THEN 'PARTITIONED TABLE' WHEN 'i' THEN 'INDEX' WHEN 'S' THEN 'SEQUENCE' WHEN 'v' THEN 'VIEW' WHEN 'c' THEN 'TYPE' WHEN 'f' THEN 'FOREIGN TABLE' WHEN 'm' THEN 'MATERIALIZED VIEW' ELSE NULL END ELSE NULL END AS TABLE_TYPE, d.description AS REMARKS, '' as TYPE_CAT, '' as TYPE_SCHEM, '' as TYPE_NAME, '' AS SELF_REFERENCING_COL_NAME, '' AS REF_GENERATION FROM pg_catalog.pg_namespace n, pg_catalog.pg_class c LEFT JOIN pg_catalog.pg_description d ON (c.oid = d.objoid AND d.objsubid = 0) LEFT JOIN pg_catalog.pg_class dc ON (d.classoid = dc.oid AND dc.relname = 'pg_class') LEFT JOIN pg_catalog.pg_namespace dn ON (dn.oid = dc.relnamespace AND dn.nspname = 'pg_catalog') WHERE c.relnamespace = n.oid AND c.relname LIKE 'PROBABLYNOT' AND (false OR (c.relkind = 'r' AND n.nspname !~ '^pg_' AND n.nspname <> 'information_schema')) ORDER BY TABLE_TYPE, TABLE_SCHEM, TABLE_NAME;Times Reported Time consuming prepare #13
Day Hour Count Duration Avg duration 06 300 64ms 0ms -
SELECT NULL AS TABLE_CAT, n.nspname AS TABLE_SCHEM, c.relname AS TABLE_NAME, CASE n.nspname ~ '^pg_' OR n.nspname = 'information_schema' WHEN true THEN CASE WHEN n.nspname = 'pg_catalog' OR n.nspname = 'information_schema' THEN CASE c.relkind WHEN 'r' THEN 'SYSTEM TABLE' WHEN 'v' THEN 'SYSTEM VIEW' WHEN 'i' THEN 'SYSTEM INDEX' ELSE NULL END WHEN n.nspname = 'pg_toast' THEN CASE c.relkind WHEN 'r' THEN 'SYSTEM TOAST TABLE' WHEN 'i' THEN 'SYSTEM TOAST INDEX' ELSE NULL END ELSE CASE c.relkind WHEN 'r' THEN 'TEMPORARY TABLE' WHEN 'p' THEN 'TEMPORARY TABLE' WHEN 'i' THEN 'TEMPORARY INDEX' WHEN 'S' THEN 'TEMPORARY SEQUENCE' WHEN 'v' THEN 'TEMPORARY VIEW' ELSE NULL END END WHEN false THEN CASE c.relkind WHEN 'r' THEN 'TABLE' WHEN 'p' THEN 'PARTITIONED TABLE' WHEN 'i' THEN 'INDEX' WHEN 'S' THEN 'SEQUENCE' WHEN 'v' THEN 'VIEW' WHEN 'c' THEN 'TYPE' WHEN 'f' THEN 'FOREIGN TABLE' WHEN 'm' THEN 'MATERIALIZED VIEW' ELSE NULL END ELSE NULL END AS TABLE_TYPE, d.description AS REMARKS, '' as TYPE_CAT, '' as TYPE_SCHEM, '' as TYPE_NAME, '' AS SELF_REFERENCING_COL_NAME, '' AS REF_GENERATION FROM pg_catalog.pg_namespace n, pg_catalog.pg_class c LEFT JOIN pg_catalog.pg_description d ON (c.oid = d.objoid AND d.objsubid = 0) LEFT JOIN pg_catalog.pg_class dc ON (d.classoid = dc.oid AND dc.relname = 'pg_class') LEFT JOIN pg_catalog.pg_namespace dn ON (dn.oid = dc.relnamespace AND dn.nspname = 'pg_catalog') WHERE c.relnamespace = n.oid AND c.relname LIKE 'PROBABLYNOT' AND (false OR (c.relkind = 'r' AND n.nspname !~ '^pg_' AND n.nspname <> 'information_schema')) ORDER BY TABLE_TYPE, TABLE_SCHEM, TABLE_NAME;
Date: 2026-02-10 06:11:50 Duration: 0ms Database: postgres
-
SELECT NULL AS TABLE_CAT, n.nspname AS TABLE_SCHEM, c.relname AS TABLE_NAME, CASE n.nspname ~ '^pg_' OR n.nspname = 'information_schema' WHEN true THEN CASE WHEN n.nspname = 'pg_catalog' OR n.nspname = 'information_schema' THEN CASE c.relkind WHEN 'r' THEN 'SYSTEM TABLE' WHEN 'v' THEN 'SYSTEM VIEW' WHEN 'i' THEN 'SYSTEM INDEX' ELSE NULL END WHEN n.nspname = 'pg_toast' THEN CASE c.relkind WHEN 'r' THEN 'SYSTEM TOAST TABLE' WHEN 'i' THEN 'SYSTEM TOAST INDEX' ELSE NULL END ELSE CASE c.relkind WHEN 'r' THEN 'TEMPORARY TABLE' WHEN 'p' THEN 'TEMPORARY TABLE' WHEN 'i' THEN 'TEMPORARY INDEX' WHEN 'S' THEN 'TEMPORARY SEQUENCE' WHEN 'v' THEN 'TEMPORARY VIEW' ELSE NULL END END WHEN false THEN CASE c.relkind WHEN 'r' THEN 'TABLE' WHEN 'p' THEN 'PARTITIONED TABLE' WHEN 'i' THEN 'INDEX' WHEN 'S' THEN 'SEQUENCE' WHEN 'v' THEN 'VIEW' WHEN 'c' THEN 'TYPE' WHEN 'f' THEN 'FOREIGN TABLE' WHEN 'm' THEN 'MATERIALIZED VIEW' ELSE NULL END ELSE NULL END AS TABLE_TYPE, d.description AS REMARKS, '' as TYPE_CAT, '' as TYPE_SCHEM, '' as TYPE_NAME, '' AS SELF_REFERENCING_COL_NAME, '' AS REF_GENERATION FROM pg_catalog.pg_namespace n, pg_catalog.pg_class c LEFT JOIN pg_catalog.pg_description d ON (c.oid = d.objoid AND d.objsubid = 0) LEFT JOIN pg_catalog.pg_class dc ON (d.classoid = dc.oid AND dc.relname = 'pg_class') LEFT JOIN pg_catalog.pg_namespace dn ON (dn.oid = dc.relnamespace AND dn.nspname = 'pg_catalog') WHERE c.relnamespace = n.oid AND c.relname LIKE 'PROBABLYNOT' AND (false OR (c.relkind = 'r' AND n.nspname !~ '^pg_' AND n.nspname <> 'information_schema')) ORDER BY TABLE_TYPE, TABLE_SCHEM, TABLE_NAME;
Date: 2026-02-10 06:11:52 Duration: 0ms Database: postgres
-
SELECT NULL AS TABLE_CAT, n.nspname AS TABLE_SCHEM, c.relname AS TABLE_NAME, CASE n.nspname ~ '^pg_' OR n.nspname = 'information_schema' WHEN true THEN CASE WHEN n.nspname = 'pg_catalog' OR n.nspname = 'information_schema' THEN CASE c.relkind WHEN 'r' THEN 'SYSTEM TABLE' WHEN 'v' THEN 'SYSTEM VIEW' WHEN 'i' THEN 'SYSTEM INDEX' ELSE NULL END WHEN n.nspname = 'pg_toast' THEN CASE c.relkind WHEN 'r' THEN 'SYSTEM TOAST TABLE' WHEN 'i' THEN 'SYSTEM TOAST INDEX' ELSE NULL END ELSE CASE c.relkind WHEN 'r' THEN 'TEMPORARY TABLE' WHEN 'p' THEN 'TEMPORARY TABLE' WHEN 'i' THEN 'TEMPORARY INDEX' WHEN 'S' THEN 'TEMPORARY SEQUENCE' WHEN 'v' THEN 'TEMPORARY VIEW' ELSE NULL END END WHEN false THEN CASE c.relkind WHEN 'r' THEN 'TABLE' WHEN 'p' THEN 'PARTITIONED TABLE' WHEN 'i' THEN 'INDEX' WHEN 'S' THEN 'SEQUENCE' WHEN 'v' THEN 'VIEW' WHEN 'c' THEN 'TYPE' WHEN 'f' THEN 'FOREIGN TABLE' WHEN 'm' THEN 'MATERIALIZED VIEW' ELSE NULL END ELSE NULL END AS TABLE_TYPE, d.description AS REMARKS, '' as TYPE_CAT, '' as TYPE_SCHEM, '' as TYPE_NAME, '' AS SELF_REFERENCING_COL_NAME, '' AS REF_GENERATION FROM pg_catalog.pg_namespace n, pg_catalog.pg_class c LEFT JOIN pg_catalog.pg_description d ON (c.oid = d.objoid AND d.objsubid = 0) LEFT JOIN pg_catalog.pg_class dc ON (d.classoid = dc.oid AND dc.relname = 'pg_class') LEFT JOIN pg_catalog.pg_namespace dn ON (dn.oid = dc.relnamespace AND dn.nspname = 'pg_catalog') WHERE c.relnamespace = n.oid AND c.relname LIKE 'PROBABLYNOT' AND (false OR (c.relkind = 'r' AND n.nspname !~ '^pg_' AND n.nspname <> 'information_schema')) ORDER BY TABLE_TYPE, TABLE_SCHEM, TABLE_NAME;
Date: 2026-02-10 06:11:51 Duration: 0ms Database: postgres
14 59ms 18 1ms 9ms 3ms select cast(count(*) / cast(setting as numeric) * 100 as int) from pg_stat_activity, pg_settings WHERE name = 'max_connections' group by setting;Times Reported Time consuming prepare #14
Day Hour Count Duration Avg duration 06 18 59ms 3ms -
select cast(count(*) / cast(setting as numeric) * 100 as int) from pg_stat_activity, pg_settings WHERE name = 'max_connections' group by setting;
Date: 2026-02-10 06:40:03 Duration: 9ms Database: postgres
-
select cast(count(*) / cast(setting as numeric) * 100 as int) from pg_stat_activity, pg_settings WHERE name = 'max_connections' group by setting;
Date: 2026-02-10 06:30:03 Duration: 4ms Database: postgres
-
select cast(count(*) / cast(setting as numeric) * 100 as int) from pg_stat_activity, pg_settings WHERE name = 'max_connections' group by setting;
Date: 2026-02-10 06:31:10 Duration: 3ms Database: postgres
15 40ms 288 0ms 0ms 0ms INSERT INTO T1440_underlying (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;Times Reported Time consuming prepare #15
Day Hour Count Duration Avg duration 06 288 40ms 0ms -
INSERT INTO T1440_underlying (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;
Date: 2026-02-10 06:32:33 Duration: 0ms Database: postgres
-
INSERT INTO T1440_underlying (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;
Date: 2026-02-10 06:16:23 Duration: 0ms Database: postgres
-
INSERT INTO T1440_underlying (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;
Date: 2026-02-10 06:30:42 Duration: 0ms Database: postgres
16 39ms 349 0ms 0ms 0ms INSERT INTO T240 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;Times Reported Time consuming prepare #16
Day Hour Count Duration Avg duration 06 349 39ms 0ms -
INSERT INTO T240 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;
Date: 2026-02-10 06:30:30 Duration: 0ms Database: postgres
-
INSERT INTO T240 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;
Date: 2026-02-10 06:16:23 Duration: 0ms Database: postgres
-
INSERT INTO T240 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;
Date: 2026-02-10 06:02:24 Duration: 0ms Database: postgres
17 31ms 2,418 0ms 0ms 0ms SET application_name = 'PostgreSQL JDBC Driver';Times Reported Time consuming prepare #17
Day Hour Count Duration Avg duration 06 2,418 31ms 0ms -
SET application_name = 'PostgreSQL JDBC Driver';
Date: 2026-02-10 06:31:55 Duration: 0ms Database: postgres
-
SET application_name = 'PostgreSQL JDBC Driver';
Date: 2026-02-10 06:05:29 Duration: 0ms Database: postgres
-
SET application_name = 'PostgreSQL JDBC Driver';
Date: 2026-02-10 06:33:56 Duration: 0ms Database: postgres
18 18ms 6 2ms 3ms 3ms with rankedmt4 as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors ), last_feed_entry as ( select * from rankedmt4 where r = 1 ), ok_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where status = 'OK' ), earliest_entry_after_ok as ( select m.datafeedname, min(m.eventtimestamp) as eventtimestamp from mt4datafeederrors m left outer join ( select datafeedname, eventtimestamp from ok_entries where r = 1) oo on m.datafeedname = oo.datafeedname where m.eventtimestamp > coalesce(oo.eventtimestamp, '1900-01-01'::timestamp without time zone) group by m.datafeedname ), notified_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where notified is not null and notified <> '' ), broker as ( select *, row_number() over (partition by feedname order by brokerid) r from ( select distinct b.brokerid, b.name as brokername, dss.classname as feedname from downloadersymbolsettings dss inner join brokersymbollist bsl on dss.symbolid = bsl.symbolid inner join broker b on bsl.brokerid = b.brokerid where dss.enabled = 1) a ) select last.id, last.datafeedname, last.eventtimestamp, last.status, last.errordescription, last.serveraddress, last.username, note.notified, note.eventtimestamp, broker.brokername from last_feed_entry last inner join earliest_entry_after_ok after_ok on last.datafeedname = after_ok.datafeedname inner join broker on last.datafeedname = broker.feedname left outer join ok_entries ok on ok.datafeedname = last.datafeedname left outer join notified_entries note on note.datafeedname = last.datafeedname and note.r = 1 where (ok.r is null or ok.r = 1) and last.datafeedname not in ( select distinct datafeedname from last_feed_entry where status = 'OK') and extract(epoch from (last.eventtimestamp - after_ok.eventtimestamp)) > 60 * 60 and last.eventtimestamp > current_timestamp - interval '1 day' and (note.eventtimestamp is null or note.eventtimestamp < current_timestamp - interval '10 hours') and last.eventtimestamp > current_timestamp - interval '1 hour' and broker.r = 1;Times Reported Time consuming prepare #18
Day Hour Count Duration Avg duration 06 6 18ms 3ms -
with rankedmt4 as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors ), last_feed_entry as ( select * from rankedmt4 where r = 1 ), ok_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where status = 'OK' ), earliest_entry_after_ok as ( select m.datafeedname, min(m.eventtimestamp) as eventtimestamp from mt4datafeederrors m left outer join ( select datafeedname, eventtimestamp from ok_entries where r = 1) oo on m.datafeedname = oo.datafeedname where m.eventtimestamp > coalesce(oo.eventtimestamp, '1900-01-01'::timestamp without time zone) group by m.datafeedname ), notified_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where notified is not null and notified <> '' ), broker as ( select *, row_number() over (partition by feedname order by brokerid) r from ( select distinct b.brokerid, b.name as brokername, dss.classname as feedname from downloadersymbolsettings dss inner join brokersymbollist bsl on dss.symbolid = bsl.symbolid inner join broker b on bsl.brokerid = b.brokerid where dss.enabled = 1) a ) select last.id, last.datafeedname, last.eventtimestamp, last.status, last.errordescription, last.serveraddress, last.username, note.notified, note.eventtimestamp, broker.brokername from last_feed_entry last inner join earliest_entry_after_ok after_ok on last.datafeedname = after_ok.datafeedname inner join broker on last.datafeedname = broker.feedname left outer join ok_entries ok on ok.datafeedname = last.datafeedname left outer join notified_entries note on note.datafeedname = last.datafeedname and note.r = 1 where (ok.r is null or ok.r = 1) and last.datafeedname not in ( select distinct datafeedname from last_feed_entry where status = 'OK') and extract(epoch from (last.eventtimestamp - after_ok.eventtimestamp)) > 60 * 60 and last.eventtimestamp > current_timestamp - interval '1 day' and (note.eventtimestamp is null or note.eventtimestamp < current_timestamp - interval '10 hours') and last.eventtimestamp > current_timestamp - interval '1 hour' and broker.r = 1;
Date: 2026-02-10 06:20:02 Duration: 3ms Database: postgres
-
with rankedmt4 as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors ), last_feed_entry as ( select * from rankedmt4 where r = 1 ), ok_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where status = 'OK' ), earliest_entry_after_ok as ( select m.datafeedname, min(m.eventtimestamp) as eventtimestamp from mt4datafeederrors m left outer join ( select datafeedname, eventtimestamp from ok_entries where r = 1) oo on m.datafeedname = oo.datafeedname where m.eventtimestamp > coalesce(oo.eventtimestamp, '1900-01-01'::timestamp without time zone) group by m.datafeedname ), notified_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where notified is not null and notified <> '' ), broker as ( select *, row_number() over (partition by feedname order by brokerid) r from ( select distinct b.brokerid, b.name as brokername, dss.classname as feedname from downloadersymbolsettings dss inner join brokersymbollist bsl on dss.symbolid = bsl.symbolid inner join broker b on bsl.brokerid = b.brokerid where dss.enabled = 1) a ) select last.id, last.datafeedname, last.eventtimestamp, last.status, last.errordescription, last.serveraddress, last.username, note.notified, note.eventtimestamp, broker.brokername from last_feed_entry last inner join earliest_entry_after_ok after_ok on last.datafeedname = after_ok.datafeedname inner join broker on last.datafeedname = broker.feedname left outer join ok_entries ok on ok.datafeedname = last.datafeedname left outer join notified_entries note on note.datafeedname = last.datafeedname and note.r = 1 where (ok.r is null or ok.r = 1) and last.datafeedname not in ( select distinct datafeedname from last_feed_entry where status = 'OK') and extract(epoch from (last.eventtimestamp - after_ok.eventtimestamp)) > 60 * 60 and last.eventtimestamp > current_timestamp - interval '1 day' and (note.eventtimestamp is null or note.eventtimestamp < current_timestamp - interval '10 hours') and last.eventtimestamp > current_timestamp - interval '1 hour' and broker.r = 1;
Date: 2026-02-10 06:50:02 Duration: 3ms Database: postgres
-
with rankedmt4 as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors ), last_feed_entry as ( select * from rankedmt4 where r = 1 ), ok_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where status = 'OK' ), earliest_entry_after_ok as ( select m.datafeedname, min(m.eventtimestamp) as eventtimestamp from mt4datafeederrors m left outer join ( select datafeedname, eventtimestamp from ok_entries where r = 1) oo on m.datafeedname = oo.datafeedname where m.eventtimestamp > coalesce(oo.eventtimestamp, '1900-01-01'::timestamp without time zone) group by m.datafeedname ), notified_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where notified is not null and notified <> '' ), broker as ( select *, row_number() over (partition by feedname order by brokerid) r from ( select distinct b.brokerid, b.name as brokername, dss.classname as feedname from downloadersymbolsettings dss inner join brokersymbollist bsl on dss.symbolid = bsl.symbolid inner join broker b on bsl.brokerid = b.brokerid where dss.enabled = 1) a ) select last.id, last.datafeedname, last.eventtimestamp, last.status, last.errordescription, last.serveraddress, last.username, note.notified, note.eventtimestamp, broker.brokername from last_feed_entry last inner join earliest_entry_after_ok after_ok on last.datafeedname = after_ok.datafeedname inner join broker on last.datafeedname = broker.feedname left outer join ok_entries ok on ok.datafeedname = last.datafeedname left outer join notified_entries note on note.datafeedname = last.datafeedname and note.r = 1 where (ok.r is null or ok.r = 1) and last.datafeedname not in ( select distinct datafeedname from last_feed_entry where status = 'OK') and extract(epoch from (last.eventtimestamp - after_ok.eventtimestamp)) > 60 * 60 and last.eventtimestamp > current_timestamp - interval '1 day' and (note.eventtimestamp is null or note.eventtimestamp < current_timestamp - interval '10 hours') and last.eventtimestamp > current_timestamp - interval '1 hour' and broker.r = 1;
Date: 2026-02-10 06:00:02 Duration: 3ms Database: postgres
19 17ms 6 2ms 3ms 2ms select client_addr, count(1) from pg_stat_activity, pg_settings WHERE name = 'max_connections' group by client_addr, setting having (client_addr is not null OR (client_addr is null and count(1) > (cast(setting as numeric) / 3 * 2))) order by count desc;Times Reported Time consuming prepare #19
Day Hour Count Duration Avg duration 06 6 17ms 2ms -
select client_addr, count(1) from pg_stat_activity, pg_settings WHERE name = 'max_connections' group by client_addr, setting having (client_addr is not null OR (client_addr is null and count(1) > (cast(setting as numeric) / 3 * 2))) order by count desc;
Date: 2026-02-10 06:40:05 Duration: 3ms Database: postgres
-
select client_addr, count(1) from pg_stat_activity, pg_settings WHERE name = 'max_connections' group by client_addr, setting having (client_addr is not null OR (client_addr is null and count(1) > (cast(setting as numeric) / 3 * 2))) order by count desc;
Date: 2026-02-10 06:00:05 Duration: 3ms Database: postgres
-
select client_addr, count(1) from pg_stat_activity, pg_settings WHERE name = 'max_connections' group by client_addr, setting having (client_addr is not null OR (client_addr is null and count(1) > (cast(setting as numeric) / 3 * 2))) order by count desc;
Date: 2026-02-10 06:30:05 Duration: 3ms Database: postgres
20 17ms 4 0ms 13ms 4ms select distinct replace(recognitionengine, ' ', '_') || '%' || replace(datetimeupdate::text, ' ', '_') as t from latest_candle_datetime_per_receng order by t;Times Reported Time consuming prepare #20
Day Hour Count Duration Avg duration 06 4 17ms 4ms -
select distinct replace(recognitionengine, ' ', '_') || '%' || replace(datetimeupdate::text, ' ', '_') as t from latest_candle_datetime_per_receng order by t;
Date: 2026-02-10 06:40:22 Duration: 13ms Database: postgres
-
select distinct replace(recognitionengine, ' ', '_') || '%' || replace(datetimeupdate::text, ' ', '_') as t from latest_candle_datetime_per_receng order by t;
Date: 2026-02-10 06:10:23 Duration: 1ms Database: postgres
-
select distinct replace(recognitionengine, ' ', '_') || '%' || replace(datetimeupdate::text, ' ', '_') as t from latest_candle_datetime_per_receng order by t;
Date: 2026-02-10 06:10:23 Duration: 1ms Database: postgres
Time consuming bind
Rank Total duration Times executed Min duration Max duration Avg duration Query 1 49s3ms 6,929 0ms 72ms 7ms WITH rar_max as ( ;Times Reported Time consuming bind #1
Day Hour Count Duration Avg duration Feb 10 06 6,929 49s3ms 7ms -
WITH rar_max as ( ;
Date: 2026-02-10 06:20:48 Duration: 72ms Database: postgres parameters: $1 = 't', $2 = '538', $3 = '7', $4 = '15', $5 = '30', $6 = '60', $7 = '120', $8 = '240', $9 = '480', $10 = '1440', $11 = '0', $12 = '', $13 = '11', $14 = 'EURUSD', $15 = 'USDJPY', $16 = 'GBPUSD', $17 = 'GBPJPY', $18 = 'AUDUSD', $19 = 'EURJPY', $20 = 'USDCAD', $21 = 'EURGBP', $22 = 'XAUUSD', $23 = 'UKOIL', $24 = 'USOIL', $25 = '0', $26 = '', $27 = '0', $28 = '0', $29 = '0', $30 = '500', $31 = '500', $32 = 't', $33 = '10', $34 = '10'
-
WITH rar_max as ( ;
Date: 2026-02-10 06:36:41 Duration: 59ms Database: postgres parameters: $1 = 't', $2 = '489', $3 = '0', $4 = '0', $5 = '0', $6 = '', $7 = '0', $8 = '', $9 = '0', $10 = '', $11 = '0', $12 = '0', $13 = '0', $14 = '0', $15 = '0', $16 = 't', $17 = '0', $18 = '0'
-
WITH rar_max as ( ;
Date: 2026-02-10 06:36:48 Duration: 54ms Database: postgres parameters: $1 = '489', $2 = '0', $3 = '0', $4 = '0', $5 = '', $6 = '0', $7 = '', $8 = '0', $9 = '0', $10 = 't', $11 = '0', $12 = '0'
2 10s544ms 27,427 0ms 20ms 0ms SELECT ;Times Reported Time consuming bind #2
Day Hour Count Duration Avg duration 06 27,427 10s544ms 0ms -
SELECT ;
Date: 2026-02-10 06:45:04 Duration: 20ms Database: postgres parameters: $1 = '515840233927091300'
-
SELECT ;
Date: 2026-02-10 06:17:01 Duration: 18ms Database: postgres parameters: $1 = '558', $2 = '558', $3 = '515840233917023300'
-
SELECT ;
Date: 2026-02-10 06:29:49 Duration: 17ms Database: postgres parameters: $1 = '515840233916163300'
3 2s850ms 1,093 1ms 8ms 2ms SELECT symbolid, ;Times Reported Time consuming bind #3
Day Hour Count Duration Avg duration 06 1,093 2s850ms 2ms -
SELECT symbolid, ;
Date: 2026-02-10 06:16:21 Duration: 8ms Database: postgres parameters: $1 = 'PEPPERSTONEMT5', $2 = '15', $3 = 'AMD.US', $4 = 'ABNB.US', $5 = 'AAPL.US', $6 = 'AXP.US', $7 = 'AMZN.US'
-
SELECT symbolid, ;
Date: 2026-02-10 06:17:23 Duration: 6ms Database: postgres parameters: $1 = 'GLOBALGTMT5', $2 = '15', $3 = 'ADAUSD', $4 = '#AAPL', $5 = '#AMZN', $6 = 'AUDJPY', $7 = 'BCHJPY', $8 = 'AUDUSD'
-
SELECT symbolid, ;
Date: 2026-02-10 06:32:23 Duration: 5ms Database: postgres parameters: $1 = 'GLOBALGTMT5', $2 = '15', $3 = '#AAPL', $4 = '#AMZN'
4 947ms 105 4ms 34ms 9ms WITH last_candle AS ( ;Times Reported Time consuming bind #4
Day Hour Count Duration Avg duration 06 105 947ms 9ms -
WITH last_candle AS ( ;
Date: 2026-02-10 06:36:02 Duration: 34ms Database: postgres parameters: $1 = '667', $2 = '667'
-
WITH last_candle AS ( ;
Date: 2026-02-10 06:16:01 Duration: 28ms Database: postgres parameters: $1 = '558', $2 = '558'
-
WITH last_candle AS ( ;
Date: 2026-02-10 06:36:00 Duration: 27ms Database: postgres parameters: $1 = '558', $2 = '558'
5 858ms 28,342 0ms 13ms 0ms select 1;Times Reported Time consuming bind #5
Day Hour Count Duration Avg duration 06 28,342 858ms 0ms -
select 1;
Date: 2026-02-10 06:21:15 Duration: 13ms Database: postgres
-
select 1;
Date: 2026-02-10 06:21:22 Duration: 13ms Database: postgres
-
select 1;
Date: 2026-02-10 06:20:44 Duration: 10ms Database: postgres
6 788ms 417 1ms 8ms 1ms SELECT s.symbolid, dss.downloadfrequency, dss.downloadersymbol;Times Reported Time consuming bind #6
Day Hour Count Duration Avg duration 06 417 788ms 1ms -
SELECT s.symbolid, dss.downloadfrequency, dss.downloadersymbol;
Date: 2026-02-10 06:15:31 Duration: 8ms Database: postgres parameters: $1 = 'ICMARKETS'
-
SELECT s.symbolid, dss.downloadfrequency, dss.downloadersymbol;
Date: 2026-02-10 06:16:00 Duration: 6ms Database: postgres parameters: $1 = 'FPMARKETS'
-
SELECT s.symbolid, dss.downloadfrequency, dss.downloadersymbol;
Date: 2026-02-10 06:16:53 Duration: 4ms Database: postgres parameters: $1 = 'BDSWISS'
7 636ms 23 0ms 71ms 27ms with wh_patitioned as ( ;Times Reported Time consuming bind #7
Day Hour Count Duration Avg duration 06 23 636ms 27ms -
with wh_patitioned as ( ;
Date: 2026-02-10 06:40:01 Duration: 71ms Database: postgres parameters: $1 = '558', $2 = '558', $3 = '558', $4 = '558', $5 = '558', $6 = '558', $7 = '558', $8 = '558', $9 = '558'
-
with wh_patitioned as ( ;
Date: 2026-02-10 06:05:43 Duration: 50ms Database: postgres parameters: $1 = '621', $2 = '621', $3 = '621', $4 = '621', $5 = '621', $6 = '621', $7 = '621', $8 = '621', $9 = '621'
-
with wh_patitioned as ( ;
Date: 2026-02-10 06:29:49 Duration: 49ms Database: postgres parameters: $1 = '558', $2 = '558', $3 = '558', $4 = '558', $5 = '558', $6 = '558', $7 = '558', $8 = '558', $9 = '558'
8 534ms 12 28ms 60ms 44ms with sym_info as ( ;Times Reported Time consuming bind #8
Day Hour Count Duration Avg duration 06 12 534ms 44ms -
with sym_info as ( ;
Date: 2026-02-10 06:36:41 Duration: 60ms Database: postgres parameters: $1 = '620', $2 = 'Forex', $3 = 'Forex', $4 = '620', $5 = 'Forex', $6 = '620', $7 = '620', $8 = 'Forex', $9 = '620'
-
with sym_info as ( ;
Date: 2026-02-10 06:51:46 Duration: 54ms Database: postgres parameters: $1 = '620', $2 = 'Forex', $3 = 'Forex', $4 = '620', $5 = 'Forex', $6 = '620', $7 = '620', $8 = 'Forex', $9 = '620'
-
with sym_info as ( ;
Date: 2026-02-10 06:06:42 Duration: 54ms Database: postgres parameters: $1 = '620', $2 = 'Forex', $3 = 'Forex', $4 = '620', $5 = 'Forex', $6 = '620', $7 = '620', $8 = 'Forex', $9 = '620'
9 402ms 57 0ms 19ms 7ms WITH /*Latest.JapSticks*/ all_results AS ( SELECT ;Times Reported Time consuming bind #9
Day Hour Count Duration Avg duration 06 57 402ms 7ms -
WITH /*Latest.JapSticks*/ all_results AS ( SELECT ;
Date: 2026-02-10 06:21:26 Duration: 19ms Database: postgres parameters: $1 = '689', $2 = '0', $3 = '0', $4 = '0', $5 = '', $6 = '0', $7 = '', $8 = '0', $9 = '', $10 = '0', $11 = '0'
-
WITH /*Latest.JapSticks*/ all_results AS ( SELECT ;
Date: 2026-02-10 06:12:47 Duration: 18ms Database: postgres parameters: $1 = '489', $2 = '0', $3 = '0', $4 = '0', $5 = '', $6 = '0', $7 = '', $8 = '0', $9 = '', $10 = '0', $11 = '0'
-
WITH /*Latest.JapSticks*/ all_results AS ( SELECT ;
Date: 2026-02-10 06:06:49 Duration: 18ms Database: postgres parameters: $1 = '689', $2 = '0', $3 = '0', $4 = '0', $5 = '', $6 = '0', $7 = '', $8 = '0', $9 = '', $10 = '0', $11 = '0'
10 257ms 5,001 0ms 1ms 0ms INSERT INTO T15 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;Times Reported Time consuming bind #10
Day Hour Count Duration Avg duration 06 5,001 257ms 0ms -
INSERT INTO T15 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;
Date: 2026-02-10 06:46:52 Duration: 1ms Database: postgres parameters: $1 = '2026-02-10 07:30:00', $2 = '155.2875', $3 = '155.3285', $4 = '155.2185', $5 = '155.2235', $6 = '744', $7 = '515840249375237300', $8 = '0', $9 = '2026-02-10 06:46:52.133', $10 = '2026-02-10 06:46:52.057', $11 = '155.2875', $12 = '155.3285', $13 = '155.2185', $14 = '155.2235', $15 = '744', $16 = '0', $17 = '2026-02-10 06:46:52.134', $18 = '2026-02-10 06:46:52.057'
-
INSERT INTO T15 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;
Date: 2026-02-10 06:15:43 Duration: 0ms Database: postgres parameters: $1 = '2026-02-10 06:00:00', $2 = '81.42', $3 = '81.493', $4 = '80.963', $5 = '81.249', $6 = '4008', $7 = '515840245849884300', $8 = '0', $9 = '2026-02-10 06:15:43.269', $10 = '2026-02-10 06:15:43.031', $11 = '81.42', $12 = '81.493', $13 = '80.963', $14 = '81.249', $15 = '4008', $16 = '0', $17 = '2026-02-10 06:15:43.269', $18 = '2026-02-10 06:15:43.031'
-
INSERT INTO T15 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;
Date: 2026-02-10 06:32:23 Duration: 0ms Database: postgres parameters: $1 = '2026-02-09 23:45:00', $2 = '209.405', $3 = '209.525', $4 = '208.66', $5 = '208.685', $6 = '421', $7 = '515840249387066300', $8 = '0', $9 = '2026-02-10 06:32:23.113', $10 = '2026-02-10 06:32:23.051', $11 = '209.405', $12 = '209.525', $13 = '208.66', $14 = '208.685', $15 = '421', $16 = '0', $17 = '2026-02-10 06:32:23.113', $18 = '2026-02-10 06:32:23.051'
11 249ms 2,991 0ms 1ms 0ms INSERT INTO T30 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;Times Reported Time consuming bind #11
Day Hour Count Duration Avg duration 06 2,991 249ms 0ms -
INSERT INTO T30 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;
Date: 2026-02-10 06:06:20 Duration: 1ms Database: postgres parameters: $1 = '2026-02-10 05:30:00', $2 = '159.24', $3 = '159.58', $4 = '158.845', $5 = '158.845', $6 = '1013', $7 = '606715250665191300', $8 = '0', $9 = '2026-02-10 06:06:20.736', $10 = '2026-02-10 06:06:20.632', $11 = '159.24', $12 = '159.58', $13 = '158.845', $14 = '158.845', $15 = '1013', $16 = '0', $17 = '2026-02-10 06:06:20.736', $18 = '2026-02-10 06:06:20.632'
-
INSERT INTO T30 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;
Date: 2026-02-10 06:30:58 Duration: 0ms Database: postgres parameters: $1 = '2026-02-10 06:00:00', $2 = '20.2968', $3 = '20.3039', $4 = '20.2916', $5 = '20.3014', $6 = '2467', $7 = '515840233485100300', $8 = '0', $9 = '2026-02-10 06:30:58.859', $10 = '2026-02-10 06:30:58.858', $11 = '20.2968', $12 = '20.3039', $13 = '20.2916', $14 = '20.3014', $15 = '2467', $16 = '0', $17 = '2026-02-10 06:30:58.859', $18 = '2026-02-10 06:30:58.858'
-
INSERT INTO T30 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;
Date: 2026-02-10 06:32:49 Duration: 0ms Database: postgres parameters: $1 = '2026-02-10 05:00:00', $2 = '27155.2', $3 = '27202.7', $4 = '27136.2', $5 = '27150.4', $6 = '9335', $7 = '515840247933633300', $8 = '0', $9 = '2026-02-10 06:32:49.99', $10 = '2026-02-10 06:32:49.92', $11 = '27155.2', $12 = '27202.7', $13 = '27136.2', $14 = '27150.4', $15 = '9335', $16 = '0', $17 = '2026-02-10 06:32:49.99', $18 = '2026-02-10 06:32:49.92'
12 188ms 2,172 0ms 0ms 0ms INSERT INTO T60 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;Times Reported Time consuming bind #12
Day Hour Count Duration Avg duration 06 2,172 188ms 0ms -
INSERT INTO T60 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;
Date: 2026-02-10 06:16:25 Duration: 0ms Database: postgres parameters: $1 = '2026-02-09 22:00:00', $2 = '95.09', $3 = '95.5', $4 = '94.84', $5 = '95.395', $6 = '963', $7 = '515840249458435300', $8 = '0', $9 = '2026-02-10 06:16:25.469', $10 = '2026-02-10 06:16:25.469', $11 = '95.09', $12 = '95.5', $13 = '94.84', $14 = '95.395', $15 = '963', $16 = '0', $17 = '2026-02-10 06:16:25.469', $18 = '2026-02-10 06:16:25.469'
-
INSERT INTO T60 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;
Date: 2026-02-10 06:00:30 Duration: 0ms Database: postgres parameters: $1 = '2026-02-10 05:00:00', $2 = '1.19129', $3 = '1.19163', $4 = '1.19092', $5 = '1.19106', $6 = '5868', $7 = '515840245852694300', $8 = '0', $9 = '2026-02-10 06:00:30.219', $10 = '2026-02-10 06:00:30.219', $11 = '1.19129', $12 = '1.19163', $13 = '1.19092', $14 = '1.19106', $15 = '5868', $16 = '0', $17 = '2026-02-10 06:00:30.219', $18 = '2026-02-10 06:00:30.219'
-
INSERT INTO T60 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;
Date: 2026-02-10 06:02:51 Duration: 0ms Database: postgres parameters: $1 = '2026-02-09 22:00:00', $2 = '194.09', $3 = '195.07', $4 = '192.035', $5 = '193.815', $6 = '1803', $7 = '515840249455331300', $8 = '0', $9 = '2026-02-10 06:02:51.275', $10 = '2026-02-10 06:02:51.275', $11 = '194.09', $12 = '195.07', $13 = '192.035', $14 = '193.815', $15 = '1803', $16 = '0', $17 = '2026-02-10 06:02:51.275', $18 = '2026-02-10 06:02:51.275'
13 162ms 300 0ms 2ms 0ms SELECT NULL AS TABLE_CAT, n.nspname AS TABLE_SCHEM, c.relname AS TABLE_NAME, CASE n.nspname ~ '^pg_' OR n.nspname = 'information_schema' WHEN true THEN CASE WHEN n.nspname = 'pg_catalog' OR n.nspname = 'information_schema' THEN CASE c.relkind WHEN 'r' THEN 'SYSTEM TABLE' WHEN 'v' THEN 'SYSTEM VIEW' WHEN 'i' THEN 'SYSTEM INDEX' ELSE NULL END WHEN n.nspname = 'pg_toast' THEN CASE c.relkind WHEN 'r' THEN 'SYSTEM TOAST TABLE' WHEN 'i' THEN 'SYSTEM TOAST INDEX' ELSE NULL END ELSE CASE c.relkind WHEN 'r' THEN 'TEMPORARY TABLE' WHEN 'p' THEN 'TEMPORARY TABLE' WHEN 'i' THEN 'TEMPORARY INDEX' WHEN 'S' THEN 'TEMPORARY SEQUENCE' WHEN 'v' THEN 'TEMPORARY VIEW' ELSE NULL END END WHEN false THEN CASE c.relkind WHEN 'r' THEN 'TABLE' WHEN 'p' THEN 'PARTITIONED TABLE' WHEN 'i' THEN 'INDEX' WHEN 'S' THEN 'SEQUENCE' WHEN 'v' THEN 'VIEW' WHEN 'c' THEN 'TYPE' WHEN 'f' THEN 'FOREIGN TABLE' WHEN 'm' THEN 'MATERIALIZED VIEW' ELSE NULL END ELSE NULL END AS TABLE_TYPE, d.description AS REMARKS, '' as TYPE_CAT, '' as TYPE_SCHEM, '' as TYPE_NAME, '' AS SELF_REFERENCING_COL_NAME, '' AS REF_GENERATION FROM pg_catalog.pg_namespace n, pg_catalog.pg_class c LEFT JOIN pg_catalog.pg_description d ON (c.oid = d.objoid AND d.objsubid = 0) LEFT JOIN pg_catalog.pg_class dc ON (d.classoid = dc.oid AND dc.relname = 'pg_class') LEFT JOIN pg_catalog.pg_namespace dn ON (dn.oid = dc.relnamespace AND dn.nspname = 'pg_catalog') WHERE c.relnamespace = n.oid AND c.relname LIKE 'PROBABLYNOT' AND (false OR (c.relkind = 'r' AND n.nspname !~ '^pg_' AND n.nspname <> 'information_schema')) ORDER BY TABLE_TYPE, TABLE_SCHEM, TABLE_NAME;Times Reported Time consuming bind #13
Day Hour Count Duration Avg duration 06 300 162ms 0ms -
SELECT NULL AS TABLE_CAT, n.nspname AS TABLE_SCHEM, c.relname AS TABLE_NAME, CASE n.nspname ~ '^pg_' OR n.nspname = 'information_schema' WHEN true THEN CASE WHEN n.nspname = 'pg_catalog' OR n.nspname = 'information_schema' THEN CASE c.relkind WHEN 'r' THEN 'SYSTEM TABLE' WHEN 'v' THEN 'SYSTEM VIEW' WHEN 'i' THEN 'SYSTEM INDEX' ELSE NULL END WHEN n.nspname = 'pg_toast' THEN CASE c.relkind WHEN 'r' THEN 'SYSTEM TOAST TABLE' WHEN 'i' THEN 'SYSTEM TOAST INDEX' ELSE NULL END ELSE CASE c.relkind WHEN 'r' THEN 'TEMPORARY TABLE' WHEN 'p' THEN 'TEMPORARY TABLE' WHEN 'i' THEN 'TEMPORARY INDEX' WHEN 'S' THEN 'TEMPORARY SEQUENCE' WHEN 'v' THEN 'TEMPORARY VIEW' ELSE NULL END END WHEN false THEN CASE c.relkind WHEN 'r' THEN 'TABLE' WHEN 'p' THEN 'PARTITIONED TABLE' WHEN 'i' THEN 'INDEX' WHEN 'S' THEN 'SEQUENCE' WHEN 'v' THEN 'VIEW' WHEN 'c' THEN 'TYPE' WHEN 'f' THEN 'FOREIGN TABLE' WHEN 'm' THEN 'MATERIALIZED VIEW' ELSE NULL END ELSE NULL END AS TABLE_TYPE, d.description AS REMARKS, '' as TYPE_CAT, '' as TYPE_SCHEM, '' as TYPE_NAME, '' AS SELF_REFERENCING_COL_NAME, '' AS REF_GENERATION FROM pg_catalog.pg_namespace n, pg_catalog.pg_class c LEFT JOIN pg_catalog.pg_description d ON (c.oid = d.objoid AND d.objsubid = 0) LEFT JOIN pg_catalog.pg_class dc ON (d.classoid = dc.oid AND dc.relname = 'pg_class') LEFT JOIN pg_catalog.pg_namespace dn ON (dn.oid = dc.relnamespace AND dn.nspname = 'pg_catalog') WHERE c.relnamespace = n.oid AND c.relname LIKE 'PROBABLYNOT' AND (false OR (c.relkind = 'r' AND n.nspname !~ '^pg_' AND n.nspname <> 'information_schema')) ORDER BY TABLE_TYPE, TABLE_SCHEM, TABLE_NAME;
Date: 2026-02-10 06:11:50 Duration: 2ms Database: postgres
-
SELECT NULL AS TABLE_CAT, n.nspname AS TABLE_SCHEM, c.relname AS TABLE_NAME, CASE n.nspname ~ '^pg_' OR n.nspname = 'information_schema' WHEN true THEN CASE WHEN n.nspname = 'pg_catalog' OR n.nspname = 'information_schema' THEN CASE c.relkind WHEN 'r' THEN 'SYSTEM TABLE' WHEN 'v' THEN 'SYSTEM VIEW' WHEN 'i' THEN 'SYSTEM INDEX' ELSE NULL END WHEN n.nspname = 'pg_toast' THEN CASE c.relkind WHEN 'r' THEN 'SYSTEM TOAST TABLE' WHEN 'i' THEN 'SYSTEM TOAST INDEX' ELSE NULL END ELSE CASE c.relkind WHEN 'r' THEN 'TEMPORARY TABLE' WHEN 'p' THEN 'TEMPORARY TABLE' WHEN 'i' THEN 'TEMPORARY INDEX' WHEN 'S' THEN 'TEMPORARY SEQUENCE' WHEN 'v' THEN 'TEMPORARY VIEW' ELSE NULL END END WHEN false THEN CASE c.relkind WHEN 'r' THEN 'TABLE' WHEN 'p' THEN 'PARTITIONED TABLE' WHEN 'i' THEN 'INDEX' WHEN 'S' THEN 'SEQUENCE' WHEN 'v' THEN 'VIEW' WHEN 'c' THEN 'TYPE' WHEN 'f' THEN 'FOREIGN TABLE' WHEN 'm' THEN 'MATERIALIZED VIEW' ELSE NULL END ELSE NULL END AS TABLE_TYPE, d.description AS REMARKS, '' as TYPE_CAT, '' as TYPE_SCHEM, '' as TYPE_NAME, '' AS SELF_REFERENCING_COL_NAME, '' AS REF_GENERATION FROM pg_catalog.pg_namespace n, pg_catalog.pg_class c LEFT JOIN pg_catalog.pg_description d ON (c.oid = d.objoid AND d.objsubid = 0) LEFT JOIN pg_catalog.pg_class dc ON (d.classoid = dc.oid AND dc.relname = 'pg_class') LEFT JOIN pg_catalog.pg_namespace dn ON (dn.oid = dc.relnamespace AND dn.nspname = 'pg_catalog') WHERE c.relnamespace = n.oid AND c.relname LIKE 'PROBABLYNOT' AND (false OR (c.relkind = 'r' AND n.nspname !~ '^pg_' AND n.nspname <> 'information_schema')) ORDER BY TABLE_TYPE, TABLE_SCHEM, TABLE_NAME;
Date: 2026-02-10 06:11:50 Duration: 1ms Database: postgres
-
SELECT NULL AS TABLE_CAT, n.nspname AS TABLE_SCHEM, c.relname AS TABLE_NAME, CASE n.nspname ~ '^pg_' OR n.nspname = 'information_schema' WHEN true THEN CASE WHEN n.nspname = 'pg_catalog' OR n.nspname = 'information_schema' THEN CASE c.relkind WHEN 'r' THEN 'SYSTEM TABLE' WHEN 'v' THEN 'SYSTEM VIEW' WHEN 'i' THEN 'SYSTEM INDEX' ELSE NULL END WHEN n.nspname = 'pg_toast' THEN CASE c.relkind WHEN 'r' THEN 'SYSTEM TOAST TABLE' WHEN 'i' THEN 'SYSTEM TOAST INDEX' ELSE NULL END ELSE CASE c.relkind WHEN 'r' THEN 'TEMPORARY TABLE' WHEN 'p' THEN 'TEMPORARY TABLE' WHEN 'i' THEN 'TEMPORARY INDEX' WHEN 'S' THEN 'TEMPORARY SEQUENCE' WHEN 'v' THEN 'TEMPORARY VIEW' ELSE NULL END END WHEN false THEN CASE c.relkind WHEN 'r' THEN 'TABLE' WHEN 'p' THEN 'PARTITIONED TABLE' WHEN 'i' THEN 'INDEX' WHEN 'S' THEN 'SEQUENCE' WHEN 'v' THEN 'VIEW' WHEN 'c' THEN 'TYPE' WHEN 'f' THEN 'FOREIGN TABLE' WHEN 'm' THEN 'MATERIALIZED VIEW' ELSE NULL END ELSE NULL END AS TABLE_TYPE, d.description AS REMARKS, '' as TYPE_CAT, '' as TYPE_SCHEM, '' as TYPE_NAME, '' AS SELF_REFERENCING_COL_NAME, '' AS REF_GENERATION FROM pg_catalog.pg_namespace n, pg_catalog.pg_class c LEFT JOIN pg_catalog.pg_description d ON (c.oid = d.objoid AND d.objsubid = 0) LEFT JOIN pg_catalog.pg_class dc ON (d.classoid = dc.oid AND dc.relname = 'pg_class') LEFT JOIN pg_catalog.pg_namespace dn ON (dn.oid = dc.relnamespace AND dn.nspname = 'pg_catalog') WHERE c.relnamespace = n.oid AND c.relname LIKE 'PROBABLYNOT' AND (false OR (c.relkind = 'r' AND n.nspname !~ '^pg_' AND n.nspname <> 'information_schema')) ORDER BY TABLE_TYPE, TABLE_SCHEM, TABLE_NAME;
Date: 2026-02-10 06:11:50 Duration: 1ms Database: postgres
14 92ms 6 5ms 50ms 15ms with rankedmt4 as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors ), last_feed_entry as ( select * from rankedmt4 where r = 1 ), ok_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where status = 'OK' ), earliest_entry_after_ok as ( select m.datafeedname, min(m.eventtimestamp) as eventtimestamp from mt4datafeederrors m left outer join ( select datafeedname, eventtimestamp from ok_entries where r = 1) oo on m.datafeedname = oo.datafeedname where m.eventtimestamp > coalesce(oo.eventtimestamp, '1900-01-01'::timestamp without time zone) group by m.datafeedname ), notified_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where notified is not null and notified <> '' ), broker as ( select *, row_number() over (partition by feedname order by brokerid) r from ( select distinct b.brokerid, b.name as brokername, dss.classname as feedname from downloadersymbolsettings dss inner join brokersymbollist bsl on dss.symbolid = bsl.symbolid inner join broker b on bsl.brokerid = b.brokerid where dss.enabled = 1) a ) select last.id, last.datafeedname, last.eventtimestamp, last.status, last.errordescription, last.serveraddress, last.username, note.notified, note.eventtimestamp, broker.brokername from last_feed_entry last inner join earliest_entry_after_ok after_ok on last.datafeedname = after_ok.datafeedname inner join broker on last.datafeedname = broker.feedname left outer join ok_entries ok on ok.datafeedname = last.datafeedname left outer join notified_entries note on note.datafeedname = last.datafeedname and note.r = 1 where (ok.r is null or ok.r = 1) and last.datafeedname not in ( select distinct datafeedname from last_feed_entry where status = 'OK') and extract(epoch from (last.eventtimestamp - after_ok.eventtimestamp)) > 60 * 60 and last.eventtimestamp > current_timestamp - interval '1 day' and (note.eventtimestamp is null or note.eventtimestamp < current_timestamp - interval '10 hours') and last.eventtimestamp > current_timestamp - interval '1 hour' and broker.r = 1;Times Reported Time consuming bind #14
Day Hour Count Duration Avg duration 06 6 92ms 15ms -
with rankedmt4 as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors ), last_feed_entry as ( select * from rankedmt4 where r = 1 ), ok_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where status = 'OK' ), earliest_entry_after_ok as ( select m.datafeedname, min(m.eventtimestamp) as eventtimestamp from mt4datafeederrors m left outer join ( select datafeedname, eventtimestamp from ok_entries where r = 1) oo on m.datafeedname = oo.datafeedname where m.eventtimestamp > coalesce(oo.eventtimestamp, '1900-01-01'::timestamp without time zone) group by m.datafeedname ), notified_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where notified is not null and notified <> '' ), broker as ( select *, row_number() over (partition by feedname order by brokerid) r from ( select distinct b.brokerid, b.name as brokername, dss.classname as feedname from downloadersymbolsettings dss inner join brokersymbollist bsl on dss.symbolid = bsl.symbolid inner join broker b on bsl.brokerid = b.brokerid where dss.enabled = 1) a ) select last.id, last.datafeedname, last.eventtimestamp, last.status, last.errordescription, last.serveraddress, last.username, note.notified, note.eventtimestamp, broker.brokername from last_feed_entry last inner join earliest_entry_after_ok after_ok on last.datafeedname = after_ok.datafeedname inner join broker on last.datafeedname = broker.feedname left outer join ok_entries ok on ok.datafeedname = last.datafeedname left outer join notified_entries note on note.datafeedname = last.datafeedname and note.r = 1 where (ok.r is null or ok.r = 1) and last.datafeedname not in ( select distinct datafeedname from last_feed_entry where status = 'OK') and extract(epoch from (last.eventtimestamp - after_ok.eventtimestamp)) > 60 * 60 and last.eventtimestamp > current_timestamp - interval '1 day' and (note.eventtimestamp is null or note.eventtimestamp < current_timestamp - interval '10 hours') and last.eventtimestamp > current_timestamp - interval '1 hour' and broker.r = 1;
Date: 2026-02-10 06:40:02 Duration: 50ms Database: postgres
-
with rankedmt4 as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors ), last_feed_entry as ( select * from rankedmt4 where r = 1 ), ok_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where status = 'OK' ), earliest_entry_after_ok as ( select m.datafeedname, min(m.eventtimestamp) as eventtimestamp from mt4datafeederrors m left outer join ( select datafeedname, eventtimestamp from ok_entries where r = 1) oo on m.datafeedname = oo.datafeedname where m.eventtimestamp > coalesce(oo.eventtimestamp, '1900-01-01'::timestamp without time zone) group by m.datafeedname ), notified_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where notified is not null and notified <> '' ), broker as ( select *, row_number() over (partition by feedname order by brokerid) r from ( select distinct b.brokerid, b.name as brokername, dss.classname as feedname from downloadersymbolsettings dss inner join brokersymbollist bsl on dss.symbolid = bsl.symbolid inner join broker b on bsl.brokerid = b.brokerid where dss.enabled = 1) a ) select last.id, last.datafeedname, last.eventtimestamp, last.status, last.errordescription, last.serveraddress, last.username, note.notified, note.eventtimestamp, broker.brokername from last_feed_entry last inner join earliest_entry_after_ok after_ok on last.datafeedname = after_ok.datafeedname inner join broker on last.datafeedname = broker.feedname left outer join ok_entries ok on ok.datafeedname = last.datafeedname left outer join notified_entries note on note.datafeedname = last.datafeedname and note.r = 1 where (ok.r is null or ok.r = 1) and last.datafeedname not in ( select distinct datafeedname from last_feed_entry where status = 'OK') and extract(epoch from (last.eventtimestamp - after_ok.eventtimestamp)) > 60 * 60 and last.eventtimestamp > current_timestamp - interval '1 day' and (note.eventtimestamp is null or note.eventtimestamp < current_timestamp - interval '10 hours') and last.eventtimestamp > current_timestamp - interval '1 hour' and broker.r = 1;
Date: 2026-02-10 06:50:03 Duration: 11ms Database: postgres
-
with rankedmt4 as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors ), last_feed_entry as ( select * from rankedmt4 where r = 1 ), ok_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where status = 'OK' ), earliest_entry_after_ok as ( select m.datafeedname, min(m.eventtimestamp) as eventtimestamp from mt4datafeederrors m left outer join ( select datafeedname, eventtimestamp from ok_entries where r = 1) oo on m.datafeedname = oo.datafeedname where m.eventtimestamp > coalesce(oo.eventtimestamp, '1900-01-01'::timestamp without time zone) group by m.datafeedname ), notified_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where notified is not null and notified <> '' ), broker as ( select *, row_number() over (partition by feedname order by brokerid) r from ( select distinct b.brokerid, b.name as brokername, dss.classname as feedname from downloadersymbolsettings dss inner join brokersymbollist bsl on dss.symbolid = bsl.symbolid inner join broker b on bsl.brokerid = b.brokerid where dss.enabled = 1) a ) select last.id, last.datafeedname, last.eventtimestamp, last.status, last.errordescription, last.serveraddress, last.username, note.notified, note.eventtimestamp, broker.brokername from last_feed_entry last inner join earliest_entry_after_ok after_ok on last.datafeedname = after_ok.datafeedname inner join broker on last.datafeedname = broker.feedname left outer join ok_entries ok on ok.datafeedname = last.datafeedname left outer join notified_entries note on note.datafeedname = last.datafeedname and note.r = 1 where (ok.r is null or ok.r = 1) and last.datafeedname not in ( select distinct datafeedname from last_feed_entry where status = 'OK') and extract(epoch from (last.eventtimestamp - after_ok.eventtimestamp)) > 60 * 60 and last.eventtimestamp > current_timestamp - interval '1 day' and (note.eventtimestamp is null or note.eventtimestamp < current_timestamp - interval '10 hours') and last.eventtimestamp > current_timestamp - interval '1 hour' and broker.r = 1;
Date: 2026-02-10 06:20:02 Duration: 9ms Database: postgres
15 57ms 239 0ms 4ms 0ms /*server.CPResult*/ SELECT patternid, resy0, resy1, supporty0, supporty1, predictiontimeto, patternstarttime, s.symbolid, resx0, resx1, supportx0, supportx1, symbol, longname, shortname, timegranularity, patternendtime, pattern, a.direction, trendchange, patternlengthbars, patternquality, resultuid as uid, breakout, initialtrend, volumeincrease, symmetry as uniformity, predictionpricefrom, predictionpriceto, noise, exchange, breakout, dtt.absolutetimezoneoffset as tzOs, dtt.timezone as tz FROM autochartist_results a INNER JOIN downloadersymbolsettings dss on a.symbolid = dss.symbolid INNER JOIN datafeedstimetable dtt ON dss.classname = dtt.classname inner join symbols s on a.symbolid = s.symbolid inner join patterns p on p.patternname = a.pattern where resultuid = $1;Times Reported Time consuming bind #15
Day Hour Count Duration Avg duration 06 239 57ms 0ms -
/*server.CPResult*/ SELECT patternid, resy0, resy1, supporty0, supporty1, predictiontimeto, patternstarttime, s.symbolid, resx0, resx1, supportx0, supportx1, symbol, longname, shortname, timegranularity, patternendtime, pattern, a.direction, trendchange, patternlengthbars, patternquality, resultuid as uid, breakout, initialtrend, volumeincrease, symmetry as uniformity, predictionpricefrom, predictionpriceto, noise, exchange, breakout, dtt.absolutetimezoneoffset as tzOs, dtt.timezone as tz FROM autochartist_results a INNER JOIN downloadersymbolsettings dss on a.symbolid = dss.symbolid INNER JOIN datafeedstimetable dtt ON dss.classname = dtt.classname inner join symbols s on a.symbolid = s.symbolid inner join patterns p on p.patternname = a.pattern where resultuid = $1;
Date: 2026-02-10 06:32:38 Duration: 4ms Database: postgres parameters: $1 = '607652888547499301'
-
/*server.CPResult*/ SELECT patternid, resy0, resy1, supporty0, supporty1, predictiontimeto, patternstarttime, s.symbolid, resx0, resx1, supportx0, supportx1, symbol, longname, shortname, timegranularity, patternendtime, pattern, a.direction, trendchange, patternlengthbars, patternquality, resultuid as uid, breakout, initialtrend, volumeincrease, symmetry as uniformity, predictionpricefrom, predictionpriceto, noise, exchange, breakout, dtt.absolutetimezoneoffset as tzOs, dtt.timezone as tz FROM autochartist_results a INNER JOIN downloadersymbolsettings dss on a.symbolid = dss.symbolid INNER JOIN datafeedstimetable dtt ON dss.classname = dtt.classname inner join symbols s on a.symbolid = s.symbolid inner join patterns p on p.patternname = a.pattern where resultuid = $1;
Date: 2026-02-10 06:37:02 Duration: 4ms Database: postgres parameters: $1 = '607652890103653301'
-
/*server.CPResult*/ SELECT patternid, resy0, resy1, supporty0, supporty1, predictiontimeto, patternstarttime, s.symbolid, resx0, resx1, supportx0, supportx1, symbol, longname, shortname, timegranularity, patternendtime, pattern, a.direction, trendchange, patternlengthbars, patternquality, resultuid as uid, breakout, initialtrend, volumeincrease, symmetry as uniformity, predictionpricefrom, predictionpriceto, noise, exchange, breakout, dtt.absolutetimezoneoffset as tzOs, dtt.timezone as tz FROM autochartist_results a INNER JOIN downloadersymbolsettings dss on a.symbolid = dss.symbolid INNER JOIN datafeedstimetable dtt ON dss.classname = dtt.classname inner join symbols s on a.symbolid = s.symbolid inner join patterns p on p.patternname = a.pattern where resultuid = $1;
Date: 2026-02-10 06:15:56 Duration: 4ms Database: postgres parameters: $1 = '607652659900137301'
16 55ms 45 0ms 3ms 1ms SELECT timegranularity FROM brokersymbollist bsl INNER JOIN symbols s ON bsl.symbolid = s.symbolid INNER JOIN downloadersymbolsettings dss on s.symbolid = dss.symbolid LEFT OUTER JOIN brokerinstrumentmapping bdfi ON bdfi.brokerid = $1 AND dss.datafeedinstrumentid = bdfi.datafeedinstrumentid WHERE s.nonliquid = 0 and s.deleted = 0 and dss.enabled = 1 AND s.symbol ILIKE $2 AND bsl.brokerid = $3 AND timegranularity >= 15 ORDER BY timegranularity LIMIT 1;Times Reported Time consuming bind #16
Day Hour Count Duration Avg duration 06 45 55ms 1ms -
SELECT timegranularity FROM brokersymbollist bsl INNER JOIN symbols s ON bsl.symbolid = s.symbolid INNER JOIN downloadersymbolsettings dss on s.symbolid = dss.symbolid LEFT OUTER JOIN brokerinstrumentmapping bdfi ON bdfi.brokerid = $1 AND dss.datafeedinstrumentid = bdfi.datafeedinstrumentid WHERE s.nonliquid = 0 and s.deleted = 0 and dss.enabled = 1 AND s.symbol ILIKE $2 AND bsl.brokerid = $3 AND timegranularity >= 15 ORDER BY timegranularity LIMIT 1;
Date: 2026-02-10 06:20:09 Duration: 3ms Database: postgres parameters: $1 = '689', $2 = 'XAUUSD', $3 = '689'
-
SELECT timegranularity FROM brokersymbollist bsl INNER JOIN symbols s ON bsl.symbolid = s.symbolid INNER JOIN downloadersymbolsettings dss on s.symbolid = dss.symbolid LEFT OUTER JOIN brokerinstrumentmapping bdfi ON bdfi.brokerid = $1 AND dss.datafeedinstrumentid = bdfi.datafeedinstrumentid WHERE s.nonliquid = 0 and s.deleted = 0 and dss.enabled = 1 AND s.symbol ILIKE $2 AND bsl.brokerid = $3 AND timegranularity >= 15 ORDER BY timegranularity LIMIT 1;
Date: 2026-02-10 06:05:09 Duration: 2ms Database: postgres parameters: $1 = '689', $2 = 'XAUUSD', $3 = '689'
-
SELECT timegranularity FROM brokersymbollist bsl INNER JOIN symbols s ON bsl.symbolid = s.symbolid INNER JOIN downloadersymbolsettings dss on s.symbolid = dss.symbolid LEFT OUTER JOIN brokerinstrumentmapping bdfi ON bdfi.brokerid = $1 AND dss.datafeedinstrumentid = bdfi.datafeedinstrumentid WHERE s.nonliquid = 0 and s.deleted = 0 and dss.enabled = 1 AND s.symbol ILIKE $2 AND bsl.brokerid = $3 AND timegranularity >= 15 ORDER BY timegranularity LIMIT 1;
Date: 2026-02-10 06:35:09 Duration: 2ms Database: postgres parameters: $1 = '689', $2 = 'XAUUSD', $3 = '689'
17 49ms 9 4ms 6ms 5ms SELECT DISTINCT ON (basegroupname, symbol) ;Times Reported Time consuming bind #17
Day Hour Count Duration Avg duration 06 9 49ms 5ms -
SELECT DISTINCT ON (basegroupname, symbol) ;
Date: 2026-02-10 06:31:26 Duration: 6ms Database: postgres parameters: $1 = '667', $2 = '667'
-
SELECT DISTINCT ON (basegroupname, symbol) ;
Date: 2026-02-10 06:46:06 Duration: 6ms Database: postgres parameters: $1 = '627', $2 = '627'
-
SELECT DISTINCT ON (basegroupname, symbol) ;
Date: 2026-02-10 06:31:07 Duration: 6ms Database: postgres parameters: $1 = '667', $2 = '667'
18 44ms 123 0ms 4ms 0ms /*server.KeyLevelResult*/ SELECT ResultUID AS ruid, s.symbolid AS sid, symbol AS sym, longname, shortname, Exchange AS e, timegranularity AS tg, a.PatternID AS pid, a.direction AS d, a.patternprice as pp, atbaridentified AS pet, CASE WHEN (x9 != '') THEN x9 WHEN (x8 != '') THEN x8 WHEN (x7 != '') THEN x7 WHEN (x6 != '') THEN x6 WHEN (x5 != '') THEN x5 WHEN (x4 != '') THEN x4 WHEN (x3 != '') THEN x3 WHEN (x2 != '') THEN x2 END AS pst, PatternPrice AS patp, x0, x1, x2, CASE WHEN (x3 != '') THEN x3 ELSE '1900-01-01' END as x3, CASE WHEN (x4 != '') THEN x4 ELSE '1900-01-01' END as x4, CASE WHEN (x5 != '') THEN x5 ELSE '1900-01-01' END as x5, CASE WHEN (x6 != '') THEN x6 ELSE '1900-01-01' END as x6, CASE WHEN (x7 != '') THEN x7 ELSE '1900-01-01' END as x7, CASE WHEN (x8 != '') THEN x8 ELSE '1900-01-01' END as x8, CASE WHEN (x9 != '') THEN x9 ELSE '1900-01-01' END as x9, errorMargin as erm, breakoutprice as pE, breakoutbars as be, breakout, atbaridentified as atBar, PatternLengthBars AS l, Bandwidth AS bw, QtyTP AS qtp, p.patternname as patternname, dtt.absolutetimezoneoffset as tzOs, dtt.timezone as tz, approachingtimestamp AS apt, approachingregion as apr, predictionpricefrom as ppf, predictionpriceto as ppt, predictiontimefrom as ptf, predictiontimebars as ptb FROM keylevels_results a INNER JOIN downloadersymbolsettings dss on a.symbolid = dss.symbolid INNER JOIN datafeedstimetable dtt ON dss.classname = dtt.classname inner join symbols s on a.symbolid = s.symbolid INNER JOIN hrspatterns p on a.patternid = p.patternid where resultuid = $1 and dtt.dayofweek = 3;Times Reported Time consuming bind #18
Day Hour Count Duration Avg duration 06 123 44ms 0ms -
/*server.KeyLevelResult*/ SELECT ResultUID AS ruid, s.symbolid AS sid, symbol AS sym, longname, shortname, Exchange AS e, timegranularity AS tg, a.PatternID AS pid, a.direction AS d, a.patternprice as pp, atbaridentified AS pet, CASE WHEN (x9 != '') THEN x9 WHEN (x8 != '') THEN x8 WHEN (x7 != '') THEN x7 WHEN (x6 != '') THEN x6 WHEN (x5 != '') THEN x5 WHEN (x4 != '') THEN x4 WHEN (x3 != '') THEN x3 WHEN (x2 != '') THEN x2 END AS pst, PatternPrice AS patp, x0, x1, x2, CASE WHEN (x3 != '') THEN x3 ELSE '1900-01-01' END as x3, CASE WHEN (x4 != '') THEN x4 ELSE '1900-01-01' END as x4, CASE WHEN (x5 != '') THEN x5 ELSE '1900-01-01' END as x5, CASE WHEN (x6 != '') THEN x6 ELSE '1900-01-01' END as x6, CASE WHEN (x7 != '') THEN x7 ELSE '1900-01-01' END as x7, CASE WHEN (x8 != '') THEN x8 ELSE '1900-01-01' END as x8, CASE WHEN (x9 != '') THEN x9 ELSE '1900-01-01' END as x9, errorMargin as erm, breakoutprice as pE, breakoutbars as be, breakout, atbaridentified as atBar, PatternLengthBars AS l, Bandwidth AS bw, QtyTP AS qtp, p.patternname as patternname, dtt.absolutetimezoneoffset as tzOs, dtt.timezone as tz, approachingtimestamp AS apt, approachingregion as apr, predictionpricefrom as ppf, predictionpriceto as ppt, predictiontimefrom as ptf, predictiontimebars as ptb FROM keylevels_results a INNER JOIN downloadersymbolsettings dss on a.symbolid = dss.symbolid INNER JOIN datafeedstimetable dtt ON dss.classname = dtt.classname inner join symbols s on a.symbolid = s.symbolid INNER JOIN hrspatterns p on a.patternid = p.patternid where resultuid = $1 and dtt.dayofweek = 3;
Date: 2026-02-10 06:05:27 Duration: 4ms Database: postgres parameters: $1 = '607652827547040303'
-
/*server.KeyLevelResult*/ SELECT ResultUID AS ruid, s.symbolid AS sid, symbol AS sym, longname, shortname, Exchange AS e, timegranularity AS tg, a.PatternID AS pid, a.direction AS d, a.patternprice as pp, atbaridentified AS pet, CASE WHEN (x9 != '') THEN x9 WHEN (x8 != '') THEN x8 WHEN (x7 != '') THEN x7 WHEN (x6 != '') THEN x6 WHEN (x5 != '') THEN x5 WHEN (x4 != '') THEN x4 WHEN (x3 != '') THEN x3 WHEN (x2 != '') THEN x2 END AS pst, PatternPrice AS patp, x0, x1, x2, CASE WHEN (x3 != '') THEN x3 ELSE '1900-01-01' END as x3, CASE WHEN (x4 != '') THEN x4 ELSE '1900-01-01' END as x4, CASE WHEN (x5 != '') THEN x5 ELSE '1900-01-01' END as x5, CASE WHEN (x6 != '') THEN x6 ELSE '1900-01-01' END as x6, CASE WHEN (x7 != '') THEN x7 ELSE '1900-01-01' END as x7, CASE WHEN (x8 != '') THEN x8 ELSE '1900-01-01' END as x8, CASE WHEN (x9 != '') THEN x9 ELSE '1900-01-01' END as x9, errorMargin as erm, breakoutprice as pE, breakoutbars as be, breakout, atbaridentified as atBar, PatternLengthBars AS l, Bandwidth AS bw, QtyTP AS qtp, p.patternname as patternname, dtt.absolutetimezoneoffset as tzOs, dtt.timezone as tz, approachingtimestamp AS apt, approachingregion as apr, predictionpricefrom as ppf, predictionpriceto as ppt, predictiontimefrom as ptf, predictiontimebars as ptb FROM keylevels_results a INNER JOIN downloadersymbolsettings dss on a.symbolid = dss.symbolid INNER JOIN datafeedstimetable dtt ON dss.classname = dtt.classname inner join symbols s on a.symbolid = s.symbolid INNER JOIN hrspatterns p on a.patternid = p.patternid where resultuid = $1 and dtt.dayofweek = 3;
Date: 2026-02-10 06:20:46 Duration: 3ms Database: postgres parameters: $1 = '607652888411092303'
-
/*server.KeyLevelResult*/ SELECT ResultUID AS ruid, s.symbolid AS sid, symbol AS sym, longname, shortname, Exchange AS e, timegranularity AS tg, a.PatternID AS pid, a.direction AS d, a.patternprice as pp, atbaridentified AS pet, CASE WHEN (x9 != '') THEN x9 WHEN (x8 != '') THEN x8 WHEN (x7 != '') THEN x7 WHEN (x6 != '') THEN x6 WHEN (x5 != '') THEN x5 WHEN (x4 != '') THEN x4 WHEN (x3 != '') THEN x3 WHEN (x2 != '') THEN x2 END AS pst, PatternPrice AS patp, x0, x1, x2, CASE WHEN (x3 != '') THEN x3 ELSE '1900-01-01' END as x3, CASE WHEN (x4 != '') THEN x4 ELSE '1900-01-01' END as x4, CASE WHEN (x5 != '') THEN x5 ELSE '1900-01-01' END as x5, CASE WHEN (x6 != '') THEN x6 ELSE '1900-01-01' END as x6, CASE WHEN (x7 != '') THEN x7 ELSE '1900-01-01' END as x7, CASE WHEN (x8 != '') THEN x8 ELSE '1900-01-01' END as x8, CASE WHEN (x9 != '') THEN x9 ELSE '1900-01-01' END as x9, errorMargin as erm, breakoutprice as pE, breakoutbars as be, breakout, atbaridentified as atBar, PatternLengthBars AS l, Bandwidth AS bw, QtyTP AS qtp, p.patternname as patternname, dtt.absolutetimezoneoffset as tzOs, dtt.timezone as tz, approachingtimestamp AS apt, approachingregion as apr, predictionpricefrom as ppf, predictionpriceto as ppt, predictiontimefrom as ptf, predictiontimebars as ptb FROM keylevels_results a INNER JOIN downloadersymbolsettings dss on a.symbolid = dss.symbolid INNER JOIN datafeedstimetable dtt ON dss.classname = dtt.classname inner join symbols s on a.symbolid = s.symbolid INNER JOIN hrspatterns p on a.patternid = p.patternid where resultuid = $1 and dtt.dayofweek = 3;
Date: 2026-02-10 06:20:18 Duration: 3ms Database: postgres parameters: $1 = '607652829799758303'
19 43ms 349 0ms 2ms 0ms INSERT INTO T240 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;Times Reported Time consuming bind #19
Day Hour Count Duration Avg duration 06 349 43ms 0ms -
INSERT INTO T240 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;
Date: 2026-02-10 06:31:53 Duration: 2ms Database: postgres parameters: $1 = '2026-02-09 20:00:00', $2 = '50127.05', $3 = '50201.05', $4 = '50025', $5 = '50123.5', $6 = '33503', $7 = '515840249388410300', $8 = '0', $9 = '2026-02-10 06:31:53.084', $10 = '2026-02-10 06:31:53.025', $11 = '50127.05', $12 = '50201.05', $13 = '50025', $14 = '50123.5', $15 = '33503', $16 = '0', $17 = '2026-02-10 06:31:53.084', $18 = '2026-02-10 06:31:53.025'
-
INSERT INTO T240 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;
Date: 2026-02-10 06:31:50 Duration: 0ms Database: postgres parameters: $1 = '2026-02-09 20:00:00', $2 = '86.09555', $3 = '88.6816', $4 = '85.6638', $5 = '87.77365', $6 = '38462', $7 = '515840249476671300', $8 = '0', $9 = '2026-02-10 06:31:50.959', $10 = '2026-02-10 06:31:50.938', $11 = '86.09555', $12 = '88.6816', $13 = '85.6638', $14 = '87.77365', $15 = '38462', $16 = '0', $17 = '2026-02-10 06:31:50.959', $18 = '2026-02-10 06:31:50.938'
-
INSERT INTO T240 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;
Date: 2026-02-10 06:16:23 Duration: 0ms Database: postgres parameters: $1 = '2026-02-09 20:00:00', $2 = '110.6125', $3 = '110.644', $4 = '110.472', $5 = '110.567', $6 = '18238', $7 = '515840249384281300', $8 = '0', $9 = '2026-02-10 06:16:23.522', $10 = '2026-02-10 06:16:23.401', $11 = '110.6125', $12 = '110.644', $13 = '110.472', $14 = '110.567', $15 = '18238', $16 = '0', $17 = '2026-02-10 06:16:23.522', $18 = '2026-02-10 06:16:23.401'
20 41ms 288 0ms 0ms 0ms INSERT INTO T1440_underlying (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;Times Reported Time consuming bind #20
Day Hour Count Duration Avg duration 06 288 41ms 0ms -
INSERT INTO T1440_underlying (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;
Date: 2026-02-10 06:31:50 Duration: 0ms Database: postgres parameters: $1 = '2026-02-09 00:00:00', $2 = '86.894', $3 = '89.03405', $4 = '82.883', $5 = '87.77365', $6 = '198348', $7 = '515840249476835300', $8 = '0', $9 = '2026-02-10 06:31:50.95', $10 = '2026-02-10 06:31:50.95', $11 = '86.894', $12 = '89.03405', $13 = '82.883', $14 = '87.77365', $15 = '198348', $16 = '0', $17 = '2026-02-10 06:31:50.95', $18 = '2026-02-10 06:31:50.95'
-
INSERT INTO T1440_underlying (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;
Date: 2026-02-10 06:32:33 Duration: 0ms Database: postgres parameters: $1 = '2026-02-09 00:00:00', $2 = '0.09535', $3 = '0.09775', $4 = '0.09085', $5 = '0.09425', $6 = '15007', $7 = '515840249475376300', $8 = '0', $9 = '2026-02-10 06:32:33.309', $10 = '2026-02-10 06:32:33.308', $11 = '0.09535', $12 = '0.09775', $13 = '0.09085', $14 = '0.09425', $15 = '15007', $16 = '0', $17 = '2026-02-10 06:32:33.309', $18 = '2026-02-10 06:32:33.308'
-
INSERT INTO T1440_underlying (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;
Date: 2026-02-10 06:17:42 Duration: 0ms Database: postgres parameters: $1 = '2026-02-09 00:00:00', $2 = '0.6014', $3 = '0.60599', $4 = '0.599725', $5 = '0.60556', $6 = '48226', $7 = '515840249465415300', $8 = '0', $9 = '2026-02-10 06:17:42.023', $10 = '2026-02-10 06:17:42.022', $11 = '0.6014', $12 = '0.60599', $13 = '0.599725', $14 = '0.60556', $15 = '48226', $16 = '0', $17 = '2026-02-10 06:17:42.023', $18 = '2026-02-10 06:17:42.022'
-
Events
Log levels
Key values
- 449,365 Log entries
Events distribution
Key values
- 0 PANIC entries
- 0 FATAL entries
- 8 ERROR entries
- 0 WARNING entries
Most Frequent Errors/Events
Key values
- 5 Max number of times the same event was reported
- 8 Total events found
Rank Times reported Error 1 5 LOG: process ... still waiting for AccessShareLock on relation ... of database ... after ... ms
Times Reported Most Frequent Error / Event #1
Day Hour Count Feb 10 06 5 - LOG: process 29076 still waiting for AccessShareLock on relation 5894441 of database 5881926 after 1000.062 ms at character 28
- LOG: process 29278 still waiting for AccessShareLock on relation 5894441 of database 5881926 after 1000.048 ms at character 28
- LOG: process 29278 still waiting for AccessShareLock on relation 5894441 of database 5881926 after 33274.279 ms at character 28
Detail: Process holding the lock: 15939. Wait queue: 19602, 29076.
Statement: SELECT datetimeupdate FROM latest_candle_datetime_per_receng WHERE recognitionengine ILIKE 'IQFEED_FX - 1'Date: 2026-02-10 06:17:15
Detail: Process holding the lock: 15939. Wait queue: 19602, 29076, 29278.
Statement: SELECT datetimeupdate FROM latest_candle_datetime_per_receng WHERE recognitionengine ILIKE 'PEPPERSTONE - 1'Date: 2026-02-10 06:17:30
Detail: Process holding the lock: 15939. Wait queue: 19602, 29076, 29278.
Statement: SELECT datetimeupdate FROM latest_candle_datetime_per_receng WHERE recognitionengine ILIKE 'PEPPERSTONE - 1'Date: 2026-02-10 06:18:02
2 3 LOG: process ... still waiting for AccessExclusiveLock on relation ... of database ... after ... ms
Times Reported Most Frequent Error / Event #2
Day Hour Count Feb 10 06 3 - LOG: process 15993 still waiting for AccessExclusiveLock on relation 5883477 of database 5881926 after 1000.064 ms
- LOG: process 19602 still waiting for AccessExclusiveLock on relation 5894441 of database 5881926 after 1000.058 ms
- LOG: process 19602 still waiting for AccessExclusiveLock on relation 5894441 of database 5881926 after 89083.669 ms
Detail: Process holding the lock: 15939. Wait queue: 15993.
Statement: TRUNCATE TABLE solr_relevance_old;Date: 2026-02-10 06:05:12
Detail: Process holding the lock: 15939. Wait queue: 19602.
Statement: refresh materialized view latest_candle_datetime_per_recengDate: 2026-02-10 06:17:02
Detail: Process holding the lock: 15939. Wait queue: 19602, 29076, 29278.
Statement: refresh materialized view latest_candle_datetime_per_recengDate: 2026-02-10 06:18:30