-
Global information
- Generated on Wed Feb 25 13:59:50 2026
- Log file: /home/postgres/pg_data/data/pg_log/postgresql-2026-02-25_150000.log
- Parsed 1,889,656 log entries in 49s
- Log start from 2026-02-25 15:00:00 to 2026-02-25 15:59:48
-
Overview
Global Stats
- 254 Number of unique normalized queries
- 223,064 Number of queries
- 2h34m2s Total query duration
- 2026-02-25 15:00:00 First query
- 2026-02-25 15:59:48 Last query
- 6,468 queries/s at 2026-02-25 15:45:04 Query peak
- 2h34m2s Total query duration
- 8s239ms Prepare/parse total duration
- 57s798ms Bind total duration
- 2h32m56s Execute total duration
- 406 Number of events
- 2 Number of unique normalized events
- 359 Max number of times the same event was reported
- 0 Number of cancellation
- 54 Total number of automatic vacuums
- 59 Total number of automatic analyzes
- 801 Number temporary file
- 150.43 MiB Max size of temporary file
- 7.40 MiB Average size of temporary file
- 3,854 Total number of sessions
- 332 sessions at 2026-02-25 15:57:04 Session peak
- 10d23h15m35s Total duration of sessions
- 4m5s Average duration of sessions
- 57 Average queries per session
- 2s398ms Average queries duration per session
- 4m3s Average idle time per session
- 3,845 Total number of connections
- 159 connections/s at 2026-02-25 15:38:40 Connection peak
- 3 Total number of databases
SQL Traffic
Key values
- 6,468 queries/s Query Peak
- 2026-02-25 15:45:04 Date
SELECT Traffic
Key values
- 3,194 queries/s Query Peak
- 2026-02-25 15:45:04 Date
INSERT/UPDATE/DELETE Traffic
Key values
- 157 queries/s Query Peak
- 2026-02-25 15:00:52 Date
Queries duration
Key values
- 2h34m2s Total query duration
Prepared queries ratio
Key values
- 0.00 Ratio of bind vs prepare
- 0.00 % Ratio between prepared and "usual" statements
General Activity
↑ Back to the top of the General Activity tableDay Hour Count Min duration Max duration Avg duration Latency Percentile(90) Latency Percentile(95) Latency Percentile(99) Feb 25 15 223,064 0ms 41s640ms 41ms 3m47s 4m6s 5m Day Hour SELECT COPY TO Average Duration Latency Percentile(90) Latency Percentile(95) Latency Percentile(99) Feb 25 15 73,394 26 0ms 0ms 0ms 0ms Day Hour INSERT UPDATE DELETE COPY FROM Average Duration Latency Percentile(90) Latency Percentile(95) Latency Percentile(99) Feb 25 15 24,088 2,642 16 96 0ms 0ms 0ms 0ms Day Hour Prepare Bind Bind/Prepare Percentage of prepare Feb 25 15 22,178 74,773 3.37 20.79% Day Hour Count Average / Second Feb 25 15 3,845 1.07/s Day Hour Count Average Duration Average idle time Feb 25 15 3,854 4m5s 4m3s -
Connections
Established Connections
Key values
- 159 connections Connection Peak
- 2026-02-25 15:38:40 Date
Connections per database
Key values
- acaweb_fx Main Database
- 3,845 connections Total
Connections per user
Key values
- postgres Main User
- 3,845 connections Total
Connections per host
Key values
- 192.168.4.142 Main host with 1423 connections
- 3,845 Total connections
Host Count 104.30.164.187 641 127.0.0.1 112 192.168.0.114 10 192.168.0.216 104 192.168.0.239 3 192.168.0.74 203 192.168.1.127 7 192.168.1.145 198 192.168.1.15 65 192.168.1.20 231 192.168.1.239 15 192.168.1.90 54 192.168.2.126 39 192.168.2.182 24 192.168.3.199 36 192.168.4.142 1,423 192.168.4.145 1 192.168.4.150 10 192.168.4.175 4 192.168.4.238 12 192.168.4.33 75 192.168.4.9 4 192.168.4.98 330 [local] 244 -
Sessions
Simultaneous sessions
Key values
- 332 sessions Session Peak
- 2026-02-25 15:57:04 Date
Histogram of session times
Key values
- 2,611 0-500ms duration
Sessions per database
Key values
- acaweb_fx Main Database
- 3,854 sessions Total
Sessions per user
Key values
- postgres Main User
- 3,854 sessions Total
Sessions per host
Key values
- 192.168.4.142 Main Host
- 3,854 sessions Total
Host Count Total Duration Average Duration 104.30.164.187 641 14h24m57s 1m20s 127.0.0.1 112 13s413ms 119ms 192.168.0.114 9 49m3s 5m27s 192.168.0.171 4 12h7m8s 3h1m47s 192.168.0.216 104 2m14s 1s288ms 192.168.0.239 3 5d10h39m12s 1d19h33m4s 192.168.0.74 203 5h47m40s 1m42s 192.168.1.127 7 13s893ms 1s984ms 192.168.1.145 198 4h44m7s 1m26s 192.168.1.15 66 3h6m10s 2m49s 192.168.1.154 4 11h31m25s 2h52m51s 192.168.1.20 231 14h56m18s 3m52s 192.168.1.239 15 105ms 7ms 192.168.1.28 1 1d20h35m6s 1d20h35m6s 192.168.1.90 54 44s322ms 820ms 192.168.2.126 39 6s788ms 174ms 192.168.2.182 24 16s684ms 695ms 192.168.3.199 36 4s699ms 130ms 192.168.4.142 1,423 8m35s 362ms 192.168.4.145 1 283ms 283ms 192.168.4.150 10 20h12m17s 2h1m13s 192.168.4.175 4 26s16ms 6s504ms 192.168.4.238 12 15s545ms 1s295ms 192.168.4.33 75 4m36s 3s689ms 192.168.4.9 4 35ms 8ms 192.168.4.98 330 13s348ms 40ms [local] 244 4m4s 1s2ms -
Checkpoints / Restartpoints
Checkpoints Buffers
Key values
- 10,564 buffers Checkpoint Peak
- 2026-02-25 15:08:32 Date
- 209.971 seconds Highest write time
- 0.059 seconds Sync time
Checkpoints Wal files
Key values
- 6 files Wal files usage Peak
- 2026-02-25 15:08:32 Date
Checkpoints distance
Key values
- 184.28 Mo Distance Peak
- 2026-02-25 15:08:32 Date
Checkpoints Activity
↑ Back to the top of the Checkpoint Activity tableDay Hour Written buffers Write time Sync time Total time Feb 25 15 33,871 1,785.867s 0.096s 1,786.356s Day Hour Added Removed Recycled Synced files Longest sync Average sync Feb 25 15 0 0 23 1,881 0.052s 0s Day Hour Count Avg time (sec) Feb 25 15 0 0s Day Hour Mean distance Mean estimate Feb 25 15 30,717.50 kB 67,218.42 kB -
Temporary Files
Size of temporary files
Key values
- 184.76 MiB Temp Files size Peak
- 2026-02-25 15:40:08 Date
Number of temporary files
Key values
- 30 per second Temp Files Peak
- 2026-02-25 15:02:09 Date
Temporary Files Activity
↑ Back to the top of the Temporary Files Activity tableDay Hour Count Total size Average size Feb 25 15 801 5.79 GiB 7.40 MiB Queries generating the most temporary files (N)
Rank Count Total size Min size Max size Avg size Query 1 91 373.77 MiB 3.99 MiB 4.25 MiB 4.11 MiB select resultuid from relevance_fibonacci_results order by resultuid desc limit ?), fr as ( select a.*, rr.age, rr.relevant from fibonacci_results a left outer join relevance_fibonacci_results rr on a.resultuid = rr.resultuid where case when false = ? then true else a.resultuid > ( select min(resultuid) from relevance_fibonacci_results) end), all_results as ( select fr.resultuid as resultuid, fr.direction as direction, s.exchange as exchange, s.symbolid as symbolid, coalesce(bim.code, s.symbol) as symbol_code, s.longname as symbol_name, s.timegranularity as interval, fr.pattern as pattern_name, fr.timed as timed, fr.patternendtime as identified, dtt.timezone as timezone, fr.patternlengthbars as length, g.basegroupname, newlevels.filtered, case when fr.age is not null then fr.age when fr.resultuid <= rm.resultuid then ? else ? end as age, case when fr.relevant is not null then fr.relevant when fr.resultuid <= rm.resultuid then ? else ? end as relevant, cps.pip from fr inner join brokersymbollist bsl on bsl.brokerid = ? and bsl.symbolid = fr.symbolid inner join symbols s on fr.symbolid = s.symbolid and s.nonliquid = ? inner join symbolgroup sg on fr.symbolid = sg.symbolid inner join groups g on sg.groupid = g.groupid inner join brokergroups bg on g.groupid = bg.groupid and bsl.brokerid = bg.brokerid inner join downloadersymbolsettings dss on fr.symbolid = dss.symbolid inner join datafeedstimetable dtt on dss.classname = dtt.classname and dtt.dayofweek = ? inner join rar_max rm on ? = ? left join lateral calc_fib_signal_filter (fr.resultuid) newlevels on true left join currencypips cps on cps.symbol = s.symbol left outer join brokerinstrumentmap bim on dss.datafeedinstrumentid = bim.datafeedinstrumentid and bim.brokerid = bsl.brokerid and bim.type = ? where fr.gmttimefound > now() - interval ? and dss.enabled = ? and s.deleted = ? and (fr.simulation = ? or fr.simulation is null) and (? = ? or s.timegranularity in (...)) and (? = ? or s.exchange in (...)) and (? = ? or coalesce(bim.code, s.symbol) in (...)) and (? = ? or fr.pattern in (...)) and (? = ? or fr.patternlengthbars <= ?) and (? = ? or (? = ? and fr.timed > cast(? as timestamp)) or (? = ? and fr.timed < cast(? as timestamp)))), results as ( select distinct on (symbolid) * from all_results where (false = ? or relevant = ?) and (? = ? or age <= ?) order by symbolid, resultuid ) select * from results order by identified desc, length desc;-
SELECT resultuid FROM relevance_fibonacci_results ORDER BY resultuid DESC LIMIT 1), fr AS ( SELECT a.*, rr.age, rr.relevant from fibonacci_results a LEFT OUTER JOIN relevance_fibonacci_results rr on a.resultuid = rr.resultuid WHERE CASE WHEN FALSE = $1 THEN true ELSE a.resultuid > ( select min(resultuid) from relevance_fibonacci_results) END), all_results AS ( SELECT fr.resultuid AS resultuid, fr.direction AS direction, s.exchange AS exchange, s.symbolid AS symbolid, coalesce(bim.code, s.symbol) AS symbol_code, s.longname AS symbol_name, s.timegranularity AS interval, fr.pattern AS pattern_name, fr.timed AS timed, fr.patternendtime AS identified, dtt.timezone AS timezone, fr.patternlengthbars AS length, g.basegroupname, newLevels.filtered, CASE WHEN fr.age IS NOT NULL THEN fr.age WHEN fr.resultuid <= rm.resultuid THEN 11 ELSE 0 END as age, CASE WHEN fr.relevant IS NOT NULL THEN fr.relevant WHEN fr.resultuid <= rm.resultuid THEN 0 ELSE 1 END as relevant, cps.pip FROM fr INNER JOIN brokersymbollist bsl ON bsl.brokerid = $2 AND bsl.symbolid = fr.symbolid INNER JOIN symbols s ON fr.symbolid = s.symbolid AND s.nonliquid = 0 INNER JOIN symbolgroup sg on fr.symbolid = sg.symbolid INNER JOIN groups g ON sg.groupid = g.groupid INNER JOIN brokergroups bg on g.groupid = bg.groupid AND bsl.brokerid = bg.brokerid INNER JOIN downloadersymbolsettings dss ON fr.symbolid = dss.symbolid INNER JOIN datafeedstimetable dtt ON dss.classname = dtt.classname AND dtt.dayofweek = 3 INNER JOIN rar_max rm ON 1 = 1 LEFT JOIN LATERAL calc_fib_signal_filter (fr.resultuid) newLevels on true LEFT JOIN currencypips cps on cps.symbol = s.symbol LEFT OUTER JOIN brokerinstrumentmap bim ON dss.datafeedinstrumentid = bim.datafeedinstrumentid AND bim.brokerid = bsl.brokerid AND bim.TYPE = 'OUTBOUND' WHERE fr.gmttimefound > now() - INTERVAL '7 DAYS' AND dss.enabled = 1 AND s.deleted = 0 AND (fr.simulation = 0 OR fr.simulation IS NULL) AND ($3 = 0 OR s.timegranularity in ($4)) AND ($5 = 0 OR s.exchange in ($6)) AND ($7 = 0 OR coalesce(bim.code, s.symbol) in ($8)) AND ($9 = 0 OR fr.pattern in ($10)) AND ($11 = 0 OR fr.patternlengthbars <= $12) AND ($13 = 0 OR ($14 = 1 AND fr.timed > cast('1970-01-01' as timestamp)) OR ($15 = 2 AND fr.timed < cast('1970-01-01' as timestamp)))), results AS ( SELECT DISTINCT ON (symbolid) * FROM all_results WHERE (FALSE = $16 OR relevant = 1) AND ($17 = 0 OR age <= $18) ORDER BY symbolid, resultuid ) SELECT * from results ORDER BY identified DESC, length DESC;
Date: 2026-02-25 15:00:14 Duration: 0ms
2 54 326.80 MiB 3.26 MiB 9.10 MiB 6.05 MiB jr.resultuid as resultuid, jr.direction as direction, jr.patternendtime as identified, jr.patternlengthbars as length, jr.patternstarttime as patternstarttime, case when jr.trendchangeid = ? then ? else ? end as trendchange, s.exchange as exchange, s.symbolid as symbolid, coalesce(bim.code, s.symbol) as symbol_code, s.longname as symbol_name, s.timegranularity as interval, jp.patternname as pattern_name, dtt.timezone as timezone, ? as age, cps.pip, g.basegroupname from japsticks_results jr inner join brokersymbollist bsl on bsl.brokerid = ? and bsl.symbolid = jr.symbolid inner join relevance_japsticks_results rar on rar.resultuid = jr.resultuid inner join symbols s on jr.symbolid = s.symbolid and s.nonliquid = ? inner join japsticks_patterns jp on jr.patternid = jp.id inner join downloadersymbolsettings dss on jr.symbolid = dss.symbolid inner join datafeedstimetable dtt on dss.classname = dtt.classname and dtt.dayofweek = ? inner join symbolgroup sg on s.symbolid = sg.symbolid inner join groups g on sg.groupid = g.groupid inner join brokergroups bg on g.groupid = bg.groupid and bsl.brokerid = bg.brokerid left join currencypips cps on cps.symbol = s.symbol left outer join brokerinstrumentmap bim on dss.datafeedinstrumentid = bim.datafeedinstrumentid and bim.brokerid = bsl.brokerid and bim.type = ? where jr.gmttimefound > now() - interval ? and s.deleted = ? and (jr.simulation = ? or jr.simulation is null) and (rar.relevant = ?) --and (semicolon_age = ? or rar.age <= semicolon_age) and (? = ? or s.timegranularity in (...)) and (? = ? or s.exchange in (...)) and (? = ? or coalesce(bim.code, s.symbol) in (...)) and (? = ? or jp.patternname in (...)) and (? = ? or jr.patternlengthbars <= ?) ), results as ( select distinct on (symbolid) * from all_results order by symbolid, resultuid ) select * from results order by identified desc, length desc ;-
jr.resultuid AS resultuid, jr.direction AS direction, jr.patternendtime AS identified, jr.patternlengthbars AS length, jr.patternstarttime AS patternstarttime, case when jr.trendchangeid = 1 then 'Continuation' else 'Reversal' end AS trendchange, s.exchange AS exchange, s.symbolid AS symbolid, coalesce(bim.code, s.symbol) AS symbol_code, s.longname AS symbol_name, s.timegranularity AS interval, jp.patternname AS pattern_name, dtt.timezone AS timezone, 0 AS age, cps.pip, g.basegroupname FROM japsticks_results jr INNER JOIN brokersymbollist bsl ON bsl.brokerid = $1 AND bsl.symbolid = jr.symbolid INNER JOIN relevance_japsticks_results rar ON rar.resultuid = jr.resultuid INNER JOIN symbols s ON jr.symbolid = s.symbolid AND s.nonliquid = 0 INNER JOIN japsticks_patterns jp ON jr.patternid = jp.id INNER JOIN downloadersymbolsettings dss ON jr.symbolid = dss.symbolid INNER JOIN datafeedstimetable dtt ON dss.classname = dtt.classname AND dtt.dayofweek = 3 INNER JOIN symbolgroup sg on s.symbolid = sg.symbolid INNER JOIN groups g ON sg.groupid = g.groupid INNER JOIN brokergroups bg on g.groupid = bg.groupid AND bsl.brokerid = bg.brokerid LEFT JOIN currencypips cps on cps.symbol = s.symbol LEFT OUTER JOIN brokerinstrumentmap bim ON dss.datafeedinstrumentid = bim.datafeedinstrumentid AND bim.brokerid = bsl.brokerid AND bim.TYPE = 'OUTBOUND' WHERE jr.gmttimefound > now() - INTERVAL '7 DAYS' AND s.deleted = 0 AND (jr.simulation = 0 OR jr.simulation IS NULL) AND (rar.relevant = 1) --AND (semicolon_age = 0 OR rar.age <= semicolon_age) AND ($2 = 0 OR s.timegranularity in ($3)) AND ($4 = 0 OR s.exchange in ($5)) AND ($6 = 0 OR coalesce(bim.code, s.symbol) in ($7)) AND ($8 = 0 OR jp.patternname in ($9)) AND ($10 = 0 OR jr.patternlengthbars <= $11)), results AS ( SELECT DISTINCT ON (symbolid) * FROM all_results ORDER BY symbolid, resultuid ) SELECT * from results ORDER BY identified DESC, length DESC;
Date: 2026-02-25 15:01:00 Duration: 0ms
3 53 190.66 MiB 3.38 MiB 3.92 MiB 3.60 MiB select resultuid from relevance_consecutivecandles_results order by resultuid desc limit ?), all_results as ( select ccr.resultuid as resultuid, ccr.direction as direction, s.exchange as exchange, s.symbolid as symbolid, coalesce(bim.code, s.symbol) as symbol_code, s.longname as symbol_name, s.timegranularity as interval, ccr.patternendtime as identified, dtt.timezone as timezone, ccr.qtyconsecutivecandles as length, g.basegroupname, case when rcr.age is not null then rcr.age when ccr.resultuid <= rm.resultuid then ? else ? end as age, case when rcr.relevant is not null then rcr.relevant when ccr.resultuid <= rm.resultuid then ? else ? end as relevant, cps.pip, newlevels.filtered from consecutivecandles_results ccr inner join brokersymbollist bsl on bsl.brokerid = ? and bsl.symbolid = ccr.symbolid inner join symbols s on ccr.symbolid = s.symbolid and s.nonliquid = ? inner join downloadersymbolsettings dss on ccr.symbolid = dss.symbolid inner join datafeedstimetable dtt on dss.classname = dtt.classname and dtt.dayofweek = ? inner join symbolgroup sg on ccr.symbolid = sg.symbolid inner join groups g on sg.groupid = g.groupid inner join brokergroups bg on g.groupid = bg.groupid and bsl.brokerid = bg.brokerid inner join rar_max rm on ? = ? left outer join relevance_consecutivecandles_results rcr on rcr.resultuid = ccr.resultuid left join currencypips cps on cps.symbol = s.symbol left outer join brokerinstrumentmap bim on dss.datafeedinstrumentid = bim.datafeedinstrumentid and bim.brokerid = bsl.brokerid and bim.type = ? left join lateral calc_cc_signal_filter (ccr.resultuid) newlevels on true where ccr.gmttimefound > now() - interval ? and s.deleted = ? and (ccr.simulation = ? or ccr.simulation is null) and (? = ? or s.timegranularity in (...)) and (? = ? or s.exchange in (...)) and (? = ? or coalesce(bim.code, s.symbol) in (...)) and (? = ? or ccr.patternlengthbars <= ?)), results as ( select distinct on (symbolid) * from all_results where (false = ? or relevant = ?) and (? = ? or age <= ?) order by symbolid, resultuid ) select * from results order by identified desc, length desc;-
SELECT resultuid FROM relevance_consecutivecandles_results ORDER BY resultuid DESC LIMIT 1), all_results AS ( SELECT ccr.resultuid AS resultuid, ccr.direction AS direction, s.exchange AS exchange, s.symbolid AS symbolid, coalesce(bim.code, s.symbol) AS symbol_code, s.longname AS symbol_name, s.timegranularity AS interval, ccr.patternendtime AS identified, dtt.timezone AS timezone, ccr.qtyconsecutivecandles AS length, g.basegroupname, CASE WHEN rcr.age IS NOT NULL THEN rcr.age WHEN ccr.resultuid <= rm.resultuid THEN 1 ELSE 0 END as age, CASE WHEN rcr.relevant IS NOT NULL THEN rcr.relevant WHEN ccr.resultuid <= rm.resultuid THEN 0 ELSE 1 END as relevant, cps.pip, newLevels.filtered FROM consecutivecandles_results ccr INNER JOIN brokersymbollist bsl ON bsl.brokerid = $1 AND bsl.symbolid = ccr.symbolid INNER JOIN symbols s ON ccr.symbolid = s.symbolid AND s.nonliquid = 0 INNER JOIN downloadersymbolsettings dss ON ccr.symbolid = dss.symbolid INNER JOIN datafeedstimetable dtt ON dss.classname = dtt.classname AND dtt.dayofweek = 3 INNER JOIN symbolgroup sg on ccr.symbolid = sg.symbolid INNER JOIN groups g ON sg.groupid = g.groupid INNER JOIN brokergroups bg on g.groupid = bg.groupid AND bsl.brokerid = bg.brokerid INNER JOIN rar_max rm ON 1 = 1 LEFT OUTER JOIN relevance_consecutivecandles_results rcr ON rcr.resultuid = ccr.resultuid LEFT JOIN currencypips cps on cps.symbol = s.symbol LEFT OUTER JOIN brokerinstrumentmap bim ON dss.datafeedinstrumentid = bim.datafeedinstrumentid AND bim.brokerid = bsl.brokerid AND bim.TYPE = 'OUTBOUND' LEFT JOIN LATERAL calc_cc_signal_filter (ccr.resultuid) newLevels on true WHERE ccr.gmttimefound > now() - INTERVAL '7 DAYS' AND s.deleted = 0 AND (ccr.simulation = 0 OR ccr.simulation IS NULL) AND ($2 = 0 OR s.timegranularity in ($3, $4, $5, $6, $7, $8, $9)) AND ($10 = 0 OR s.exchange in ($11)) AND ($12 = 0 OR coalesce(bim.code, s.symbol) in ($13, $14, $15, $16, $17, $18, $19, $20, $21, $22, $23, $24, $25, $26, $27, $28, $29, $30, $31, $32, $33, $34, $35, $36, $37, $38, $39, $40, $41, $42, $43, $44, $45, $46, $47, $48, $49, $50, $51, $52, $53, $54, $55, $56, $57, $58, $59, $60, $61, $62, $63, $64, $65, $66, $67, $68, $69, $70, $71, $72, $73, $74, $75, $76, $77, $78, $79, $80, $81, $82, $83, $84, $85, $86, $87, $88, $89, $90, $91, $92, $93, $94, $95, $96, $97, $98, $99, $100, $101, $102, $103, $104, $105, $106, $107, $108, $109, $110, $111, $112, $113, $114, $115, $116, $117, $118, $119, $120, $121, $122, $123, $124, $125, $126, $127, $128, $129, $130, $131, $132, $133, $134, $135, $136, $137, $138, $139, $140, $141, $142, $143, $144, $145, $146, $147, $148, $149, $150, $151, $152, $153, $154, $155, $156, $157, $158, $159, $160, $161, $162, $163, $164, $165, $166, $167, $168, $169, $170, $171, $172, $173, $174, $175, $176, $177, $178, $179, $180, $181, $182, $183, $184, $185, $186, $187, $188, $189, $190, $191, $192, $193, $194, $195, $196, $197, $198, $199, $200, $201, $202, $203, $204, $205, $206, $207, $208, $209, $210, $211, $212, $213, $214, $215, $216, $217, $218, $219, $220, $221, $222, $223, $224, $225)) AND ($226 = 0 OR ccr.patternlengthbars <= $227)), results AS ( SELECT DISTINCT ON (symbolid) * FROM all_results WHERE (FALSE = $228 OR relevant = 1) AND ($229 = 0 OR age <= $230) ORDER BY symbolid, resultuid ) SELECT * from results ORDER BY identified DESC, length DESC;
Date: 2026-02-25 15:02:06 Duration: 0ms
4 30 1.66 GiB 3.83 MiB 150.43 MiB 56.53 MiB with rankedmt4 as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors ), last_feed_entry as ( select * from rankedmt4 where r = ? ), ok_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where status = ? ), earliest_entry_after_ok as ( select m.datafeedname, min(m.eventtimestamp) as eventtimestamp from mt4datafeederrors m left outer join ( select datafeedname, eventtimestamp from ok_entries where r = ?) oo on m.datafeedname = oo.datafeedname where m.eventtimestamp > coalesce(oo.eventtimestamp, ?::timestamp without time zone) group by m.datafeedname ), notified_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where notified is not null and notified <> ? ), broker as ( select *, row_number() over (partition by feedname order by brokerid) r from ( select distinct b.brokerid, b.name as brokername, dss.classname as feedname from downloadersymbolsettings dss inner join brokersymbollist bsl on dss.symbolid = bsl.symbolid inner join broker b on bsl.brokerid = b.brokerid where dss.enabled = ?) a ) select last.id, last.datafeedname, last.eventtimestamp, last.status, last.errordescription, last.serveraddress, last.username, note.notified, note.eventtimestamp, broker.brokername from last_feed_entry last inner join earliest_entry_after_ok after_ok on last.datafeedname = after_ok.datafeedname inner join broker on last.datafeedname = broker.feedname left outer join ok_entries ok on ok.datafeedname = last.datafeedname left outer join notified_entries note on note.datafeedname = last.datafeedname and note.r = ? where (ok.r is null or ok.r = ?) and last.datafeedname not in ( select distinct datafeedname from last_feed_entry where status = ?) and extract(epoch from (last.eventtimestamp - after_ok.eventtimestamp)) > ? * ? and last.eventtimestamp > current_timestamp - interval ? and (note.eventtimestamp is null or note.eventtimestamp < current_timestamp - interval ?) and last.eventtimestamp > current_timestamp - interval ? and broker.r = ?;-
with rankedmt4 as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors ), last_feed_entry as ( select * from rankedmt4 where r = 1 ), ok_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where status = 'OK' ), earliest_entry_after_ok as ( select m.datafeedname, min(m.eventtimestamp) as eventtimestamp from mt4datafeederrors m left outer join ( select datafeedname, eventtimestamp from ok_entries where r = 1) oo on m.datafeedname = oo.datafeedname where m.eventtimestamp > coalesce(oo.eventtimestamp, '1900-01-01'::timestamp without time zone) group by m.datafeedname ), notified_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where notified is not null and notified <> '' ), broker as ( select *, row_number() over (partition by feedname order by brokerid) r from ( select distinct b.brokerid, b.name as brokername, dss.classname as feedname from downloadersymbolsettings dss inner join brokersymbollist bsl on dss.symbolid = bsl.symbolid inner join broker b on bsl.brokerid = b.brokerid where dss.enabled = 1) a ) select last.id, last.datafeedname, last.eventtimestamp, last.status, last.errordescription, last.serveraddress, last.username, note.notified, note.eventtimestamp, broker.brokername from last_feed_entry last inner join earliest_entry_after_ok after_ok on last.datafeedname = after_ok.datafeedname inner join broker on last.datafeedname = broker.feedname left outer join ok_entries ok on ok.datafeedname = last.datafeedname left outer join notified_entries note on note.datafeedname = last.datafeedname and note.r = 1 where (ok.r is null or ok.r = 1) and last.datafeedname not in ( select distinct datafeedname from last_feed_entry where status = 'OK') and extract(epoch from (last.eventtimestamp - after_ok.eventtimestamp)) > 60 * 60 and last.eventtimestamp > current_timestamp - interval '1 day' and (note.eventtimestamp is null or note.eventtimestamp < current_timestamp - interval '10 hours') and last.eventtimestamp > current_timestamp - interval '1 hour' and broker.r = 1;
Date: 2026-02-25 15:00:06 Duration: 0ms
5 16 621.98 MiB 38.87 MiB 38.88 MiB 38.87 MiB update solr_relevance_old set new_hod_correct = sub.hod_correct, new_hod_percent = sub.hod_percent, new_hod_total = sub.hod_total, new_pattern_correct = sub.pattern_correct, new_pattern_percent = sub.pattern_percent, new_pattern_total = sub.pattern_total, new_percent = sub.percent, new_symbol_correct = sub.symbol_correct, new_symbol_percent = sub.symbol_percent, new_symbol_total = sub.symbol_total from ( select distinct resultuid, hod_correct, hod_percent, hod_total, hod, pattern_correct, pattern_percent, pattern_total, percent, symbol_correct, symbol_percent, symbol_total from whatshot_probability where type = ?) sub where result_uid = sub.resultuid;-
UPDATE solr_relevance_old SET new_hod_correct = sub.hod_correct, new_hod_percent = sub.hod_percent, new_hod_total = sub.hod_total, new_pattern_correct = sub.pattern_correct, new_pattern_percent = sub.pattern_percent, new_pattern_total = sub.pattern_total, new_percent = sub.percent, new_symbol_correct = sub.symbol_correct, new_symbol_percent = sub.symbol_percent, new_symbol_total = sub.symbol_total FROM ( select distinct resultuid, hod_correct, hod_percent, hod_total, hod, pattern_correct, pattern_percent, pattern_total, percent, symbol_correct, symbol_percent, symbol_total FROM whatshot_probability WHERE type = 'cp') sub WHERE result_uid = sub.resultuid;
Date: 2026-02-25 15:01:13 Duration: 0ms
6 16 1.18 GiB 74.61 MiB 75.51 MiB 75.34 MiB with max_ra as ( select resultuid from relevance_keylevels_results order by resultuid desc limit ?) update solr_relevance_old set newrelevant = sub.relevant, newage = sub.age from ( select so.uuid, case when ra.relevant is not null then ra.relevant when so.result_uid < max_ra.resultuid then ? else ? end as relevant, case when ra.age is not null then ra.age when so.result_uid < max_ra.resultuid then ? else ? end as age, so.result_uid from max_ra, solr_relevance_old so inner join keylevels_results k on so.result_uid = k.resultuid and so.uuid ilike ? inner join downloadersymbolsettings dss on k.symbolid = dss.symbolid left outer join relevance_keylevels_results ra on so.result_uid = ra.resultuid and so.uuid ilike ?) sub where solr_relevance_old.result_uid = sub.result_uid and solr_relevance_old.uuid ilike ?; update solr_relevance_old set newrelevant = ? where result_uid in ( select result_uid from solr_relevance_old s left outer join keylevels_results a on a.resultuid = s.result_uid where s.uuid ilike ? and a.resultuid is null); update solr_relevance_old set new_hod_correct = sub.hod_correct, new_hod_percent = sub.hod_percent, new_hod_total = sub.hod_total, new_pattern_correct = sub.pattern_correct, new_pattern_percent = sub.pattern_percent, new_pattern_total = sub.pattern_total, new_percent = sub.percent, new_symbol_correct = sub.symbol_correct, new_symbol_percent = sub.symbol_percent, new_symbol_total = sub.symbol_total from ( select distinct resultuid, hod_correct, hod_percent, hod_total, hod, pattern_correct, pattern_percent, pattern_total, percent, symbol_correct, symbol_percent, symbol_total from whatshot_probability where type in (...)) sub where result_uid = sub.resultuid;-
with max_ra as ( select resultuid from relevance_keylevels_results order by resultuid desc limit 1) update solr_relevance_old set newrelevant = sub.relevant, newage = sub.age from ( select so.uuid, case when ra.relevant is not null then ra.relevant when so.result_uid < max_ra.resultuid then 0 else 1 end as relevant, case when ra.age is not null then ra.age when so.result_uid < max_ra.resultuid then 11 else 0 end as age, so.result_uid from max_ra, solr_relevance_old so inner join keylevels_results k on so.result_uid = k.resultuid and so.uuid ilike 'kl_%' inner join downloadersymbolsettings dss on k.symbolid = dss.symbolid left outer join relevance_keylevels_results ra on so.result_uid = ra.resultuid and so.uuid ilike 'kl_%') sub where solr_relevance_old.result_uid = sub.result_uid and solr_relevance_old.uuid ilike 'kl_%'; update solr_relevance_old set newrelevant = 0 where result_uid in ( select result_uid from solr_relevance_old s left outer join keylevels_results a on a.resultuid = s.result_uid where s.uuid ilike 'kl_%' and a.resultuid is null); UPDATE solr_relevance_old SET new_hod_correct = sub.hod_correct, new_hod_percent = sub.hod_percent, new_hod_total = sub.hod_total, new_pattern_correct = sub.pattern_correct, new_pattern_percent = sub.pattern_percent, new_pattern_total = sub.pattern_total, new_percent = sub.percent, new_symbol_correct = sub.symbol_correct, new_symbol_percent = sub.symbol_percent, new_symbol_total = sub.symbol_total FROM ( select distinct resultuid, hod_correct, hod_percent, hod_total, hod, pattern_correct, pattern_percent, pattern_total, percent, symbol_correct, symbol_percent, symbol_total FROM whatshot_probability WHERE type in ('kl', 'ekl')) sub WHERE result_uid = sub.resultuid;
Date: 2026-02-25 15:01:16 Duration: 0ms
7 8 1.06 GiB 135.52 MiB 135.56 MiB 135.54 MiB select updateresultsmaterializedview ();-
select updateresultsmaterializedview ();
Date: 2026-02-25 15:02:12 Duration: 0ms
8 4 198.75 MiB 49.64 MiB 49.77 MiB 49.69 MiB select updateageforrelevantresults ();-
select updateageforrelevantresults ();
Date: 2026-02-25 15:02:05 Duration: 0ms
Queries generating the largest temporary files
Rank Size Query 1 150.43 MiB with rankedmt4 as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors ), last_feed_entry as ( select * from rankedmt4 where r = 1 ), ok_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where status = 'OK' ), earliest_entry_after_ok as ( select m.datafeedname, min(m.eventtimestamp) as eventtimestamp from mt4datafeederrors m left outer join ( select datafeedname, eventtimestamp from ok_entries where r = 1) oo on m.datafeedname = oo.datafeedname where m.eventtimestamp > coalesce(oo.eventtimestamp, '1900-01-01'::timestamp without time zone) group by m.datafeedname ), notified_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where notified is not null and notified <> '' ), broker as ( select *, row_number() over (partition by feedname order by brokerid) r from ( select distinct b.brokerid, b.name as brokername, dss.classname as feedname from downloadersymbolsettings dss inner join brokersymbollist bsl on dss.symbolid = bsl.symbolid inner join broker b on bsl.brokerid = b.brokerid where dss.enabled = 1) a ) select last.id, last.datafeedname, last.eventtimestamp, last.status, last.errordescription, last.serveraddress, last.username, note.notified, note.eventtimestamp, broker.brokername from last_feed_entry last inner join earliest_entry_after_ok after_ok on last.datafeedname = after_ok.datafeedname inner join broker on last.datafeedname = broker.feedname left outer join ok_entries ok on ok.datafeedname = last.datafeedname left outer join notified_entries note on note.datafeedname = last.datafeedname and note.r = 1 where (ok.r is null or ok.r = 1) and last.datafeedname not in ( select distinct datafeedname from last_feed_entry where status = 'OK') and extract(epoch from (last.eventtimestamp - after_ok.eventtimestamp)) > 60 * 60 and last.eventtimestamp > current_timestamp - interval '1 day' and (note.eventtimestamp is null or note.eventtimestamp < current_timestamp - interval '10 hours') and last.eventtimestamp > current_timestamp - interval '1 hour' and broker.r = 1;[ Date: 2026-02-25 15:30:07 ]
2 135.56 MiB select updateresultsmaterializedview ();[ Date: 2026-02-25 15:32:12 ]
3 135.55 MiB select updateresultsmaterializedview ();[ Date: 2026-02-25 15:47:16 ]
4 135.54 MiB select updateresultsmaterializedview ();[ Date: 2026-02-25 15:02:12 ]
5 135.54 MiB select updateresultsmaterializedview ();[ Date: 2026-02-25 15:50:33 ]
6 135.54 MiB select updateresultsmaterializedview ();[ Date: 2026-02-25 15:35:32 ]
7 135.54 MiB select updateresultsmaterializedview ();[ Date: 2026-02-25 15:17:10 ]
8 135.53 MiB select updateresultsmaterializedview ();[ Date: 2026-02-25 15:20:33 ]
9 135.52 MiB select updateresultsmaterializedview ();[ Date: 2026-02-25 15:05:33 ]
10 122.79 MiB with rankedmt4 as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors ), last_feed_entry as ( select * from rankedmt4 where r = 1 ), ok_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where status = 'OK' ), earliest_entry_after_ok as ( select m.datafeedname, min(m.eventtimestamp) as eventtimestamp from mt4datafeederrors m left outer join ( select datafeedname, eventtimestamp from ok_entries where r = 1) oo on m.datafeedname = oo.datafeedname where m.eventtimestamp > coalesce(oo.eventtimestamp, '1900-01-01'::timestamp without time zone) group by m.datafeedname ), notified_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where notified is not null and notified <> '' ), broker as ( select *, row_number() over (partition by feedname order by brokerid) r from ( select distinct b.brokerid, b.name as brokername, dss.classname as feedname from downloadersymbolsettings dss inner join brokersymbollist bsl on dss.symbolid = bsl.symbolid inner join broker b on bsl.brokerid = b.brokerid where dss.enabled = 1) a ) select last.id, last.datafeedname, last.eventtimestamp, last.status, last.errordescription, last.serveraddress, last.username, note.notified, note.eventtimestamp, broker.brokername from last_feed_entry last inner join earliest_entry_after_ok after_ok on last.datafeedname = after_ok.datafeedname inner join broker on last.datafeedname = broker.feedname left outer join ok_entries ok on ok.datafeedname = last.datafeedname left outer join notified_entries note on note.datafeedname = last.datafeedname and note.r = 1 where (ok.r is null or ok.r = 1) and last.datafeedname not in ( select distinct datafeedname from last_feed_entry where status = 'OK') and extract(epoch from (last.eventtimestamp - after_ok.eventtimestamp)) > 60 * 60 and last.eventtimestamp > current_timestamp - interval '1 day' and (note.eventtimestamp is null or note.eventtimestamp < current_timestamp - interval '10 hours') and last.eventtimestamp > current_timestamp - interval '1 hour' and broker.r = 1;[ Date: 2026-02-25 15:20:04 ]
11 114.87 MiB with rankedmt4 as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors ), last_feed_entry as ( select * from rankedmt4 where r = 1 ), ok_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where status = 'OK' ), earliest_entry_after_ok as ( select m.datafeedname, min(m.eventtimestamp) as eventtimestamp from mt4datafeederrors m left outer join ( select datafeedname, eventtimestamp from ok_entries where r = 1) oo on m.datafeedname = oo.datafeedname where m.eventtimestamp > coalesce(oo.eventtimestamp, '1900-01-01'::timestamp without time zone) group by m.datafeedname ), notified_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where notified is not null and notified <> '' ), broker as ( select *, row_number() over (partition by feedname order by brokerid) r from ( select distinct b.brokerid, b.name as brokername, dss.classname as feedname from downloadersymbolsettings dss inner join brokersymbollist bsl on dss.symbolid = bsl.symbolid inner join broker b on bsl.brokerid = b.brokerid where dss.enabled = 1) a ) select last.id, last.datafeedname, last.eventtimestamp, last.status, last.errordescription, last.serveraddress, last.username, note.notified, note.eventtimestamp, broker.brokername from last_feed_entry last inner join earliest_entry_after_ok after_ok on last.datafeedname = after_ok.datafeedname inner join broker on last.datafeedname = broker.feedname left outer join ok_entries ok on ok.datafeedname = last.datafeedname left outer join notified_entries note on note.datafeedname = last.datafeedname and note.r = 1 where (ok.r is null or ok.r = 1) and last.datafeedname not in ( select distinct datafeedname from last_feed_entry where status = 'OK') and extract(epoch from (last.eventtimestamp - after_ok.eventtimestamp)) > 60 * 60 and last.eventtimestamp > current_timestamp - interval '1 day' and (note.eventtimestamp is null or note.eventtimestamp < current_timestamp - interval '10 hours') and last.eventtimestamp > current_timestamp - interval '1 hour' and broker.r = 1;[ Date: 2026-02-25 15:00:04 ]
12 95.73 MiB with rankedmt4 as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors ), last_feed_entry as ( select * from rankedmt4 where r = 1 ), ok_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where status = 'OK' ), earliest_entry_after_ok as ( select m.datafeedname, min(m.eventtimestamp) as eventtimestamp from mt4datafeederrors m left outer join ( select datafeedname, eventtimestamp from ok_entries where r = 1) oo on m.datafeedname = oo.datafeedname where m.eventtimestamp > coalesce(oo.eventtimestamp, '1900-01-01'::timestamp without time zone) group by m.datafeedname ), notified_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where notified is not null and notified <> '' ), broker as ( select *, row_number() over (partition by feedname order by brokerid) r from ( select distinct b.brokerid, b.name as brokername, dss.classname as feedname from downloadersymbolsettings dss inner join brokersymbollist bsl on dss.symbolid = bsl.symbolid inner join broker b on bsl.brokerid = b.brokerid where dss.enabled = 1) a ) select last.id, last.datafeedname, last.eventtimestamp, last.status, last.errordescription, last.serveraddress, last.username, note.notified, note.eventtimestamp, broker.brokername from last_feed_entry last inner join earliest_entry_after_ok after_ok on last.datafeedname = after_ok.datafeedname inner join broker on last.datafeedname = broker.feedname left outer join ok_entries ok on ok.datafeedname = last.datafeedname left outer join notified_entries note on note.datafeedname = last.datafeedname and note.r = 1 where (ok.r is null or ok.r = 1) and last.datafeedname not in ( select distinct datafeedname from last_feed_entry where status = 'OK') and extract(epoch from (last.eventtimestamp - after_ok.eventtimestamp)) > 60 * 60 and last.eventtimestamp > current_timestamp - interval '1 day' and (note.eventtimestamp is null or note.eventtimestamp < current_timestamp - interval '10 hours') and last.eventtimestamp > current_timestamp - interval '1 hour' and broker.r = 1;[ Date: 2026-02-25 15:40:05 ]
13 88.67 MiB with rankedmt4 as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors ), last_feed_entry as ( select * from rankedmt4 where r = 1 ), ok_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where status = 'OK' ), earliest_entry_after_ok as ( select m.datafeedname, min(m.eventtimestamp) as eventtimestamp from mt4datafeederrors m left outer join ( select datafeedname, eventtimestamp from ok_entries where r = 1) oo on m.datafeedname = oo.datafeedname where m.eventtimestamp > coalesce(oo.eventtimestamp, '1900-01-01'::timestamp without time zone) group by m.datafeedname ), notified_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where notified is not null and notified <> '' ), broker as ( select *, row_number() over (partition by feedname order by brokerid) r from ( select distinct b.brokerid, b.name as brokername, dss.classname as feedname from downloadersymbolsettings dss inner join brokersymbollist bsl on dss.symbolid = bsl.symbolid inner join broker b on bsl.brokerid = b.brokerid where dss.enabled = 1) a ) select last.id, last.datafeedname, last.eventtimestamp, last.status, last.errordescription, last.serveraddress, last.username, note.notified, note.eventtimestamp, broker.brokername from last_feed_entry last inner join earliest_entry_after_ok after_ok on last.datafeedname = after_ok.datafeedname inner join broker on last.datafeedname = broker.feedname left outer join ok_entries ok on ok.datafeedname = last.datafeedname left outer join notified_entries note on note.datafeedname = last.datafeedname and note.r = 1 where (ok.r is null or ok.r = 1) and last.datafeedname not in ( select distinct datafeedname from last_feed_entry where status = 'OK') and extract(epoch from (last.eventtimestamp - after_ok.eventtimestamp)) > 60 * 60 and last.eventtimestamp > current_timestamp - interval '1 day' and (note.eventtimestamp is null or note.eventtimestamp < current_timestamp - interval '10 hours') and last.eventtimestamp > current_timestamp - interval '1 hour' and broker.r = 1;[ Date: 2026-02-25 15:10:05 ]
14 84.37 MiB with rankedmt4 as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors ), last_feed_entry as ( select * from rankedmt4 where r = 1 ), ok_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where status = 'OK' ), earliest_entry_after_ok as ( select m.datafeedname, min(m.eventtimestamp) as eventtimestamp from mt4datafeederrors m left outer join ( select datafeedname, eventtimestamp from ok_entries where r = 1) oo on m.datafeedname = oo.datafeedname where m.eventtimestamp > coalesce(oo.eventtimestamp, '1900-01-01'::timestamp without time zone) group by m.datafeedname ), notified_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where notified is not null and notified <> '' ), broker as ( select *, row_number() over (partition by feedname order by brokerid) r from ( select distinct b.brokerid, b.name as brokername, dss.classname as feedname from downloadersymbolsettings dss inner join brokersymbollist bsl on dss.symbolid = bsl.symbolid inner join broker b on bsl.brokerid = b.brokerid where dss.enabled = 1) a ) select last.id, last.datafeedname, last.eventtimestamp, last.status, last.errordescription, last.serveraddress, last.username, note.notified, note.eventtimestamp, broker.brokername from last_feed_entry last inner join earliest_entry_after_ok after_ok on last.datafeedname = after_ok.datafeedname inner join broker on last.datafeedname = broker.feedname left outer join ok_entries ok on ok.datafeedname = last.datafeedname left outer join notified_entries note on note.datafeedname = last.datafeedname and note.r = 1 where (ok.r is null or ok.r = 1) and last.datafeedname not in ( select distinct datafeedname from last_feed_entry where status = 'OK') and extract(epoch from (last.eventtimestamp - after_ok.eventtimestamp)) > 60 * 60 and last.eventtimestamp > current_timestamp - interval '1 day' and (note.eventtimestamp is null or note.eventtimestamp < current_timestamp - interval '10 hours') and last.eventtimestamp > current_timestamp - interval '1 hour' and broker.r = 1;[ Date: 2026-02-25 15:40:06 ]
15 84.17 MiB with rankedmt4 as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors ), last_feed_entry as ( select * from rankedmt4 where r = 1 ), ok_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where status = 'OK' ), earliest_entry_after_ok as ( select m.datafeedname, min(m.eventtimestamp) as eventtimestamp from mt4datafeederrors m left outer join ( select datafeedname, eventtimestamp from ok_entries where r = 1) oo on m.datafeedname = oo.datafeedname where m.eventtimestamp > coalesce(oo.eventtimestamp, '1900-01-01'::timestamp without time zone) group by m.datafeedname ), notified_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where notified is not null and notified <> '' ), broker as ( select *, row_number() over (partition by feedname order by brokerid) r from ( select distinct b.brokerid, b.name as brokername, dss.classname as feedname from downloadersymbolsettings dss inner join brokersymbollist bsl on dss.symbolid = bsl.symbolid inner join broker b on bsl.brokerid = b.brokerid where dss.enabled = 1) a ) select last.id, last.datafeedname, last.eventtimestamp, last.status, last.errordescription, last.serveraddress, last.username, note.notified, note.eventtimestamp, broker.brokername from last_feed_entry last inner join earliest_entry_after_ok after_ok on last.datafeedname = after_ok.datafeedname inner join broker on last.datafeedname = broker.feedname left outer join ok_entries ok on ok.datafeedname = last.datafeedname left outer join notified_entries note on note.datafeedname = last.datafeedname and note.r = 1 where (ok.r is null or ok.r = 1) and last.datafeedname not in ( select distinct datafeedname from last_feed_entry where status = 'OK') and extract(epoch from (last.eventtimestamp - after_ok.eventtimestamp)) > 60 * 60 and last.eventtimestamp > current_timestamp - interval '1 day' and (note.eventtimestamp is null or note.eventtimestamp < current_timestamp - interval '10 hours') and last.eventtimestamp > current_timestamp - interval '1 hour' and broker.r = 1;[ Date: 2026-02-25 15:00:06 ]
16 76.08 MiB with rankedmt4 as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors ), last_feed_entry as ( select * from rankedmt4 where r = 1 ), ok_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where status = 'OK' ), earliest_entry_after_ok as ( select m.datafeedname, min(m.eventtimestamp) as eventtimestamp from mt4datafeederrors m left outer join ( select datafeedname, eventtimestamp from ok_entries where r = 1) oo on m.datafeedname = oo.datafeedname where m.eventtimestamp > coalesce(oo.eventtimestamp, '1900-01-01'::timestamp without time zone) group by m.datafeedname ), notified_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where notified is not null and notified <> '' ), broker as ( select *, row_number() over (partition by feedname order by brokerid) r from ( select distinct b.brokerid, b.name as brokername, dss.classname as feedname from downloadersymbolsettings dss inner join brokersymbollist bsl on dss.symbolid = bsl.symbolid inner join broker b on bsl.brokerid = b.brokerid where dss.enabled = 1) a ) select last.id, last.datafeedname, last.eventtimestamp, last.status, last.errordescription, last.serveraddress, last.username, note.notified, note.eventtimestamp, broker.brokername from last_feed_entry last inner join earliest_entry_after_ok after_ok on last.datafeedname = after_ok.datafeedname inner join broker on last.datafeedname = broker.feedname left outer join ok_entries ok on ok.datafeedname = last.datafeedname left outer join notified_entries note on note.datafeedname = last.datafeedname and note.r = 1 where (ok.r is null or ok.r = 1) and last.datafeedname not in ( select distinct datafeedname from last_feed_entry where status = 'OK') and extract(epoch from (last.eventtimestamp - after_ok.eventtimestamp)) > 60 * 60 and last.eventtimestamp > current_timestamp - interval '1 day' and (note.eventtimestamp is null or note.eventtimestamp < current_timestamp - interval '10 hours') and last.eventtimestamp > current_timestamp - interval '1 hour' and broker.r = 1;[ Date: 2026-02-25 15:50:05 ]
17 75.51 MiB with max_ra as ( select resultuid from relevance_keylevels_results order by resultuid desc limit 1) update solr_relevance_old set newrelevant = sub.relevant, newage = sub.age from ( select so.uuid, case when ra.relevant is not null then ra.relevant when so.result_uid < max_ra.resultuid then 0 else 1 end as relevant, case when ra.age is not null then ra.age when so.result_uid < max_ra.resultuid then 11 else 0 end as age, so.result_uid from max_ra, solr_relevance_old so inner join keylevels_results k on so.result_uid = k.resultuid and so.uuid ilike 'kl_%' inner join downloadersymbolsettings dss on k.symbolid = dss.symbolid left outer join relevance_keylevels_results ra on so.result_uid = ra.resultuid and so.uuid ilike 'kl_%') sub where solr_relevance_old.result_uid = sub.result_uid and solr_relevance_old.uuid ilike 'kl_%'; update solr_relevance_old set newrelevant = 0 where result_uid in ( select result_uid from solr_relevance_old s left outer join keylevels_results a on a.resultuid = s.result_uid where s.uuid ilike 'kl_%' and a.resultuid is null); UPDATE solr_relevance_old SET new_hod_correct = sub.hod_correct, new_hod_percent = sub.hod_percent, new_hod_total = sub.hod_total, new_pattern_correct = sub.pattern_correct, new_pattern_percent = sub.pattern_percent, new_pattern_total = sub.pattern_total, new_percent = sub.percent, new_symbol_correct = sub.symbol_correct, new_symbol_percent = sub.symbol_percent, new_symbol_total = sub.symbol_total FROM ( select distinct resultuid, hod_correct, hod_percent, hod_total, hod, pattern_correct, pattern_percent, pattern_total, percent, symbol_correct, symbol_percent, symbol_total FROM whatshot_probability WHERE type in ('kl', 'ekl')) sub WHERE result_uid = sub.resultuid;[ Date: 2026-02-25 15:11:16 ]
18 75.51 MiB with max_ra as ( select resultuid from relevance_keylevels_results order by resultuid desc limit 1) update solr_relevance_old set newrelevant = sub.relevant, newage = sub.age from ( select so.uuid, case when ra.relevant is not null then ra.relevant when so.result_uid < max_ra.resultuid then 0 else 1 end as relevant, case when ra.age is not null then ra.age when so.result_uid < max_ra.resultuid then 11 else 0 end as age, so.result_uid from max_ra, solr_relevance_old so inner join keylevels_results k on so.result_uid = k.resultuid and so.uuid ilike 'kl_%' inner join downloadersymbolsettings dss on k.symbolid = dss.symbolid left outer join relevance_keylevels_results ra on so.result_uid = ra.resultuid and so.uuid ilike 'kl_%') sub where solr_relevance_old.result_uid = sub.result_uid and solr_relevance_old.uuid ilike 'kl_%'; update solr_relevance_old set newrelevant = 0 where result_uid in ( select result_uid from solr_relevance_old s left outer join keylevels_results a on a.resultuid = s.result_uid where s.uuid ilike 'kl_%' and a.resultuid is null); UPDATE solr_relevance_old SET new_hod_correct = sub.hod_correct, new_hod_percent = sub.hod_percent, new_hod_total = sub.hod_total, new_pattern_correct = sub.pattern_correct, new_pattern_percent = sub.pattern_percent, new_pattern_total = sub.pattern_total, new_percent = sub.percent, new_symbol_correct = sub.symbol_correct, new_symbol_percent = sub.symbol_percent, new_symbol_total = sub.symbol_total FROM ( select distinct resultuid, hod_correct, hod_percent, hod_total, hod, pattern_correct, pattern_percent, pattern_total, percent, symbol_correct, symbol_percent, symbol_total FROM whatshot_probability WHERE type in ('kl', 'ekl')) sub WHERE result_uid = sub.resultuid;[ Date: 2026-02-25 15:16:16 ]
19 75.51 MiB with max_ra as ( select resultuid from relevance_keylevels_results order by resultuid desc limit 1) update solr_relevance_old set newrelevant = sub.relevant, newage = sub.age from ( select so.uuid, case when ra.relevant is not null then ra.relevant when so.result_uid < max_ra.resultuid then 0 else 1 end as relevant, case when ra.age is not null then ra.age when so.result_uid < max_ra.resultuid then 11 else 0 end as age, so.result_uid from max_ra, solr_relevance_old so inner join keylevels_results k on so.result_uid = k.resultuid and so.uuid ilike 'kl_%' inner join downloadersymbolsettings dss on k.symbolid = dss.symbolid left outer join relevance_keylevels_results ra on so.result_uid = ra.resultuid and so.uuid ilike 'kl_%') sub where solr_relevance_old.result_uid = sub.result_uid and solr_relevance_old.uuid ilike 'kl_%'; update solr_relevance_old set newrelevant = 0 where result_uid in ( select result_uid from solr_relevance_old s left outer join keylevels_results a on a.resultuid = s.result_uid where s.uuid ilike 'kl_%' and a.resultuid is null); UPDATE solr_relevance_old SET new_hod_correct = sub.hod_correct, new_hod_percent = sub.hod_percent, new_hod_total = sub.hod_total, new_pattern_correct = sub.pattern_correct, new_pattern_percent = sub.pattern_percent, new_pattern_total = sub.pattern_total, new_percent = sub.percent, new_symbol_correct = sub.symbol_correct, new_symbol_percent = sub.symbol_percent, new_symbol_total = sub.symbol_total FROM ( select distinct resultuid, hod_correct, hod_percent, hod_total, hod, pattern_correct, pattern_percent, pattern_total, percent, symbol_correct, symbol_percent, symbol_total FROM whatshot_probability WHERE type in ('kl', 'ekl')) sub WHERE result_uid = sub.resultuid;[ Date: 2026-02-25 15:18:15 ]
20 75.51 MiB with max_ra as ( select resultuid from relevance_keylevels_results order by resultuid desc limit 1) update solr_relevance_old set newrelevant = sub.relevant, newage = sub.age from ( select so.uuid, case when ra.relevant is not null then ra.relevant when so.result_uid < max_ra.resultuid then 0 else 1 end as relevant, case when ra.age is not null then ra.age when so.result_uid < max_ra.resultuid then 11 else 0 end as age, so.result_uid from max_ra, solr_relevance_old so inner join keylevels_results k on so.result_uid = k.resultuid and so.uuid ilike 'kl_%' inner join downloadersymbolsettings dss on k.symbolid = dss.symbolid left outer join relevance_keylevels_results ra on so.result_uid = ra.resultuid and so.uuid ilike 'kl_%') sub where solr_relevance_old.result_uid = sub.result_uid and solr_relevance_old.uuid ilike 'kl_%'; update solr_relevance_old set newrelevant = 0 where result_uid in ( select result_uid from solr_relevance_old s left outer join keylevels_results a on a.resultuid = s.result_uid where s.uuid ilike 'kl_%' and a.resultuid is null); UPDATE solr_relevance_old SET new_hod_correct = sub.hod_correct, new_hod_percent = sub.hod_percent, new_hod_total = sub.hod_total, new_pattern_correct = sub.pattern_correct, new_pattern_percent = sub.pattern_percent, new_pattern_total = sub.pattern_total, new_percent = sub.percent, new_symbol_correct = sub.symbol_correct, new_symbol_percent = sub.symbol_percent, new_symbol_total = sub.symbol_total FROM ( select distinct resultuid, hod_correct, hod_percent, hod_total, hod, pattern_correct, pattern_percent, pattern_total, percent, symbol_correct, symbol_percent, symbol_total FROM whatshot_probability WHERE type in ('kl', 'ekl')) sub WHERE result_uid = sub.resultuid;[ Date: 2026-02-25 15:20:18 ]
-
Vacuums
Vacuums / Analyzes Distribution
Key values
- 0 sec Highest CPU-cost vacuum
Table
Database - Date
- 0 sec Highest CPU-cost analyze
Table
Database - Date
Analyzes per table
Key values
- public.solr_relevance_old (16) Main table analyzed (database acaweb_fx)
- 59 analyzes Total
Table Number of analyzes acaweb_fx.public.solr_relevance_old 16 acaweb_fx.pg_catalog.pg_attribute 6 acaweb_fx.pg_catalog.pg_class 6 acaweb_fx.public.relevance_keylevels_results 4 acaweb_fx.public.relevance_fibonacci_results 4 acaweb_fx.pg_catalog.pg_type 4 acaweb_fx.public.relevance_autochartist_results 4 acaweb_fx.public.datafeeds_latestrun 3 acaweb_fx.pg_catalog.pg_index 2 acaweb_fx.public.autochartist_symbolupdates 2 acaweb_fx.public.latest_t15_candle_view 2 acaweb_fx.pg_catalog.pg_depend 2 acaweb_fx.public.latest_candle_datetime_per_receng 1 acaweb_fx.public.relevance_consecutivecandles_results 1 acaweb_fx.public.symbollatestupdatetime 1 acaweb_fx.public.bigmovement_results_underlying 1 Total 59 Vacuums per table
Key values
- public.solr_relevance_old (26) Main table vacuumed on database acaweb_fx
- 54 vacuums Total
Index Buffer usage Skipped WAL usage Table Vacuums scans hits misses dirtied pins frozen records full page bytes acaweb_fx.public.solr_relevance_old 26 20 18,696 0 82 0 259 10,325 16 1,800,768 acaweb_fx.pg_catalog.pg_attribute 4 4 3,281 0 613 0 268 1,311 481 2,637,930 acaweb_fx.pg_catalog.pg_type 3 3 475 0 96 0 0 202 66 363,272 acaweb_fx.public.datafeeds_latestrun 3 0 360 0 4 0 0 37 6 36,692 acaweb_fx.public.relevance_fibonacci_results 3 3 3,337 0 95 0 166 360 127 441,357 acaweb_fx.pg_toast.pg_toast_2619 2 2 295 0 55 0 0 179 50 188,727 acaweb_fx.pg_catalog.pg_statistic 2 2 2,007 0 344 0 1,145 973 310 1,279,171 acaweb_fx.public.latest_t15_candle_view 2 2 132 0 5 0 0 12 2 17,221 acaweb_fx.public.relevance_keylevels_results 2 2 7,657 0 383 4 203 1,818 1,160 3,611,820 acaweb_fx.pg_catalog.pg_class 2 2 926 0 100 0 0 270 92 489,262 acaweb_fx.public.relevance_autochartist_results 2 2 6,587 0 148 2 498 1,068 134 440,133 acaweb_fx.pg_catalog.pg_index 1 1 108 0 13 0 0 28 11 84,864 acaweb_fx.public.relevance_consecutivecandles_results 1 1 75 0 4 0 0 22 2 18,215 socialmedia.public.processstatevariables 1 1 204 0 41 0 0 107 30 174,380 Total 54 45 44,140 20,138 1,983 6 2,539 16,712 2,487 11,583,812 Tuples removed per table
Key values
- public.solr_relevance_old (87681) Main table with removed tuples on database acaweb_fx
- 98200 tuples Total removed
Index Tuples Pages Table Vacuums scans removed remain not yet removable removed remain acaweb_fx.public.solr_relevance_old 26 20 87,681 210,019 55,253 0 5,257 acaweb_fx.pg_catalog.pg_attribute 4 4 5,665 46,272 3,960 22 1,106 acaweb_fx.public.relevance_keylevels_results 2 2 1,325 24,669 0 0 558 acaweb_fx.pg_catalog.pg_statistic 2 2 1,193 7,394 0 0 2,388 acaweb_fx.pg_catalog.pg_type 3 3 720 4,696 352 0 132 acaweb_fx.public.relevance_autochartist_results 2 2 608 17,448 558 0 760 acaweb_fx.pg_catalog.pg_class 2 2 236 3,395 95 0 300 acaweb_fx.public.datafeeds_latestrun 3 0 189 42 0 0 48 acaweb_fx.public.relevance_fibonacci_results 3 3 130 4,351 260 0 306 acaweb_fx.pg_toast.pg_toast_2619 2 2 121 355 27 6 94 acaweb_fx.public.relevance_consecutivecandles_results 1 1 110 295 0 0 7 acaweb_fx.public.latest_t15_candle_view 2 2 107 33 9 0 2 socialmedia.public.processstatevariables 1 1 98 389 0 0 43 acaweb_fx.pg_catalog.pg_index 1 1 17 825 12 0 22 Total 54 45 98,200 320,183 60,526 28 11,023 Pages removed per table
Key values
- pg_catalog.pg_attribute (22) Main table with removed pages on database acaweb_fx
- 28 pages Total removed
Table Number of vacuums Index scans Tuples removed Pages removed acaweb_fx.pg_catalog.pg_attribute 4 4 5665 22 acaweb_fx.pg_toast.pg_toast_2619 2 2 121 6 acaweb_fx.pg_catalog.pg_index 1 1 17 0 acaweb_fx.pg_catalog.pg_type 3 3 720 0 acaweb_fx.public.datafeeds_latestrun 3 0 189 0 acaweb_fx.pg_catalog.pg_statistic 2 2 1193 0 acaweb_fx.public.relevance_consecutivecandles_results 1 1 110 0 socialmedia.public.processstatevariables 1 1 98 0 acaweb_fx.public.latest_t15_candle_view 2 2 107 0 acaweb_fx.public.relevance_keylevels_results 2 2 1325 0 acaweb_fx.pg_catalog.pg_class 2 2 236 0 acaweb_fx.public.solr_relevance_old 26 20 87681 0 acaweb_fx.public.relevance_autochartist_results 2 2 608 0 acaweb_fx.public.relevance_fibonacci_results 3 3 130 0 Total 54 45 98,200 28 Autovacuum Activity
↑ Back to the top of the Autovacuum Activity tableDay Hour VACUUMs ANALYZEs Feb 25 15 54 59 - 0 sec Highest CPU-cost vacuum
-
Locks
Locks by types
Key values
- unknown Main Lock Type
- 0 locks Total
Most frequent waiting queries (N)
Rank Count Total time Min time Max time Avg duration Query NO DATASET
Queries that waited the most
Rank Wait time Query NO DATASET
-
Queries
Queries by type
Key values
- 73,394 Total read queries
- 33,276 Total write queries
Queries by database
Key values
- unknown Main database
- 222,106 Requests
- 2h32m56s (unknown)
- Main time consuming database
Database Request type Count Duration acaweb_fx Total 867 0ms copy from 80 0ms copy to 26 0ms cte 104 0ms ddl 16 0ms delete 16 0ms others 183 0ms select 72 0ms tcl 332 0ms update 38 0ms socialmedia Total 91 0ms others 7 0ms select 76 0ms tcl 8 0ms unknown Total 222,106 2h32m56s copy from 16 0ms cte 5,518 0ms insert 24,088 0ms others 4,875 0ms select 73,246 0ms tcl 430 0ms update 2,604 0ms Queries by user
Key values
- unknown Main user
- 222,106 Requests
User Request type Count Duration postgres Total 958 0ms copy from 80 0ms copy to 26 0ms cte 104 0ms ddl 16 0ms delete 16 0ms others 190 0ms select 148 0ms tcl 340 0ms update 38 0ms unknown Total 222,106 2h32m56s copy from 16 0ms cte 5,518 0ms insert 24,088 0ms others 4,875 0ms select 73,246 0ms tcl 430 0ms update 2,604 0ms Duration by user
Key values
- 2h32m56s (unknown) Main time consuming user
User Request type Count Duration postgres Total 958 0ms copy from 80 0ms copy to 26 0ms cte 104 0ms ddl 16 0ms delete 16 0ms others 190 0ms select 148 0ms tcl 340 0ms update 38 0ms unknown Total 222,106 2h32m56s copy from 16 0ms cte 5,518 0ms insert 24,088 0ms others 4,875 0ms select 73,246 0ms tcl 430 0ms update 2,604 0ms Queries by host
Key values
- unknown Main host
- 223,064 Requests
- 2h32m56s (unknown)
- Main time consuming host
Queries by application
Key values
- unknown Main application
- 222,708 Requests
- 2h32m56s (unknown)
- Main time consuming application
Number of cancelled queries
Key values
- 0 per second Cancelled query Peak
- 2026-02-25 15:35:16 Date
Number of cancelled queries (5 minutes period)
NO DATASET
-
Top Queries
Histogram of query times
Key values
- 62,610 0-1ms duration
Slowest individual queries
Rank Duration Query NO DATASET
Time consuming queries
Rank Total duration Times executed Min duration Max duration Avg duration Query 1 0ms 26 0ms 0ms 0ms select key, value from datasources ds inner join datasourceparams dsp on ds.id = dsp.datasourceid where ds.name = ?;Times Reported Time consuming queries #1
Day Hour Count Duration Avg duration Feb 25 15 26 0ms 0ms 2 0ms 303 0ms 0ms 0ms with rar_max as ( select resultuid from relevance_consecutivecandles_results order by resultuid desc limit ? ), all_results as ( select ccr.resultuid as resultuid, ccr.direction as direction, s.exchange as exchange, s.symbolid as symbolid, coalesce(bim.code, s.symbol) as symbol_code, s.longname as symbol_name, s.timegranularity as interval, ccr.patternendtime as identified, dtt.timezone as timezone, ccr.qtyconsecutivecandles as length, g.basegroupname, case when rcr.age is not null then rcr.age when ccr.resultuid <= rm.resultuid then ? else ? end as age, case when rcr.relevant is not null then rcr.relevant when ccr.resultuid <= rm.resultuid then ? else ? end as relevant, cps.pip, newlevels.filtered from consecutivecandles_results ccr inner join brokersymbollist bsl on bsl.brokerid = ? and bsl.symbolid = ccr.symbolid inner join symbols s on ccr.symbolid = s.symbolid and s.nonliquid = ? inner join downloadersymbolsettings dss on ccr.symbolid = dss.symbolid inner join datafeedstimetable dtt on dss.classname = dtt.classname and dtt.dayofweek = ? inner join symbolgroup sg on ccr.symbolid = sg.symbolid inner join groups g on sg.groupid = g.groupid inner join brokergroups bg on g.groupid = bg.groupid and bsl.brokerid = bg.brokerid inner join rar_max rm on ? = ? left outer join relevance_consecutivecandles_results rcr on rcr.resultuid = ccr.resultuid left join currencypips cps on cps.symbol = s.symbol left outer join brokerinstrumentmap bim on dss.datafeedinstrumentid = bim.datafeedinstrumentid and bim.brokerid = bsl.brokerid and bim.type = ? left join lateral calc_cc_signal_filter (ccr.resultuid) newlevels on true where ccr.gmttimefound > now() - interval ? and s.deleted = ? and (ccr.simulation = ? or ccr.simulation is null) and (? = ? or s.timegranularity in (...)) and (? = ? or s.exchange in (...)) and (? = ? or coalesce(bim.code, s.symbol) in (...)) and (? = ? or ccr.patternlengthbars <= ?) ), results as ( select distinct on (symbolid) * from all_results where (false = ? or relevant = ?) and (? = ? or age <= ?) order by symbolid, resultuid ) select * from results order by identified desc, length desc;Times Reported Time consuming queries #2
Day Hour Count Duration Avg duration Feb 25 15 303 0ms 0ms 3 0ms 1 0ms 0ms 0ms select count(*) from "public"."processes" left outer join "public"."brokers" "LT?" on "LT?"."id" = "public"."processes"."brokerid" left outer join "public"."contenttypes" "LT?" on "LT?"."id" = "public"."processes"."contenttypeid" where (brokerid = ?);Times Reported Time consuming queries #3
Day Hour Count Duration Avg duration Feb 25 15 1 0ms 0ms 4 0ms 2 0ms 0ms 0ms select trumpetsymbolid as sid, trumpettimegranularity as tg from brokersymbollist bsl left join powerstats_symboldata psd on bsl.symbolid = psd.symbolid left join downloadersymbolsettings dss on psd.symbolid = dss.symbolid left join symbols s on dss.symbolid = s.symbolid left outer join brokerinstrumentmap bdfi on code = ? and bdfi.brokerid = ? and dss.datafeedinstrumentid = bdfi.datafeedinstrumentid where (code = ? or s.symbol = ?) and bsl.brokerid = ? and dss.classname <> ? group by trumpetsymbolid, trumpettimegranularity;Times Reported Time consuming queries #4
Day Hour Count Duration Avg duration Feb 25 15 2 0ms 0ms 5 0ms 2,058 0ms 0ms 0ms insert into t60 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) values (?, ?, ?, ?, ?, ?, ?, ?, ?, ?) on conflict (pricedatetime, symbolid) do update set open = ?, high = ?, low = ?, close = ?, volume = ?, bsf = ?, sastdatetimewritten = ?, sastdatetimereceived = ?;Times Reported Time consuming queries #5
Day Hour Count Duration Avg duration Feb 25 15 2,058 0ms 0ms 6 0ms 74 0ms 0ms 0ms with rar_max as ( select resultuid from relevance_bigmovement_results order by resultuid desc limit ? ) select bmr.symbolid, patternstarttime, patternendtime, timegranularity, ? as direction, case when bmr.old_resultuid = ? then bmr.old_resultuid else bmr.resultuid end as uid, s.exchange, s.symbol, s.longname, s.shortname, dtt.timezone, bmr.patternmovement, bmr.statisticalmovement, bmr.fromprice, bmr.toprice, bmr.percentile, bmr.patternlengthbars, case when rbr.age is not null then rbr.age when bmr.resultuid <= rm.resultuid then ? else ? end as age, case when rbr.relevant is not null then rbr.relevant when bmr.resultuid <= rm.resultuid then ? else ? end as relevant, cps.pip from bigmovement_results bmr inner join downloadersymbolsettings dss on bmr.symbolid = dss.symbolid inner join datafeedstimetable dtt on dss.classname = dtt.classname inner join symbols s on bmr.symbolid = s.symbolid inner join rar_max rm on ? = ? left outer join relevance_bigmovement_results rbr on rbr.resultuid = bmr.resultuid left join currencypips cps on cps.symbol = s.symbol where (bmr.old_resultuid = ? or bmr.resultuid = ?) and dtt.dayofweek = ?;Times Reported Time consuming queries #6
Day Hour Count Duration Avg duration Feb 25 15 74 0ms 0ms 7 0ms 4 0ms 0ms 0ms select count(*) from datafeeds_latestrun where feedname ilike ? and ((latestrxtime > current_timestamp - interval ? and latestdbwritetime > current_timestamp - interval ?) or (latestdbwritetime > current_timestamp - interval ? and lateststartuptime > current_timestamp - interval ?));Times Reported Time consuming queries #7
Day Hour Count Duration Avg duration Feb 25 15 4 0ms 0ms 8 0ms 4 0ms 0ms 0ms select updaterelevantforrelevantresults ();Times Reported Time consuming queries #8
Day Hour Count Duration Avg duration Feb 25 15 4 0ms 0ms 9 0ms 114 0ms 0ms 0ms select case when a.old_resultuid = ? then a.old_resultuid else a.resultuid end as resultuid, s.symbol, timegranularity as interval, direction as direction, patternendtime as patternendtime, patternstartprice as psp, patternendprice as pep, target03 as t03, target16 as t16, patternlengthbars as length, p.patternname as patternname, dtt.timezone, cps.pip from fibonacci_results a inner join downloadersymbolsettings dss on a.symbolid = dss.symbolid inner join datafeedstimetable dtt on dss.classname = dtt.classname inner join symbols s on a.symbolid = s.symbolid inner join fibonaccipatterns p on a.pattern = p.patternname left join currencypips cps on cps.symbol = s.symbol where (a.old_resultuid = ? or a.resultuid = ?) and dtt.dayofweek = ?;Times Reported Time consuming queries #9
Day Hour Count Duration Avg duration Feb 25 15 114 0ms 0ms 10 0ms 7 0ms 0ms 0ms set datestyle = iso;Times Reported Time consuming queries #10
Day Hour Count Duration Avg duration Feb 25 15 7 0ms 0ms 11 0ms 2,123 0ms 0ms 0ms update patternresultsrelevance set relevant = ?, saxo_relevant = ?, notrelevantpricedatetime = ?, reason = ? where uniqueindex = ? and relevant = ?;Times Reported Time consuming queries #11
Day Hour Count Duration Avg duration Feb 25 15 2,123 0ms 0ms 12 0ms 9 0ms 0ms 0ms insert into resultmedia (processresultsid, type, name, filename) values (?, ?, ?, ?) returning id;Times Reported Time consuming queries #12
Day Hour Count Duration Avg duration Feb 25 15 9 0ms 0ms 13 0ms 7 0ms 0ms 0ms set client_encoding to ?;Times Reported Time consuming queries #13
Day Hour Count Duration Avg duration Feb 25 15 7 0ms 0ms 14 0ms 1 0ms 0ms 0ms select id from processes where id = ?;Times Reported Time consuming queries #14
Day Hour Count Duration Avg duration Feb 25 15 1 0ms 0ms 15 0ms 2 0ms 0ms 0ms select "public"."executions"."id" AS "id", "public"."executions"."processid" AS "processid", "public"."executions"."executiondate" AS "executiondate", "public"."executions"."errorcount" AS "errorcount", "public"."executions"."warningcount" AS "warningcount", "public"."executions"."isrunning" AS "isrunning", "public"."executions"."response" AS "response", "public"."executions"."live" AS "live", "public"."executions"."has_results" AS "has_results", "LT?"."id" AS "LA?" from "public"."executions" left outer join "public"."processes" "LT?" on "LT?"."id" = "public"."executions"."processid" where (processid = ?) order by "public"."executions"."id" desc limit ? offset ?;Times Reported Time consuming queries #15
Day Hour Count Duration Avg duration Feb 25 15 2 0ms 0ms 16 0ms 239 0ms 0ms 0ms select v.datname, c.relname, v.phase, v.heap_blks_total, v.heap_blks_scanned, v.heap_blks_vacuumed, v.index_vacuum_count, v.max_dead_tuples, v.num_dead_tuples from pg_stat_progress_vacuum as v join pg_class c on c.oid = v.relid;Times Reported Time consuming queries #16
Day Hour Count Duration Avg duration Feb 25 15 239 0ms 0ms 17 0ms 18 0ms 0ms 0ms select cast(count(*) / cast(setting as numeric) * ? as int) from pg_stat_activity, pg_settings where name = ? group by setting;Times Reported Time consuming queries #17
Day Hour Count Duration Avg duration Feb 25 15 18 0ms 0ms 18 0ms 1 0ms 0ms 0ms select distinct "public"."processes"."live" AS "live" from "public"."processes" left outer join "public"."brokers" "LT?" on "LT?"."id" = "public"."processes"."brokerid" left outer join "public"."contenttypes" "LT?" on "LT?"."id" = "public"."processes"."contenttypeid" order by ? asc;Times Reported Time consuming queries #18
Day Hour Count Duration Avg duration Feb 25 15 1 0ms 0ms 19 0ms 2 0ms 0ms 0ms select count(*) from "public"."executions" left outer join "public"."processes" "LT?" on "LT?"."id" = "public"."executions"."processid" where (processid = ?);Times Reported Time consuming queries #19
Day Hour Count Duration Avg duration Feb 25 15 2 0ms 0ms 20 0ms 57 0ms 0ms 0ms select * from ( select pricedatetime, open, high, low, close, volume, bsf from t60 where symbolid = ? and (bsf = ? or bsf is null) order by pricedatetime desc limit ?) a order by pricedatetime asc;Times Reported Time consuming queries #20
Day Hour Count Duration Avg duration Feb 25 15 57 0ms 0ms Most frequent queries (N)
Rank Times executed Total duration Min duration Max duration Avg duration Query 1 22,281 0ms 0ms 0ms 0ms select ?;Times Reported Time consuming queries #1
Day Hour Count Duration Avg duration Feb 25 15 22,281 0ms 0ms 2 13,064 0ms 0ms 0ms 0ms select distinct on (coalesce(bim.code, s.symbol) , s.exchange, s.timegranularity, df.timezone) s.symbolid as id, coalesce(bim.code, s.symbol) as name, s.symbol as symbol, dss.downloadersymbol as ticker, s.exchange as exchange, s.timegranularity as interval, df.timezone as timezone from symbols s inner join downloadersymbolsettings dss on dss.symbolid = s.symbolid inner join datafeedstimetable df on df.classname ilike dss.classname left join brokersymbollist bsl on bsl.brokerid = ? and bsl.symbolid = s.symbolid left outer join brokerinstrumentmap bim on dss.datafeedinstrumentid = bim.datafeedinstrumentid and bim.brokerid = ? and bim.type = ? where s.symbolid = ?;Times Reported Time consuming queries #2
Day Hour Count Duration Avg duration Feb 25 15 13,064 0ms 0ms 3 10,393 0ms 0ms 0ms 0ms select s.symbolid as id, s.symbol as name, s.exchange as exchange, s.timegranularity as interval, dtt.timezone as timezone from symbols s inner join downloadersymbolsettings dss on dss.symbolid = s.symbolid inner join datafeedstimetable dtt on dss.classname = dtt.classname and dtt.dayofweek = ? inner join brokersymbollist bsl on bsl.symbolid = s.symbolid where bsl.brokerid = ? and (? = ? or s.timegranularity = ?) and (s.symbol = ? or dss.downloadersymbol = ?) and dss.enabled = ?;Times Reported Time consuming queries #3
Day Hour Count Duration Avg duration Feb 25 15 10,393 0ms 0ms 4 5,551 0ms 0ms 0ms 0ms insert into t15 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) values (?, ?, ?, ?, ?, ?, ?, ?, ?, ?) on conflict (pricedatetime, symbolid) do update set open = ?, high = ?, low = ?, close = ?, volume = ?, bsf = ?, sastdatetimewritten = ?, sastdatetimereceived = ?;Times Reported Time consuming queries #4
Day Hour Count Duration Avg duration Feb 25 15 5,551 0ms 0ms 5 4,230 0ms 0ms 0ms 0ms insert into autochartist_results (resultid, symbolid, bandwidth, pattern, qtytp, gmttimefound, direction, initialtrend, breakout, volumeincrease, noise, symmetry, predictionpricefrom, predictionpriceto, predictiontimefrom, predictiontimeto, patternstarttime, patternendtime, patternstartprice, patternendprice, resx0, resx1, supportx0, supportx1, resy0, resy1, supporty0, supporty1, supportgradient, resgradient, riskreward, patternquality, trendchange, maxmovementafterbreakout, latestbaratbreakouttime, latestbaratbreakoutprice, patternlengthbars, temporarypattern, relevancestartdistance, simulation, writtendatetime) values (?, ?, ?.?, ?, ?, ?::timestamp without time zone, ?, ?.?, ?.?, ?.?, ?.?, ?.?, ?.?, ?.?, ?::timestamp without time zone, ?::timestamp without time zone, ?::timestamp without time zone, ?::timestamp without time zone, ?.?, ?.?, ?::timestamp without time zone, ?::timestamp without time zone, ?::timestamp without time zone, ?::timestamp without time zone, ?.?, ?.?, ?.?, ?.?, ?.?, ?.?, ?.?, ?.?, ?, ?.?, ?::timestamp without time zone, ?.?, ?, ?, ?.?, ?, current_timestamp::timestamp without time zone) on conflict do nothing;Times Reported Time consuming queries #5
Day Hour Count Duration Avg duration Feb 25 15 4,230 0ms 0ms 6 4,103 0ms 0ms 0ms 0ms insert into executionlogs (executionid, status, message, details, detailtype) values (null, ?, ?, null, null);Times Reported Time consuming queries #6
Day Hour Count Duration Avg duration Feb 25 15 4,103 0ms 0ms 7 3,346 0ms 0ms 0ms 0ms select datid, datname, pid, usesysid, usename, application_name, client_addr, client_hostname, client_port, backend_start, xact_start, query_start, state_change, wait_event_type, wait_event, state, backend_xid, backend_xmin, query, backend_type from pg_stat_activity where backend_type != ? or (coalesce(trim(query), ?) != ? and pid != pg_backend_pid() and query_start is not null and datname not ilike ? and datname not ilike ? and datname not ilike ? and datname not ilike ? and not (query_start < ?::timestamptz and state = ?));Times Reported Time consuming queries #7
Day Hour Count Duration Avg duration Feb 25 15 3,346 0ms 0ms 8 3,125 0ms 0ms 0ms 0ms insert into t30 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) values (?, ?, ?, ?, ?, ?, ?, ?, ?, ?) on conflict (pricedatetime, symbolid) do update set open = ?, high = ?, low = ?, close = ?, volume = ?, bsf = ?, sastdatetimewritten = ?, sastdatetimereceived = ?;Times Reported Time consuming queries #8
Day Hour Count Duration Avg duration Feb 25 15 3,125 0ms 0ms 9 2,426 0ms 0ms 0ms 0ms insert into fibonacci_results (bandwidth, pattern, gmttimefound, direction, patternstarttime, patternendtime, patternstartprice, patternendprice, qtytp, pricex, timex, pricea, timea, priceb, timeb, pricec, timec, priced, timed, averagequality, timequality, errormargin, patternlengthbars, target10, target06, target16, target07, target12, target05, target03, symbolid, noise, ratiosfound, temporarypattern, uniqueindex, completed, simulation, writtendatetime) values (?.?, ?, ?::timestamp without time zone, ?, ?::timestamp without time zone, ?::timestamp without time zone, ?.?, ?.?, ?, ?.?, ?::timestamp without time zone, ?.?, ?::timestamp without time zone, ?.?, ?::timestamp without time zone, ?.?, ?::timestamp without time zone, ?.?, ?::timestamp without time zone, ?.?, ?.?, ?.?, ?, ?.?, ?.?, ?.?, ?.?, ?.?, ?.?, ?.?, ?, ?.?, ?, ?, ?, ?, ?, current_timestamp::timestamp without time zone) on conflict do nothing;Times Reported Time consuming queries #9
Day Hour Count Duration Avg duration Feb 25 15 2,426 0ms 0ms 10 2,205 0ms 0ms 0ms 0ms set extra_float_digits = ?;Times Reported Time consuming queries #10
Day Hour Count Duration Avg duration Feb 25 15 2,205 0ms 0ms 11 2,179 0ms 0ms 0ms 0ms set application_name = ?;Times Reported Time consuming queries #11
Day Hour Count Duration Avg duration Feb 25 15 2,179 0ms 0ms 12 2,123 0ms 0ms 0ms 0ms update patternresultsrelevance set relevant = ?, saxo_relevant = ?, notrelevantpricedatetime = ?, reason = ? where uniqueindex = ? and relevant = ?;Times Reported Time consuming queries #12
Day Hour Count Duration Avg duration Feb 25 15 2,123 0ms 0ms 13 2,058 0ms 0ms 0ms 0ms insert into t60 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) values (?, ?, ?, ?, ?, ?, ?, ?, ?, ?) on conflict (pricedatetime, symbolid) do update set open = ?, high = ?, low = ?, close = ?, volume = ?, bsf = ?, sastdatetimewritten = ?, sastdatetimereceived = ?;Times Reported Time consuming queries #13
Day Hour Count Duration Avg duration Feb 25 15 2,058 0ms 0ms 14 1,745 0ms 0ms 0ms 0ms insert into keylevels_results (bandwidth, breakout, patternid, gmttimefound, approachingtimestamp, approachingregion, qtytp, patternlengthbars, patternprice, x0, x1, x2, x3, x4, x5, x6, x7, x8, x9, breakoutbars, breakoutprice, patternendtime, atbaridentified, atpriceidentified, errormargin, direction, symbolid, predictionpricefrom, predictionpriceto, predictiontimefrom, predictiontimebars, uniquepointsvalue, furthestprice, relevancestartdistance, patternclassid, patternstarttime, stoplosslevel, simulation, writtendatetime) values (?.?, ?, ?, ?::timestamp without time zone, ?, ?.?, ?, ?, ?.?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?.?, ?::timestamp without time zone, ?, ?.?, ?.?, ?, ?, ?.?, ?.?, ?::timestamp without time zone, ?, ?, ?.?, ?.?, ?, ?, ?.?, ?, current_timestamp::timestamp without time zone) on conflict do nothing;Times Reported Time consuming queries #14
Day Hour Count Duration Avg duration Feb 25 15 1,745 0ms 0ms 15 1,358 0ms 0ms 0ms 0ms with rar_max as ( select resultuid from relevance_autochartist_results order by resultuid desc limit ? ) select a.symbolid, pattern, patternid, resy0, resy1, resx0, resx1, supporty0, supporty1, supportx0, supportx1, predictiontimeto, patternstarttime, timegranularity, patternendtime, direction, trendchange, patternlengthbars, patternquality, case when a.old_resultuid = ? then a.old_resultuid else a.resultuid end as uid, breakout, initialtrend, volumeincrease, symmetry as uniformity, predictionpricefrom, predictionpriceto, noise, s.exchange, s.symbol, s.longname, s.shortname, breakout, dtt.timezone, patternstartprice, patternendprice, qtytp, newlevels.profit, newlevels.stop, newlevels.filtered, case when rar.age is not null then rar.age when a.resultuid <= rm.resultuid then ? else ? end as age, case when rar.relevant is not null then rar.relevant when a.resultuid <= rm.resultuid then ? else ? end as relevant, cps.pip from autochartist_results a inner join downloadersymbolsettings dss on a.symbolid = dss.symbolid inner join datafeedstimetable dtt on dss.classname = dtt.classname inner join symbols s on a.symbolid = s.symbolid inner join patterns p on p.patternname = a.pattern inner join rar_max rm on ? = ? left outer join relevance_autochartist_results rar on rar.resultuid = a.resultuid left join lateral calc_cp_signal (a.resultuid) newlevels on true left join currencypips cps on cps.symbol = s.symbol where (a.old_resultuid = ? or a.resultuid = ?) and dtt.dayofweek = ?;Times Reported Time consuming queries #15
Day Hour Count Duration Avg duration Feb 25 15 1,358 0ms 0ms 16 1,279 0ms 0ms 0ms 0ms select symbolid, pricedatetime, classname, downloadfrequency, downloadersymbol, open, high, low, close, volume, bsf, sastdatetimereceived from ( select pricedatetime, dss.classname, dss.downloadfrequency, dss.symbolid, dss.downloadersymbol, t.open, t.high, t.low, t.close, t.volume, t.bsf, t.sastdatetimereceived, row_number() over (partition by t.symbolid order by t.pricedatetime desc) as rn from t15 t, downloadersymbolsettings dss, symbols s where dss.classname = ? and dss.downloadfrequency = ? and dss.symbolid = t.symbolid and s.symbolid = dss.symbolid and dss.enabled = ? and s.deleted = ? and dss.downloadersymbol in (...) and t.pricedatetime > now() - interval ?) as ranked_candles_table where rn = ?;Times Reported Time consuming queries #16
Day Hour Count Duration Avg duration Feb 25 15 1,279 0ms 0ms 17 1,197 0ms 0ms 0ms 0ms select * from status_perbroker;Times Reported Time consuming queries #17
Day Hour Count Duration Avg duration Feb 25 15 1,197 0ms 0ms 18 1,195 0ms 0ms 0ms 0ms select relname, schemaname, heap_blks_read, heap_blks_hit, idx_blks_read, idx_blks_hit, toast_blks_read, toast_blks_hit, tidx_blks_read, tidx_blks_hit from pg_statio_user_tables where ((relname ~ ?));Times Reported Time consuming queries #18
Day Hour Count Duration Avg duration Feb 25 15 1,195 0ms 0ms 19 1,195 0ms 0ms 0ms 0ms select relname, schemaname, indexrelname, idx_scan, idx_tup_read, idx_tup_fetch, pg_relation_size(indexrelid) as index_size from pg_stat_user_indexes where ((relname ~ ?));Times Reported Time consuming queries #19
Day Hour Count Duration Avg duration Feb 25 15 1,195 0ms 0ms 20 1,195 0ms 0ms 0ms 0ms select n.nspname as schemaname, count(*) from ( select c.relnamespace from pg_class c where c.relkind in (...)) as subquery left join pg_namespace n on (n.oid = relnamespace) where n.nspname not in (...) group by n.nspname;Times Reported Time consuming queries #20
Day Hour Count Duration Avg duration Feb 25 15 1,195 0ms 0ms Normalized slowest queries (N)
Rank Min duration Max duration Avg duration Times executed Total duration Query 1 0ms 0ms 0ms 26 0ms select key, value from datasources ds inner join datasourceparams dsp on ds.id = dsp.datasourceid where ds.name = ?;Times Reported Time consuming queries #1
Day Hour Count Duration Avg duration Feb 25 15 26 0ms 0ms 2 0ms 0ms 0ms 303 0ms with rar_max as ( select resultuid from relevance_consecutivecandles_results order by resultuid desc limit ? ), all_results as ( select ccr.resultuid as resultuid, ccr.direction as direction, s.exchange as exchange, s.symbolid as symbolid, coalesce(bim.code, s.symbol) as symbol_code, s.longname as symbol_name, s.timegranularity as interval, ccr.patternendtime as identified, dtt.timezone as timezone, ccr.qtyconsecutivecandles as length, g.basegroupname, case when rcr.age is not null then rcr.age when ccr.resultuid <= rm.resultuid then ? else ? end as age, case when rcr.relevant is not null then rcr.relevant when ccr.resultuid <= rm.resultuid then ? else ? end as relevant, cps.pip, newlevels.filtered from consecutivecandles_results ccr inner join brokersymbollist bsl on bsl.brokerid = ? and bsl.symbolid = ccr.symbolid inner join symbols s on ccr.symbolid = s.symbolid and s.nonliquid = ? inner join downloadersymbolsettings dss on ccr.symbolid = dss.symbolid inner join datafeedstimetable dtt on dss.classname = dtt.classname and dtt.dayofweek = ? inner join symbolgroup sg on ccr.symbolid = sg.symbolid inner join groups g on sg.groupid = g.groupid inner join brokergroups bg on g.groupid = bg.groupid and bsl.brokerid = bg.brokerid inner join rar_max rm on ? = ? left outer join relevance_consecutivecandles_results rcr on rcr.resultuid = ccr.resultuid left join currencypips cps on cps.symbol = s.symbol left outer join brokerinstrumentmap bim on dss.datafeedinstrumentid = bim.datafeedinstrumentid and bim.brokerid = bsl.brokerid and bim.type = ? left join lateral calc_cc_signal_filter (ccr.resultuid) newlevels on true where ccr.gmttimefound > now() - interval ? and s.deleted = ? and (ccr.simulation = ? or ccr.simulation is null) and (? = ? or s.timegranularity in (...)) and (? = ? or s.exchange in (...)) and (? = ? or coalesce(bim.code, s.symbol) in (...)) and (? = ? or ccr.patternlengthbars <= ?) ), results as ( select distinct on (symbolid) * from all_results where (false = ? or relevant = ?) and (? = ? or age <= ?) order by symbolid, resultuid ) select * from results order by identified desc, length desc;Times Reported Time consuming queries #2
Day Hour Count Duration Avg duration Feb 25 15 303 0ms 0ms 3 0ms 0ms 0ms 1 0ms select count(*) from "public"."processes" left outer join "public"."brokers" "LT?" on "LT?"."id" = "public"."processes"."brokerid" left outer join "public"."contenttypes" "LT?" on "LT?"."id" = "public"."processes"."contenttypeid" where (brokerid = ?);Times Reported Time consuming queries #3
Day Hour Count Duration Avg duration Feb 25 15 1 0ms 0ms 4 0ms 0ms 0ms 2 0ms select trumpetsymbolid as sid, trumpettimegranularity as tg from brokersymbollist bsl left join powerstats_symboldata psd on bsl.symbolid = psd.symbolid left join downloadersymbolsettings dss on psd.symbolid = dss.symbolid left join symbols s on dss.symbolid = s.symbolid left outer join brokerinstrumentmap bdfi on code = ? and bdfi.brokerid = ? and dss.datafeedinstrumentid = bdfi.datafeedinstrumentid where (code = ? or s.symbol = ?) and bsl.brokerid = ? and dss.classname <> ? group by trumpetsymbolid, trumpettimegranularity;Times Reported Time consuming queries #4
Day Hour Count Duration Avg duration Feb 25 15 2 0ms 0ms 5 0ms 0ms 0ms 2,058 0ms insert into t60 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) values (?, ?, ?, ?, ?, ?, ?, ?, ?, ?) on conflict (pricedatetime, symbolid) do update set open = ?, high = ?, low = ?, close = ?, volume = ?, bsf = ?, sastdatetimewritten = ?, sastdatetimereceived = ?;Times Reported Time consuming queries #5
Day Hour Count Duration Avg duration Feb 25 15 2,058 0ms 0ms 6 0ms 0ms 0ms 74 0ms with rar_max as ( select resultuid from relevance_bigmovement_results order by resultuid desc limit ? ) select bmr.symbolid, patternstarttime, patternendtime, timegranularity, ? as direction, case when bmr.old_resultuid = ? then bmr.old_resultuid else bmr.resultuid end as uid, s.exchange, s.symbol, s.longname, s.shortname, dtt.timezone, bmr.patternmovement, bmr.statisticalmovement, bmr.fromprice, bmr.toprice, bmr.percentile, bmr.patternlengthbars, case when rbr.age is not null then rbr.age when bmr.resultuid <= rm.resultuid then ? else ? end as age, case when rbr.relevant is not null then rbr.relevant when bmr.resultuid <= rm.resultuid then ? else ? end as relevant, cps.pip from bigmovement_results bmr inner join downloadersymbolsettings dss on bmr.symbolid = dss.symbolid inner join datafeedstimetable dtt on dss.classname = dtt.classname inner join symbols s on bmr.symbolid = s.symbolid inner join rar_max rm on ? = ? left outer join relevance_bigmovement_results rbr on rbr.resultuid = bmr.resultuid left join currencypips cps on cps.symbol = s.symbol where (bmr.old_resultuid = ? or bmr.resultuid = ?) and dtt.dayofweek = ?;Times Reported Time consuming queries #6
Day Hour Count Duration Avg duration Feb 25 15 74 0ms 0ms 7 0ms 0ms 0ms 4 0ms select count(*) from datafeeds_latestrun where feedname ilike ? and ((latestrxtime > current_timestamp - interval ? and latestdbwritetime > current_timestamp - interval ?) or (latestdbwritetime > current_timestamp - interval ? and lateststartuptime > current_timestamp - interval ?));Times Reported Time consuming queries #7
Day Hour Count Duration Avg duration Feb 25 15 4 0ms 0ms 8 0ms 0ms 0ms 4 0ms select updaterelevantforrelevantresults ();Times Reported Time consuming queries #8
Day Hour Count Duration Avg duration Feb 25 15 4 0ms 0ms 9 0ms 0ms 0ms 114 0ms select case when a.old_resultuid = ? then a.old_resultuid else a.resultuid end as resultuid, s.symbol, timegranularity as interval, direction as direction, patternendtime as patternendtime, patternstartprice as psp, patternendprice as pep, target03 as t03, target16 as t16, patternlengthbars as length, p.patternname as patternname, dtt.timezone, cps.pip from fibonacci_results a inner join downloadersymbolsettings dss on a.symbolid = dss.symbolid inner join datafeedstimetable dtt on dss.classname = dtt.classname inner join symbols s on a.symbolid = s.symbolid inner join fibonaccipatterns p on a.pattern = p.patternname left join currencypips cps on cps.symbol = s.symbol where (a.old_resultuid = ? or a.resultuid = ?) and dtt.dayofweek = ?;Times Reported Time consuming queries #9
Day Hour Count Duration Avg duration Feb 25 15 114 0ms 0ms 10 0ms 0ms 0ms 7 0ms set datestyle = iso;Times Reported Time consuming queries #10
Day Hour Count Duration Avg duration Feb 25 15 7 0ms 0ms 11 0ms 0ms 0ms 2,123 0ms update patternresultsrelevance set relevant = ?, saxo_relevant = ?, notrelevantpricedatetime = ?, reason = ? where uniqueindex = ? and relevant = ?;Times Reported Time consuming queries #11
Day Hour Count Duration Avg duration Feb 25 15 2,123 0ms 0ms 12 0ms 0ms 0ms 9 0ms insert into resultmedia (processresultsid, type, name, filename) values (?, ?, ?, ?) returning id;Times Reported Time consuming queries #12
Day Hour Count Duration Avg duration Feb 25 15 9 0ms 0ms 13 0ms 0ms 0ms 7 0ms set client_encoding to ?;Times Reported Time consuming queries #13
Day Hour Count Duration Avg duration Feb 25 15 7 0ms 0ms 14 0ms 0ms 0ms 1 0ms select id from processes where id = ?;Times Reported Time consuming queries #14
Day Hour Count Duration Avg duration Feb 25 15 1 0ms 0ms 15 0ms 0ms 0ms 2 0ms select "public"."executions"."id" AS "id", "public"."executions"."processid" AS "processid", "public"."executions"."executiondate" AS "executiondate", "public"."executions"."errorcount" AS "errorcount", "public"."executions"."warningcount" AS "warningcount", "public"."executions"."isrunning" AS "isrunning", "public"."executions"."response" AS "response", "public"."executions"."live" AS "live", "public"."executions"."has_results" AS "has_results", "LT?"."id" AS "LA?" from "public"."executions" left outer join "public"."processes" "LT?" on "LT?"."id" = "public"."executions"."processid" where (processid = ?) order by "public"."executions"."id" desc limit ? offset ?;Times Reported Time consuming queries #15
Day Hour Count Duration Avg duration Feb 25 15 2 0ms 0ms 16 0ms 0ms 0ms 239 0ms select v.datname, c.relname, v.phase, v.heap_blks_total, v.heap_blks_scanned, v.heap_blks_vacuumed, v.index_vacuum_count, v.max_dead_tuples, v.num_dead_tuples from pg_stat_progress_vacuum as v join pg_class c on c.oid = v.relid;Times Reported Time consuming queries #16
Day Hour Count Duration Avg duration Feb 25 15 239 0ms 0ms 17 0ms 0ms 0ms 18 0ms select cast(count(*) / cast(setting as numeric) * ? as int) from pg_stat_activity, pg_settings where name = ? group by setting;Times Reported Time consuming queries #17
Day Hour Count Duration Avg duration Feb 25 15 18 0ms 0ms 18 0ms 0ms 0ms 1 0ms select distinct "public"."processes"."live" AS "live" from "public"."processes" left outer join "public"."brokers" "LT?" on "LT?"."id" = "public"."processes"."brokerid" left outer join "public"."contenttypes" "LT?" on "LT?"."id" = "public"."processes"."contenttypeid" order by ? asc;Times Reported Time consuming queries #18
Day Hour Count Duration Avg duration Feb 25 15 1 0ms 0ms 19 0ms 0ms 0ms 2 0ms select count(*) from "public"."executions" left outer join "public"."processes" "LT?" on "LT?"."id" = "public"."executions"."processid" where (processid = ?);Times Reported Time consuming queries #19
Day Hour Count Duration Avg duration Feb 25 15 2 0ms 0ms 20 0ms 0ms 0ms 57 0ms select * from ( select pricedatetime, open, high, low, close, volume, bsf from t60 where symbolid = ? and (bsf = ? or bsf is null) order by pricedatetime desc limit ?) a order by pricedatetime asc;Times Reported Time consuming queries #20
Day Hour Count Duration Avg duration Feb 25 15 57 0ms 0ms Time consuming prepare
Rank Total duration Times executed Min duration Max duration Avg duration Query 1 2s880ms 2,777 0ms 13ms 1ms WITH rar_max as ( ;Times Reported Time consuming prepare #1
Day Hour Count Duration Avg duration Feb 25 15 2,777 2s880ms 1ms -
WITH rar_max as ( ;
Date: 2026-02-25 15:35:58 Duration: 13ms Database: postgres
-
WITH rar_max as ( ;
Date: 2026-02-25 15:29:06 Duration: 9ms Database: postgres
-
WITH rar_max as ( ;
Date: 2026-02-25 15:11:38 Duration: 8ms Database: postgres
2 1s556ms 1,408 0ms 4ms 1ms SELECT symbolid, ;Times Reported Time consuming prepare #2
Day Hour Count Duration Avg duration 15 1,408 1s556ms 1ms -
SELECT symbolid, ;
Date: 2026-02-25 15:00:33 Duration: 4ms Database: postgres
-
SELECT symbolid, ;
Date: 2026-02-25 15:47:16 Duration: 3ms Database: postgres
-
SELECT symbolid, ;
Date: 2026-02-25 15:46:50 Duration: 3ms Database: postgres
3 1s198ms 3,936 0ms 22ms 0ms SELECT ;Times Reported Time consuming prepare #3
Day Hour Count Duration Avg duration 15 3,936 1s198ms 0ms -
SELECT ;
Date: 2026-02-25 15:30:05 Duration: 22ms Database: postgres
-
SELECT ;
Date: 2026-02-25 15:17:37 Duration: 8ms Database: postgres
-
SELECT ;
Date: 2026-02-25 15:00:13 Duration: 8ms Database: postgres
4 1s53ms 956 0ms 2ms 1ms SELECT s.symbolid, dss.downloadfrequency, dss.downloadersymbol;Times Reported Time consuming prepare #4
Day Hour Count Duration Avg duration 15 956 1s53ms 1ms -
SELECT s.symbolid, dss.downloadfrequency, dss.downloadersymbol;
Date: 2026-02-25 15:00:34 Duration: 2ms Database: postgres
-
SELECT s.symbolid, dss.downloadfrequency, dss.downloadersymbol;
Date: 2026-02-25 15:17:11 Duration: 1ms Database: postgres
-
SELECT s.symbolid, dss.downloadfrequency, dss.downloadersymbol;
Date: 2026-02-25 15:15:47 Duration: 1ms Database: postgres
5 337ms 2,205 0ms 5ms 0ms SET extra_float_digits = 3;Times Reported Time consuming prepare #5
Day Hour Count Duration Avg duration 15 2,205 337ms 0ms -
SET extra_float_digits = 3;
Date: 2026-02-25 15:16:55 Duration: 5ms Database: postgres
-
SET extra_float_digits = 3;
Date: 2026-02-25 15:17:37 Duration: 4ms Database: postgres
-
SET extra_float_digits = 3;
Date: 2026-02-25 15:00:04 Duration: 3ms Database: postgres
6 271ms 2,947 0ms 0ms 0ms INSERT INTO T30 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;Times Reported Time consuming prepare #6
Day Hour Count Duration Avg duration 15 2,947 271ms 0ms -
INSERT INTO T30 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;
Date: 2026-02-25 15:15:55 Duration: 0ms Database: postgres
-
INSERT INTO T30 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;
Date: 2026-02-25 15:00:21 Duration: 0ms Database: postgres
-
INSERT INTO T30 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;
Date: 2026-02-25 15:00:59 Duration: 0ms Database: postgres
7 195ms 1,890 0ms 0ms 0ms INSERT INTO T60 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;Times Reported Time consuming prepare #7
Day Hour Count Duration Avg duration 15 1,890 195ms 0ms -
INSERT INTO T60 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;
Date: 2026-02-25 15:00:35 Duration: 0ms Database: postgres
-
INSERT INTO T60 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;
Date: 2026-02-25 15:17:52 Duration: 0ms Database: postgres
-
INSERT INTO T60 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;
Date: 2026-02-25 15:16:08 Duration: 0ms Database: postgres
8 181ms 1,292 0ms 0ms 0ms INSERT INTO T15 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;Times Reported Time consuming prepare #8
Day Hour Count Duration Avg duration 15 1,292 181ms 0ms -
INSERT INTO T15 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;
Date: 2026-02-25 15:17:59 Duration: 0ms Database: postgres
-
INSERT INTO T15 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;
Date: 2026-02-25 15:32:49 Duration: 0ms Database: postgres
-
INSERT INTO T15 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;
Date: 2026-02-25 15:16:00 Duration: 0ms Database: postgres
9 90ms 1,868 0ms 7ms 0ms select 1;Times Reported Time consuming prepare #9
Day Hour Count Duration Avg duration 15 1,868 90ms 0ms -
select 1;
Date: 2026-02-25 15:35:55 Duration: 7ms Database: postgres
-
select 1;
Date: 2026-02-25 15:29:05 Duration: 2ms Database: postgres
-
select 1;
Date: 2026-02-25 15:46:57 Duration: 1ms Database: postgres
10 64ms 12 4ms 7ms 5ms with sym_info as ( ;Times Reported Time consuming prepare #10
Day Hour Count Duration Avg duration 15 12 64ms 5ms -
with sym_info as ( ;
Date: 2026-02-25 15:06:40 Duration: 7ms Database: postgres
-
with sym_info as ( ;
Date: 2026-02-25 15:36:56 Duration: 6ms Database: postgres
-
with sym_info as ( ;
Date: 2026-02-25 15:36:51 Duration: 6ms Database: postgres
11 45ms 26 0ms 5ms 1ms WITH last_candle AS ( ;Times Reported Time consuming prepare #11
Day Hour Count Duration Avg duration 15 26 45ms 1ms -
WITH last_candle AS ( ;
Date: 2026-02-25 15:48:01 Duration: 5ms Database: postgres
-
WITH last_candle AS ( ;
Date: 2026-02-25 15:00:12 Duration: 4ms Database: postgres
-
WITH last_candle AS ( ;
Date: 2026-02-25 15:32:17 Duration: 3ms Database: postgres
12 44ms 18 1ms 2ms 2ms select cast(count(*) / cast(setting as numeric) * 100 as int) from pg_stat_activity, pg_settings WHERE name = 'max_connections' group by setting;Times Reported Time consuming prepare #12
Day Hour Count Duration Avg duration 15 18 44ms 2ms -
select cast(count(*) / cast(setting as numeric) * 100 as int) from pg_stat_activity, pg_settings WHERE name = 'max_connections' group by setting;
Date: 2026-02-25 15:20:03 Duration: 2ms Database: postgres
-
select cast(count(*) / cast(setting as numeric) * 100 as int) from pg_stat_activity, pg_settings WHERE name = 'max_connections' group by setting;
Date: 2026-02-25 15:50:03 Duration: 2ms Database: postgres
-
select cast(count(*) / cast(setting as numeric) * 100 as int) from pg_stat_activity, pg_settings WHERE name = 'max_connections' group by setting;
Date: 2026-02-25 15:01:17 Duration: 2ms Database: postgres
13 28ms 25 0ms 1ms 1ms select distinct classname, to_char(created_datetime, 'yyyy-mm-dd HH24:MI'), to_char(cleared_datetime, 'yyyy-mm-dd HH24:MI'), action_to_take, description, created_datetime from datafeed_restarter_events where (is_current_entry = 1 OR cleared_datetime > current_timestamp - interval '17 hour') order by created_datetime desc;Times Reported Time consuming prepare #13
Day Hour Count Duration Avg duration 15 25 28ms 1ms -
select distinct classname, to_char(created_datetime, 'yyyy-mm-dd HH24:MI'), to_char(cleared_datetime, 'yyyy-mm-dd HH24:MI'), action_to_take, description, created_datetime from datafeed_restarter_events where (is_current_entry = 1 OR cleared_datetime > current_timestamp - interval '17 hour') order by created_datetime desc;
Date: 2026-02-25 15:10:52 Duration: 1ms Database: postgres
-
select distinct classname, to_char(created_datetime, 'yyyy-mm-dd HH24:MI'), to_char(cleared_datetime, 'yyyy-mm-dd HH24:MI'), action_to_take, description, created_datetime from datafeed_restarter_events where (is_current_entry = 1 OR cleared_datetime > current_timestamp - interval '17 hour') order by created_datetime desc;
Date: 2026-02-25 15:05:52 Duration: 1ms Database: postgres
-
select distinct classname, to_char(created_datetime, 'yyyy-mm-dd HH24:MI'), to_char(cleared_datetime, 'yyyy-mm-dd HH24:MI'), action_to_take, description, created_datetime from datafeed_restarter_events where (is_current_entry = 1 OR cleared_datetime > current_timestamp - interval '17 hour') order by created_datetime desc;
Date: 2026-02-25 15:25:33 Duration: 1ms Database: postgres
14 27ms 117 0ms 0ms 0ms SELECT NULL AS TABLE_CAT, n.nspname AS TABLE_SCHEM, c.relname AS TABLE_NAME, CASE n.nspname ~ '^pg_' OR n.nspname = 'information_schema' WHEN true THEN CASE WHEN n.nspname = 'pg_catalog' OR n.nspname = 'information_schema' THEN CASE c.relkind WHEN 'r' THEN 'SYSTEM TABLE' WHEN 'v' THEN 'SYSTEM VIEW' WHEN 'i' THEN 'SYSTEM INDEX' ELSE NULL END WHEN n.nspname = 'pg_toast' THEN CASE c.relkind WHEN 'r' THEN 'SYSTEM TOAST TABLE' WHEN 'i' THEN 'SYSTEM TOAST INDEX' ELSE NULL END ELSE CASE c.relkind WHEN 'r' THEN 'TEMPORARY TABLE' WHEN 'p' THEN 'TEMPORARY TABLE' WHEN 'i' THEN 'TEMPORARY INDEX' WHEN 'S' THEN 'TEMPORARY SEQUENCE' WHEN 'v' THEN 'TEMPORARY VIEW' ELSE NULL END END WHEN false THEN CASE c.relkind WHEN 'r' THEN 'TABLE' WHEN 'p' THEN 'PARTITIONED TABLE' WHEN 'i' THEN 'INDEX' WHEN 'S' THEN 'SEQUENCE' WHEN 'v' THEN 'VIEW' WHEN 'c' THEN 'TYPE' WHEN 'f' THEN 'FOREIGN TABLE' WHEN 'm' THEN 'MATERIALIZED VIEW' ELSE NULL END ELSE NULL END AS TABLE_TYPE, d.description AS REMARKS, '' as TYPE_CAT, '' as TYPE_SCHEM, '' as TYPE_NAME, '' AS SELF_REFERENCING_COL_NAME, '' AS REF_GENERATION FROM pg_catalog.pg_namespace n, pg_catalog.pg_class c LEFT JOIN pg_catalog.pg_description d ON (c.oid = d.objoid AND d.objsubid = 0) LEFT JOIN pg_catalog.pg_class dc ON (d.classoid = dc.oid AND dc.relname = 'pg_class') LEFT JOIN pg_catalog.pg_namespace dn ON (dn.oid = dc.relnamespace AND dn.nspname = 'pg_catalog') WHERE c.relnamespace = n.oid AND c.relname LIKE 'PROBABLYNOT' AND (false OR (c.relkind = 'r' AND n.nspname !~ '^pg_' AND n.nspname <> 'information_schema')) ORDER BY TABLE_TYPE, TABLE_SCHEM, TABLE_NAME;Times Reported Time consuming prepare #14
Day Hour Count Duration Avg duration 15 117 27ms 0ms -
SELECT NULL AS TABLE_CAT, n.nspname AS TABLE_SCHEM, c.relname AS TABLE_NAME, CASE n.nspname ~ '^pg_' OR n.nspname = 'information_schema' WHEN true THEN CASE WHEN n.nspname = 'pg_catalog' OR n.nspname = 'information_schema' THEN CASE c.relkind WHEN 'r' THEN 'SYSTEM TABLE' WHEN 'v' THEN 'SYSTEM VIEW' WHEN 'i' THEN 'SYSTEM INDEX' ELSE NULL END WHEN n.nspname = 'pg_toast' THEN CASE c.relkind WHEN 'r' THEN 'SYSTEM TOAST TABLE' WHEN 'i' THEN 'SYSTEM TOAST INDEX' ELSE NULL END ELSE CASE c.relkind WHEN 'r' THEN 'TEMPORARY TABLE' WHEN 'p' THEN 'TEMPORARY TABLE' WHEN 'i' THEN 'TEMPORARY INDEX' WHEN 'S' THEN 'TEMPORARY SEQUENCE' WHEN 'v' THEN 'TEMPORARY VIEW' ELSE NULL END END WHEN false THEN CASE c.relkind WHEN 'r' THEN 'TABLE' WHEN 'p' THEN 'PARTITIONED TABLE' WHEN 'i' THEN 'INDEX' WHEN 'S' THEN 'SEQUENCE' WHEN 'v' THEN 'VIEW' WHEN 'c' THEN 'TYPE' WHEN 'f' THEN 'FOREIGN TABLE' WHEN 'm' THEN 'MATERIALIZED VIEW' ELSE NULL END ELSE NULL END AS TABLE_TYPE, d.description AS REMARKS, '' as TYPE_CAT, '' as TYPE_SCHEM, '' as TYPE_NAME, '' AS SELF_REFERENCING_COL_NAME, '' AS REF_GENERATION FROM pg_catalog.pg_namespace n, pg_catalog.pg_class c LEFT JOIN pg_catalog.pg_description d ON (c.oid = d.objoid AND d.objsubid = 0) LEFT JOIN pg_catalog.pg_class dc ON (d.classoid = dc.oid AND dc.relname = 'pg_class') LEFT JOIN pg_catalog.pg_namespace dn ON (dn.oid = dc.relnamespace AND dn.nspname = 'pg_catalog') WHERE c.relnamespace = n.oid AND c.relname LIKE 'PROBABLYNOT' AND (false OR (c.relkind = 'r' AND n.nspname !~ '^pg_' AND n.nspname <> 'information_schema')) ORDER BY TABLE_TYPE, TABLE_SCHEM, TABLE_NAME;
Date: 2026-02-25 15:13:18 Duration: 0ms Database: postgres
-
SELECT NULL AS TABLE_CAT, n.nspname AS TABLE_SCHEM, c.relname AS TABLE_NAME, CASE n.nspname ~ '^pg_' OR n.nspname = 'information_schema' WHEN true THEN CASE WHEN n.nspname = 'pg_catalog' OR n.nspname = 'information_schema' THEN CASE c.relkind WHEN 'r' THEN 'SYSTEM TABLE' WHEN 'v' THEN 'SYSTEM VIEW' WHEN 'i' THEN 'SYSTEM INDEX' ELSE NULL END WHEN n.nspname = 'pg_toast' THEN CASE c.relkind WHEN 'r' THEN 'SYSTEM TOAST TABLE' WHEN 'i' THEN 'SYSTEM TOAST INDEX' ELSE NULL END ELSE CASE c.relkind WHEN 'r' THEN 'TEMPORARY TABLE' WHEN 'p' THEN 'TEMPORARY TABLE' WHEN 'i' THEN 'TEMPORARY INDEX' WHEN 'S' THEN 'TEMPORARY SEQUENCE' WHEN 'v' THEN 'TEMPORARY VIEW' ELSE NULL END END WHEN false THEN CASE c.relkind WHEN 'r' THEN 'TABLE' WHEN 'p' THEN 'PARTITIONED TABLE' WHEN 'i' THEN 'INDEX' WHEN 'S' THEN 'SEQUENCE' WHEN 'v' THEN 'VIEW' WHEN 'c' THEN 'TYPE' WHEN 'f' THEN 'FOREIGN TABLE' WHEN 'm' THEN 'MATERIALIZED VIEW' ELSE NULL END ELSE NULL END AS TABLE_TYPE, d.description AS REMARKS, '' as TYPE_CAT, '' as TYPE_SCHEM, '' as TYPE_NAME, '' AS SELF_REFERENCING_COL_NAME, '' AS REF_GENERATION FROM pg_catalog.pg_namespace n, pg_catalog.pg_class c LEFT JOIN pg_catalog.pg_description d ON (c.oid = d.objoid AND d.objsubid = 0) LEFT JOIN pg_catalog.pg_class dc ON (d.classoid = dc.oid AND dc.relname = 'pg_class') LEFT JOIN pg_catalog.pg_namespace dn ON (dn.oid = dc.relnamespace AND dn.nspname = 'pg_catalog') WHERE c.relnamespace = n.oid AND c.relname LIKE 'PROBABLYNOT' AND (false OR (c.relkind = 'r' AND n.nspname !~ '^pg_' AND n.nspname <> 'information_schema')) ORDER BY TABLE_TYPE, TABLE_SCHEM, TABLE_NAME;
Date: 2026-02-25 15:13:18 Duration: 0ms Database: postgres
-
SELECT NULL AS TABLE_CAT, n.nspname AS TABLE_SCHEM, c.relname AS TABLE_NAME, CASE n.nspname ~ '^pg_' OR n.nspname = 'information_schema' WHEN true THEN CASE WHEN n.nspname = 'pg_catalog' OR n.nspname = 'information_schema' THEN CASE c.relkind WHEN 'r' THEN 'SYSTEM TABLE' WHEN 'v' THEN 'SYSTEM VIEW' WHEN 'i' THEN 'SYSTEM INDEX' ELSE NULL END WHEN n.nspname = 'pg_toast' THEN CASE c.relkind WHEN 'r' THEN 'SYSTEM TOAST TABLE' WHEN 'i' THEN 'SYSTEM TOAST INDEX' ELSE NULL END ELSE CASE c.relkind WHEN 'r' THEN 'TEMPORARY TABLE' WHEN 'p' THEN 'TEMPORARY TABLE' WHEN 'i' THEN 'TEMPORARY INDEX' WHEN 'S' THEN 'TEMPORARY SEQUENCE' WHEN 'v' THEN 'TEMPORARY VIEW' ELSE NULL END END WHEN false THEN CASE c.relkind WHEN 'r' THEN 'TABLE' WHEN 'p' THEN 'PARTITIONED TABLE' WHEN 'i' THEN 'INDEX' WHEN 'S' THEN 'SEQUENCE' WHEN 'v' THEN 'VIEW' WHEN 'c' THEN 'TYPE' WHEN 'f' THEN 'FOREIGN TABLE' WHEN 'm' THEN 'MATERIALIZED VIEW' ELSE NULL END ELSE NULL END AS TABLE_TYPE, d.description AS REMARKS, '' as TYPE_CAT, '' as TYPE_SCHEM, '' as TYPE_NAME, '' AS SELF_REFERENCING_COL_NAME, '' AS REF_GENERATION FROM pg_catalog.pg_namespace n, pg_catalog.pg_class c LEFT JOIN pg_catalog.pg_description d ON (c.oid = d.objoid AND d.objsubid = 0) LEFT JOIN pg_catalog.pg_class dc ON (d.classoid = dc.oid AND dc.relname = 'pg_class') LEFT JOIN pg_catalog.pg_namespace dn ON (dn.oid = dc.relnamespace AND dn.nspname = 'pg_catalog') WHERE c.relnamespace = n.oid AND c.relname LIKE 'PROBABLYNOT' AND (false OR (c.relkind = 'r' AND n.nspname !~ '^pg_' AND n.nspname <> 'information_schema')) ORDER BY TABLE_TYPE, TABLE_SCHEM, TABLE_NAME;
Date: 2026-02-25 15:13:17 Duration: 0ms Database: postgres
15 26ms 2,179 0ms 0ms 0ms SET application_name = 'PostgreSQL JDBC Driver';Times Reported Time consuming prepare #15
Day Hour Count Duration Avg duration 15 2,179 26ms 0ms -
SET application_name = 'PostgreSQL JDBC Driver';
Date: 2026-02-25 15:25:05 Duration: 0ms Database: postgres
-
SET application_name = 'PostgreSQL JDBC Driver';
Date: 2026-02-25 15:00:26 Duration: 0ms Database: postgres
-
SET application_name = 'PostgreSQL JDBC Driver';
Date: 2026-02-25 15:46:58 Duration: 0ms Database: postgres
16 26ms 192 0ms 0ms 0ms INSERT INTO T240 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;Times Reported Time consuming prepare #16
Day Hour Count Duration Avg duration 15 192 26ms 0ms -
INSERT INTO T240 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;
Date: 2026-02-25 15:01:20 Duration: 0ms Database: postgres
-
INSERT INTO T240 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;
Date: 2026-02-25 15:02:06 Duration: 0ms Database: postgres
-
INSERT INTO T240 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;
Date: 2026-02-25 15:15:48 Duration: 0ms Database: postgres
17 23ms 25 0ms 1ms 0ms select feedname, to_char(latestrxtime, 'yyyy-mm-dd HH24:MI'), to_char(LatestDBWriteTime, 'yyyy-mm-dd HH24:MI'), to_char(LatestStartupTime, 'yyyy-mm-dd HH24:MI'), StartupTimeInMinutes, dm.source_type, dm.transport_type, case when latestrxtime < (CURRENT_TIMESTAMP - 5 * interval '1 minute') then 'X' else 'OK' end, case when (feedname ilike '%_EOD' OR feedname ilike 'IQFEED_DAILIES' or feedname ilike 'YAHOO%' or feedname ilike 'QUANDL_FUTURES%' or feedname ilike 'BAR_CHART') then case when LatestDBWriteTime < (CURRENT_TIMESTAMP - 24 * interval '1 hour') then 'X' else 'OK' end else case when (LatestDBWriteTime < (CURRENT_TIMESTAMP - 15 * interval '1 minute') and LatestStartupTime < (CURRENT_TIMESTAMP - 30 * interval '1 minute')) OR latestrxtime < CURRENT_TIMESTAMP - interval '2 hour' then 'X' else 'OK' end end as statusDB, comment from datafeeds_latestrun dlr left outer join datafeeds df on dlr.feedname ilike df.name inner join datafeeds_metadata dm on df.metadata_id = dm.id order by feedname;Times Reported Time consuming prepare #17
Day Hour Count Duration Avg duration 15 25 23ms 0ms -
select feedname, to_char(latestrxtime, 'yyyy-mm-dd HH24:MI'), to_char(LatestDBWriteTime, 'yyyy-mm-dd HH24:MI'), to_char(LatestStartupTime, 'yyyy-mm-dd HH24:MI'), StartupTimeInMinutes, dm.source_type, dm.transport_type, case when latestrxtime < (CURRENT_TIMESTAMP - 5 * interval '1 minute') then 'X' else 'OK' end, case when (feedname ilike '%_EOD' OR feedname ilike 'IQFEED_DAILIES' or feedname ilike 'YAHOO%' or feedname ilike 'QUANDL_FUTURES%' or feedname ilike 'BAR_CHART') then case when LatestDBWriteTime < (CURRENT_TIMESTAMP - 24 * interval '1 hour') then 'X' else 'OK' end else case when (LatestDBWriteTime < (CURRENT_TIMESTAMP - 15 * interval '1 minute') and LatestStartupTime < (CURRENT_TIMESTAMP - 30 * interval '1 minute')) OR latestrxtime < CURRENT_TIMESTAMP - interval '2 hour' then 'X' else 'OK' end end as statusDB, comment from datafeeds_latestrun dlr left outer join datafeeds df on dlr.feedname ilike df.name inner join datafeeds_metadata dm on df.metadata_id = dm.id order by feedname;
Date: 2026-02-25 15:10:52 Duration: 1ms Database: postgres
-
select feedname, to_char(latestrxtime, 'yyyy-mm-dd HH24:MI'), to_char(LatestDBWriteTime, 'yyyy-mm-dd HH24:MI'), to_char(LatestStartupTime, 'yyyy-mm-dd HH24:MI'), StartupTimeInMinutes, dm.source_type, dm.transport_type, case when latestrxtime < (CURRENT_TIMESTAMP - 5 * interval '1 minute') then 'X' else 'OK' end, case when (feedname ilike '%_EOD' OR feedname ilike 'IQFEED_DAILIES' or feedname ilike 'YAHOO%' or feedname ilike 'QUANDL_FUTURES%' or feedname ilike 'BAR_CHART') then case when LatestDBWriteTime < (CURRENT_TIMESTAMP - 24 * interval '1 hour') then 'X' else 'OK' end else case when (LatestDBWriteTime < (CURRENT_TIMESTAMP - 15 * interval '1 minute') and LatestStartupTime < (CURRENT_TIMESTAMP - 30 * interval '1 minute')) OR latestrxtime < CURRENT_TIMESTAMP - interval '2 hour' then 'X' else 'OK' end end as statusDB, comment from datafeeds_latestrun dlr left outer join datafeeds df on dlr.feedname ilike df.name inner join datafeeds_metadata dm on df.metadata_id = dm.id order by feedname;
Date: 2026-02-25 15:15:53 Duration: 1ms Database: postgres
-
select feedname, to_char(latestrxtime, 'yyyy-mm-dd HH24:MI'), to_char(LatestDBWriteTime, 'yyyy-mm-dd HH24:MI'), to_char(LatestStartupTime, 'yyyy-mm-dd HH24:MI'), StartupTimeInMinutes, dm.source_type, dm.transport_type, case when latestrxtime < (CURRENT_TIMESTAMP - 5 * interval '1 minute') then 'X' else 'OK' end, case when (feedname ilike '%_EOD' OR feedname ilike 'IQFEED_DAILIES' or feedname ilike 'YAHOO%' or feedname ilike 'QUANDL_FUTURES%' or feedname ilike 'BAR_CHART') then case when LatestDBWriteTime < (CURRENT_TIMESTAMP - 24 * interval '1 hour') then 'X' else 'OK' end else case when (LatestDBWriteTime < (CURRENT_TIMESTAMP - 15 * interval '1 minute') and LatestStartupTime < (CURRENT_TIMESTAMP - 30 * interval '1 minute')) OR latestrxtime < CURRENT_TIMESTAMP - interval '2 hour' then 'X' else 'OK' end end as statusDB, comment from datafeeds_latestrun dlr left outer join datafeeds df on dlr.feedname ilike df.name inner join datafeeds_metadata dm on df.metadata_id = dm.id order by feedname;
Date: 2026-02-25 15:05:52 Duration: 1ms Database: postgres
18 17ms 6 2ms 3ms 2ms with rankedmt4 as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors ), last_feed_entry as ( select * from rankedmt4 where r = 1 ), ok_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where status = 'OK' ), earliest_entry_after_ok as ( select m.datafeedname, min(m.eventtimestamp) as eventtimestamp from mt4datafeederrors m left outer join ( select datafeedname, eventtimestamp from ok_entries where r = 1) oo on m.datafeedname = oo.datafeedname where m.eventtimestamp > coalesce(oo.eventtimestamp, '1900-01-01'::timestamp without time zone) group by m.datafeedname ), notified_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where notified is not null and notified <> '' ), broker as ( select *, row_number() over (partition by feedname order by brokerid) r from ( select distinct b.brokerid, b.name as brokername, dss.classname as feedname from downloadersymbolsettings dss inner join brokersymbollist bsl on dss.symbolid = bsl.symbolid inner join broker b on bsl.brokerid = b.brokerid where dss.enabled = 1) a ) select last.id, last.datafeedname, last.eventtimestamp, last.status, last.errordescription, last.serveraddress, last.username, note.notified, note.eventtimestamp, broker.brokername from last_feed_entry last inner join earliest_entry_after_ok after_ok on last.datafeedname = after_ok.datafeedname inner join broker on last.datafeedname = broker.feedname left outer join ok_entries ok on ok.datafeedname = last.datafeedname left outer join notified_entries note on note.datafeedname = last.datafeedname and note.r = 1 where (ok.r is null or ok.r = 1) and last.datafeedname not in ( select distinct datafeedname from last_feed_entry where status = 'OK') and extract(epoch from (last.eventtimestamp - after_ok.eventtimestamp)) > 60 * 60 and last.eventtimestamp > current_timestamp - interval '1 day' and (note.eventtimestamp is null or note.eventtimestamp < current_timestamp - interval '10 hours') and last.eventtimestamp > current_timestamp - interval '1 hour' and broker.r = 1;Times Reported Time consuming prepare #18
Day Hour Count Duration Avg duration 15 6 17ms 2ms -
with rankedmt4 as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors ), last_feed_entry as ( select * from rankedmt4 where r = 1 ), ok_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where status = 'OK' ), earliest_entry_after_ok as ( select m.datafeedname, min(m.eventtimestamp) as eventtimestamp from mt4datafeederrors m left outer join ( select datafeedname, eventtimestamp from ok_entries where r = 1) oo on m.datafeedname = oo.datafeedname where m.eventtimestamp > coalesce(oo.eventtimestamp, '1900-01-01'::timestamp without time zone) group by m.datafeedname ), notified_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where notified is not null and notified <> '' ), broker as ( select *, row_number() over (partition by feedname order by brokerid) r from ( select distinct b.brokerid, b.name as brokername, dss.classname as feedname from downloadersymbolsettings dss inner join brokersymbollist bsl on dss.symbolid = bsl.symbolid inner join broker b on bsl.brokerid = b.brokerid where dss.enabled = 1) a ) select last.id, last.datafeedname, last.eventtimestamp, last.status, last.errordescription, last.serveraddress, last.username, note.notified, note.eventtimestamp, broker.brokername from last_feed_entry last inner join earliest_entry_after_ok after_ok on last.datafeedname = after_ok.datafeedname inner join broker on last.datafeedname = broker.feedname left outer join ok_entries ok on ok.datafeedname = last.datafeedname left outer join notified_entries note on note.datafeedname = last.datafeedname and note.r = 1 where (ok.r is null or ok.r = 1) and last.datafeedname not in ( select distinct datafeedname from last_feed_entry where status = 'OK') and extract(epoch from (last.eventtimestamp - after_ok.eventtimestamp)) > 60 * 60 and last.eventtimestamp > current_timestamp - interval '1 day' and (note.eventtimestamp is null or note.eventtimestamp < current_timestamp - interval '10 hours') and last.eventtimestamp > current_timestamp - interval '1 hour' and broker.r = 1;
Date: 2026-02-25 15:30:02 Duration: 3ms Database: postgres
-
with rankedmt4 as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors ), last_feed_entry as ( select * from rankedmt4 where r = 1 ), ok_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where status = 'OK' ), earliest_entry_after_ok as ( select m.datafeedname, min(m.eventtimestamp) as eventtimestamp from mt4datafeederrors m left outer join ( select datafeedname, eventtimestamp from ok_entries where r = 1) oo on m.datafeedname = oo.datafeedname where m.eventtimestamp > coalesce(oo.eventtimestamp, '1900-01-01'::timestamp without time zone) group by m.datafeedname ), notified_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where notified is not null and notified <> '' ), broker as ( select *, row_number() over (partition by feedname order by brokerid) r from ( select distinct b.brokerid, b.name as brokername, dss.classname as feedname from downloadersymbolsettings dss inner join brokersymbollist bsl on dss.symbolid = bsl.symbolid inner join broker b on bsl.brokerid = b.brokerid where dss.enabled = 1) a ) select last.id, last.datafeedname, last.eventtimestamp, last.status, last.errordescription, last.serveraddress, last.username, note.notified, note.eventtimestamp, broker.brokername from last_feed_entry last inner join earliest_entry_after_ok after_ok on last.datafeedname = after_ok.datafeedname inner join broker on last.datafeedname = broker.feedname left outer join ok_entries ok on ok.datafeedname = last.datafeedname left outer join notified_entries note on note.datafeedname = last.datafeedname and note.r = 1 where (ok.r is null or ok.r = 1) and last.datafeedname not in ( select distinct datafeedname from last_feed_entry where status = 'OK') and extract(epoch from (last.eventtimestamp - after_ok.eventtimestamp)) > 60 * 60 and last.eventtimestamp > current_timestamp - interval '1 day' and (note.eventtimestamp is null or note.eventtimestamp < current_timestamp - interval '10 hours') and last.eventtimestamp > current_timestamp - interval '1 hour' and broker.r = 1;
Date: 2026-02-25 15:50:02 Duration: 3ms Database: postgres
-
with rankedmt4 as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors ), last_feed_entry as ( select * from rankedmt4 where r = 1 ), ok_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where status = 'OK' ), earliest_entry_after_ok as ( select m.datafeedname, min(m.eventtimestamp) as eventtimestamp from mt4datafeederrors m left outer join ( select datafeedname, eventtimestamp from ok_entries where r = 1) oo on m.datafeedname = oo.datafeedname where m.eventtimestamp > coalesce(oo.eventtimestamp, '1900-01-01'::timestamp without time zone) group by m.datafeedname ), notified_entries as ( select *, row_number() over (partition by datafeedname order by eventtimestamp desc) r from mt4datafeederrors where notified is not null and notified <> '' ), broker as ( select *, row_number() over (partition by feedname order by brokerid) r from ( select distinct b.brokerid, b.name as brokername, dss.classname as feedname from downloadersymbolsettings dss inner join brokersymbollist bsl on dss.symbolid = bsl.symbolid inner join broker b on bsl.brokerid = b.brokerid where dss.enabled = 1) a ) select last.id, last.datafeedname, last.eventtimestamp, last.status, last.errordescription, last.serveraddress, last.username, note.notified, note.eventtimestamp, broker.brokername from last_feed_entry last inner join earliest_entry_after_ok after_ok on last.datafeedname = after_ok.datafeedname inner join broker on last.datafeedname = broker.feedname left outer join ok_entries ok on ok.datafeedname = last.datafeedname left outer join notified_entries note on note.datafeedname = last.datafeedname and note.r = 1 where (ok.r is null or ok.r = 1) and last.datafeedname not in ( select distinct datafeedname from last_feed_entry where status = 'OK') and extract(epoch from (last.eventtimestamp - after_ok.eventtimestamp)) > 60 * 60 and last.eventtimestamp > current_timestamp - interval '1 day' and (note.eventtimestamp is null or note.eventtimestamp < current_timestamp - interval '10 hours') and last.eventtimestamp > current_timestamp - interval '1 hour' and broker.r = 1;
Date: 2026-02-25 15:20:02 Duration: 3ms Database: postgres
19 17ms 6 2ms 3ms 2ms select client_addr, count(1) from pg_stat_activity, pg_settings WHERE name = 'max_connections' group by client_addr, setting having (client_addr is not null OR (client_addr is null and count(1) > (cast(setting as numeric) / 3 * 2))) order by count desc;Times Reported Time consuming prepare #19
Day Hour Count Duration Avg duration 15 6 17ms 2ms -
select client_addr, count(1) from pg_stat_activity, pg_settings WHERE name = 'max_connections' group by client_addr, setting having (client_addr is not null OR (client_addr is null and count(1) > (cast(setting as numeric) / 3 * 2))) order by count desc;
Date: 2026-02-25 15:50:05 Duration: 3ms Database: postgres
-
select client_addr, count(1) from pg_stat_activity, pg_settings WHERE name = 'max_connections' group by client_addr, setting having (client_addr is not null OR (client_addr is null and count(1) > (cast(setting as numeric) / 3 * 2))) order by count desc;
Date: 2026-02-25 15:00:05 Duration: 3ms Database: postgres
-
select client_addr, count(1) from pg_stat_activity, pg_settings WHERE name = 'max_connections' group by client_addr, setting having (client_addr is not null OR (client_addr is null and count(1) > (cast(setting as numeric) / 3 * 2))) order by count desc;
Date: 2026-02-25 15:30:04 Duration: 3ms Database: postgres
20 16ms 8 0ms 4ms 2ms WITH pre_symbols AS ( /* find relevant symbols */ ;Times Reported Time consuming prepare #20
Day Hour Count Duration Avg duration 15 8 16ms 2ms -
WITH pre_symbols AS ( /* find relevant symbols */ ;
Date: 2026-02-25 15:13:17 Duration: 4ms Database: postgres
-
WITH pre_symbols AS ( /* find relevant symbols */ ;
Date: 2026-02-25 15:13:17 Duration: 3ms Database: postgres
-
WITH pre_symbols AS ( /* find relevant symbols */ ;
Date: 2026-02-25 15:13:17 Duration: 3ms Database: postgres
Time consuming bind
Rank Total duration Times executed Min duration Max duration Avg duration Query 1 36s514ms 4,501 0ms 63ms 8ms WITH rar_max as ( ;Times Reported Time consuming bind #1
Day Hour Count Duration Avg duration Feb 25 15 4,501 36s514ms 8ms -
WITH rar_max as ( ;
Date: 2026-02-25 15:30:04 Duration: 63ms Database: postgres parameters: $1 = 't', $2 = '689', $3 = '7', $4 = '15', $5 = '30', $6 = '60', $7 = '120', $8 = '240', $9 = '480', $10 = '1440', $11 = '0', $12 = '', $13 = '310', $14 = '#AAPL', $15 = '#ADS', $16 = '#AIG', $17 = '#ALV', $18 = '#AMZN', $19 = '#AXP', $20 = '#BA', $21 = '#BABA', $22 = '#BAC', $23 = '#BAS', $24 = '#BAYN', $25 = '#BEI', $26 = '#BIDU', $27 = '#BMW', $28 = '#C', $29 = '#CAT', $30 = '#CBK', $31 = '#CL', $32 = '#CSCO', $33 = '#CVX', $34 = '#DAI', $35 = '#DB1', $36 = '#DBK', $37 = '#DIS', $38 = '#DPW', $39 = '#DTE', $40 = '#EBAY', $41 = '#EON', $42 = '#F', $43 = '#FB', $44 = '#FDX', $45 = '#FME', $46 = '#GE', $47 = '#GM', $48 = '#GOOG', $49 = '#GS', $50 = '#HPQ', $51 = '#IBM', $52 = '#IFX', $53 = '#INTC', $54 = '#JD', $55 = '#JNJ', $56 = '#JPM', $57 = '#KO', $58 = '#LHA', $59 = '#LMT', $60 = '#MA', $61 = '#MCD', $62 = '#META', $63 = '#MMM', $64 = '#MSFT', $65 = '#MUV2', $66 = '#NFLX', $67 = '#NKE', $68 = '#NTES', $69 = '#ORCL', $70 = '#PFE', $71 = '#PG', $72 = '#QCOM', $73 = '#RACE', $74 = '#RWE', $75 = '#SAP', $76 = '#SIE', $77 = '#T', $78 = '#UBER', $79 = '#V', $80 = '#VOW', $81 = '#WB', $82 = '#XOM', $83 = 'AUDCAD', $84 = 'AUDCHF', $85 = 'AUDJPY', $86 = 'AUDNZD', $87 = 'AUDUSD', $88 = 'AUS200', $89 = 'BRENT', $90 = 'BTCUSD', $91 = 'CADCHF', $92 = 'CADJPY', $93 = 'CHFJPY', $94 = 'CHI50', $95 = 'ESP35', $96 = 'ETHUSD', $97 = 'EU50', $98 = 'EURAUD', $99 = 'EURCAD', $100 = 'EURCHF', $101 = 'EURGBP', $102 = 'EURHUF', $103 = 'EURJPY', $104 = 'EURNZD', $105 = 'EURPLN', $106 = 'EURUSD', $107 = 'FRA40', $108 = 'GBPAUD', $109 = 'GBPCAD', $110 = 'GBPCHF', $111 = 'GBPJPY', $112 = 'GBPNZD', $113 = 'GBPUSD', $114 = 'GER30', $115 = 'HK50', $116 = 'HKCH50', $117 = 'IT40', $118 = 'JP225', $119 = 'LTCUSD', $120 = 'NAS100', $121 = 'NZDCAD', $122 = 'NZDCHF', $123 = 'NZDJPY', $124 = 'NZDUSD', $125 = 'SPX500', $126 = 'UK100', $127 = 'US30', $128 = 'USDCAD', $129 = 'USDCHF', $130 = 'USDCNH', $131 = 'USDCZK', $132 = 'USDDKK', $133 = 'USDHKD', $134 = 'USDHUF', $135 = 'USDJPY', $136 = 'USDMXN', $137 = 'USDNOK', $138 = 'USDPLN', $139 = 'USDSEK', $140 = 'USDSGD', $141 = 'USDTRY', $142 = 'USDX', $143 = 'USDZAR', $144 = 'WTI', $145 = 'XAGUSD', $146 = 'XAUUSD', $147 = '#ADS', $148 = '#ALV', $149 = '#BAS', $150 = '#BAYN', $151 = '#BEI', $152 = '#BMW', $153 = '#CBK', $154 = '#DAI', $155 = '#DB1', $156 = '#DBK', $157 = '#DPW', $158 = '#DTE', $159 = '#EON', $160 = '#FME', $161 = '#IFX', $162 = '#LHA', $163 = '#MUV2', $164 = '#RWE', $165 = '#SAP', $166 = '#SIE', $167 = '#VOW', $168 = 'AUDCAD', $169 = 'AUDCHF', $170 = 'AUDJPY', $171 = 'AUDNZD', $172 = 'AUDUSD', $173 = 'CADCHF', $174 = 'CADJPY', $175 = 'CHFJPY', $176 = 'EURAUD', $177 = 'EURCAD', $178 = 'EURCHF', $179 = 'EURGBP', $180 = 'EURHUF', $181 = 'EURJPY', $182 = 'EURNZD', $183 = 'EURPLN', $184 = 'EURUSD', $185 = 'GBPAUD', $186 = 'GBPCAD', $187 = 'GBPCHF', $188 = 'GBPJPY', $189 = 'GBPNZD', $190 = 'GBPUSD', $191 = 'NZDCAD', $192 = 'NZDCHF', $193 = 'NZDJPY', $194 = 'NZDUSD', $195 = 'USDCAD', $196 = 'USDCHF', $197 = 'USDCNH', $198 = 'USDCZK', $199 = 'USDDKK', $200 = 'USDHKD', $201 = 'USDHUF', $202 = 'USDJPY', $203 = 'USDMXN', $204 = 'USDNOK', $205 = 'USDPLN', $206 = 'USDSEK', $207 = 'USDSGD', $208 = 'USDTRY', $209 = 'USDX', $210 = 'USDZAR', $211 = 'XAGUSD', $212 = 'XAUUSD', $213 = 'BTCUSD', $214 = 'ETHUSD', $215 = 'LTCUSD', $216 = 'AUDCAD', $217 = 'AUDCHF', $218 = 'AUDJPY', $219 = 'AUDNZD', $220 = 'CADCHF', $221 = 'CADJPY', $222 = 'CHFJPY', $223 = 'EURAUD', $224 = 'EURCAD', $225 = 'EURCHF', $226 = 'EURGBP', $227 = 'EURHUF', $228 = 'EURJPY', $229 = 'EURNZD', $230 = 'EURPLN', $231 = 'GBPAUD', $232 = 'GBPCAD', $233 = 'GBPCHF', $234 = 'GBPJPY', $235 = 'GBPNZD', $236 = 'NZDCAD', $237 = 'NZDCHF', $238 = 'NZDJPY', $239 = 'USDCNH', $240 = 'USDCZK', $241 = 'USDDKK', $242 = 'USDHKD', $243 = 'USDHUF', $244 = 'USDMXN', $245 = 'USDNOK', $246 = 'USDPLN', $247 = 'USDSEK', $248 = 'USDSGD', $249 = 'USDTRY', $250 = 'USDX', $251 = 'USDZAR', $252 = 'XAGUSD', $253 = 'XAUUSD', $254 = 'BRENT', $255 = 'WTI', $256 = 'AUS200', $257 = 'CHI50', $258 = 'ESP35', $259 = 'EU50', $260 = 'FRA40', $261 = 'GER30', $262 = 'HK50', $263 = 'HKCH50', $264 = 'IT40', $265 = 'JP225', $266 = 'NAS100', $267 = 'SPX500', $268 = 'UK100', $269 = 'US30', $270 = 'AUDUSD', $271 = 'EURUSD', $272 = 'GBPUSD', $273 = 'NZDUSD', $274 = 'USDCAD', $275 = 'USDCHF', $276 = 'USDJPY', $277 = '#AAPL', $278 = '#AIG', $279 = '#AMZN', $280 = '#AXP', $281 = '#BA', $282 = '#BABA', $283 = '#BAC', $284 = '#BIDU', $285 = '#C', $286 = '#CAT', $287 = '#CL', $288 = '#CSCO', $289 = '#CVX', $290 = '#DIS', $291 = '#EBAY', $292 = '#F', $293 = '#FB', $294 = '#FDX', $295 = '#GE', $296 = '#GM', $297 = '#GOOG', $298 = '#GS', $299 = '#HPQ', $300 = '#IBM', $301 = '#INTC', $302 = '#JD', $303 = '#JNJ', $304 = '#JPM', $305 = '#KO', $306 = '#LMT', $307 = '#MA', $308 = '#MCD', $309 = '#MMM', $310 = '#MSFT', $311 = '#NFLX', $312 = '#NKE', $313 = '#NTES', $314 = '#ORCL', $315 = '#PFE', $316 = '#PG', $317 = '#QCOM', $318 = '#RACE', $319 = '#T', $320 = '#UBER', $321 = '#V', $322 = '#WB', $323 = '#XOM', $324 = '0', $325 = '', $326 = '700', $327 = '700', $328 = '0', $329 = '0', $330 = '0', $331 = 't', $332 = '10', $333 = '10'
-
WITH rar_max as ( ;
Date: 2026-02-25 15:25:05 Duration: 42ms Database: postgres parameters: $1 = '607740240690638301', $2 = '607740240690638301', $3 = '607740240690638301'
-
WITH rar_max as ( ;
Date: 2026-02-25 15:26:36 Duration: 39ms Database: postgres parameters: $1 = '607734761663424301', $2 = '607734761663424301', $3 = '607734761663424301'
2 12s342ms 27,382 0ms 20ms 0ms SELECT ;Times Reported Time consuming bind #2
Day Hour Count Duration Avg duration 15 27,382 12s342ms 0ms -
SELECT ;
Date: 2026-02-25 15:00:04 Duration: 20ms Database: postgres parameters: $1 = '558', $2 = '0', $3 = '0', $4 = 'USDHUF', $5 = 'USDHUF'
-
SELECT ;
Date: 2026-02-25 15:30:02 Duration: 19ms Database: postgres parameters: $1 = '558', $2 = '0', $3 = '0', $4 = 'ES35', $5 = 'ES35'
-
SELECT ;
Date: 2026-02-25 15:21:02 Duration: 19ms Database: postgres parameters: $1 = '667', $2 = '0', $3 = '0', $4 = 'AUDSGD', $5 = 'AUDSGD'
3 2s862ms 1,408 1ms 7ms 2ms SELECT symbolid, ;Times Reported Time consuming bind #3
Day Hour Count Duration Avg duration 15 1,408 2s862ms 2ms -
SELECT symbolid, ;
Date: 2026-02-25 15:00:33 Duration: 7ms Database: postgres parameters: $1 = 'GLOBALGTMT5', $2 = '15', $3 = 'EURGBP', $4 = 'EURHUF'
-
SELECT symbolid, ;
Date: 2026-02-25 15:47:10 Duration: 6ms Database: postgres parameters: $1 = 'MILLENNIUMPF', $2 = '15', $3 = 'GBPJPY.ID'
-
SELECT symbolid, ;
Date: 2026-02-25 15:46:13 Duration: 6ms Database: postgres parameters: $1 = 'MILLENNIUMPF', $2 = '15', $3 = 'GBPUSD.FX', $4 = 'GBPUSD'
4 1s671ms 956 1ms 20ms 1ms SELECT s.symbolid, dss.downloadfrequency, dss.downloadersymbol;Times Reported Time consuming bind #4
Day Hour Count Duration Avg duration 15 956 1s671ms 1ms -
SELECT s.symbolid, dss.downloadfrequency, dss.downloadersymbol;
Date: 2026-02-25 15:30:52 Duration: 20ms Database: postgres parameters: $1 = 'MILLENNIUMPF'
-
SELECT s.symbolid, dss.downloadfrequency, dss.downloadersymbol;
Date: 2026-02-25 15:30:52 Duration: 5ms Database: postgres parameters: $1 = 'PEPPERSTONEMT5'
-
SELECT s.symbolid, dss.downloadfrequency, dss.downloadersymbol;
Date: 2026-02-25 15:15:47 Duration: 3ms Database: postgres parameters: $1 = 'BDSWISS'
5 586ms 23 0ms 76ms 25ms with wh_patitioned as ( ;Times Reported Time consuming bind #5
Day Hour Count Duration Avg duration 15 23 586ms 25ms -
with wh_patitioned as ( ;
Date: 2026-02-25 15:30:03 Duration: 76ms Database: postgres parameters: $1 = '558', $2 = '558', $3 = '558', $4 = '558', $5 = '558', $6 = '558', $7 = '558', $8 = '558', $9 = '558'
-
with wh_patitioned as ( ;
Date: 2026-02-25 15:20:02 Duration: 48ms Database: postgres parameters: $1 = '558', $2 = '558', $3 = '558', $4 = '558', $5 = '558', $6 = '558', $7 = '558', $8 = '558', $9 = '558'
-
with wh_patitioned as ( ;
Date: 2026-02-25 15:03:43 Duration: 47ms Database: postgres parameters: $1 = '558', $2 = '558', $3 = '558', $4 = '558', $5 = '558', $6 = '558', $7 = '558', $8 = '558', $9 = '558'
6 576ms 22,172 0ms 12ms 0ms select 1;Times Reported Time consuming bind #6
Day Hour Count Duration Avg duration 15 22,172 576ms 0ms -
select 1;
Date: 2026-02-25 15:50:41 Duration: 12ms Database: postgres
-
select 1;
Date: 2026-02-25 15:05:54 Duration: 8ms Database: postgres
-
select 1;
Date: 2026-02-25 15:21:02 Duration: 8ms Database: postgres
7 575ms 82 0ms 18ms 7ms WITH /*Latest.JapSticks*/ all_results AS ( SELECT ;Times Reported Time consuming bind #7
Day Hour Count Duration Avg duration 15 82 575ms 7ms -
WITH /*Latest.JapSticks*/ all_results AS ( SELECT ;
Date: 2026-02-25 15:20:44 Duration: 18ms Database: postgres parameters: $1 = '667', $2 = '0', $3 = '0', $4 = '0', $5 = '', $6 = '0', $7 = '', $8 = '0', $9 = '', $10 = '0', $11 = '0'
-
WITH /*Latest.JapSticks*/ all_results AS ( SELECT ;
Date: 2026-02-25 15:03:36 Duration: 18ms Database: postgres parameters: $1 = '489', $2 = '0', $3 = '0', $4 = '0', $5 = '', $6 = '0', $7 = '', $8 = '0', $9 = '', $10 = '0', $11 = '0'
-
WITH /*Latest.JapSticks*/ all_results AS ( SELECT ;
Date: 2026-02-25 15:51:10 Duration: 18ms Database: postgres parameters: $1 = '489', $2 = '0', $3 = '0', $4 = '0', $5 = '', $6 = '0', $7 = '', $8 = '0', $9 = '', $10 = '0', $11 = '0'
8 567ms 75 4ms 26ms 7ms WITH last_candle AS ( ;Times Reported Time consuming bind #8
Day Hour Count Duration Avg duration 15 75 567ms 7ms -
WITH last_candle AS ( ;
Date: 2026-02-25 15:20:02 Duration: 26ms Database: postgres parameters: $1 = '558', $2 = '558'
-
WITH last_candle AS ( ;
Date: 2026-02-25 15:20:06 Duration: 14ms Database: postgres parameters: $1 = '667', $2 = '667'
-
WITH last_candle AS ( ;
Date: 2026-02-25 15:00:12 Duration: 13ms Database: postgres parameters: $1 = '558', $2 = '558'
9 387ms 12 28ms 47ms 32ms with sym_info as ( ;Times Reported Time consuming bind #9
Day Hour Count Duration Avg duration 15 12 387ms 32ms -
with sym_info as ( ;
Date: 2026-02-25 15:06:40 Duration: 47ms Database: postgres parameters: $1 = '620', $2 = 'Forex', $3 = 'Forex', $4 = '620', $5 = 'Forex', $6 = '620', $7 = '620', $8 = 'Forex', $9 = '620'
-
with sym_info as ( ;
Date: 2026-02-25 15:36:56 Duration: 44ms Database: postgres parameters: $1 = '692', $2 = 'Forex', $3 = 'Forex', $4 = '692', $5 = 'Forex', $6 = '692', $7 = '692', $8 = 'Forex', $9 = '692'
-
with sym_info as ( ;
Date: 2026-02-25 15:36:51 Duration: 32ms Database: postgres parameters: $1 = '627', $2 = 'Forex', $3 = 'Forex', $4 = '627', $5 = 'Forex', $6 = '627', $7 = '627', $8 = 'Forex', $9 = '627'
10 265ms 5,551 0ms 0ms 0ms INSERT INTO T15 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;Times Reported Time consuming bind #10
Day Hour Count Duration Avg duration 15 5,551 265ms 0ms -
INSERT INTO T15 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;
Date: 2026-02-25 15:46:11 Duration: 0ms Database: postgres parameters: $1 = '2026-02-25 15:30:00', $2 = '2.26358', $3 = '2.26431', $4 = '2.263515', $5 = '2.263785', $6 = '1586', $7 = '515840230500755300', $8 = '0', $9 = '2026-02-25 15:46:11.762', $10 = '2026-02-25 15:46:11.676', $11 = '2.26358', $12 = '2.26431', $13 = '2.263515', $14 = '2.263785', $15 = '1586', $16 = '0', $17 = '2026-02-25 15:46:11.762', $18 = '2026-02-25 15:46:11.676'
-
INSERT INTO T15 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;
Date: 2026-02-25 15:30:40 Duration: 0ms Database: postgres parameters: $1 = '2026-02-25 15:00:00', $2 = '202.272', $3 = '202.378', $4 = '202.255', $5 = '202.334', $6 = '1326', $7 = '515840230445516300', $8 = '0', $9 = '2026-02-25 15:30:40.197', $10 = '2026-02-25 15:30:39.949', $11 = '202.272', $12 = '202.378', $13 = '202.255', $14 = '202.334', $15 = '1326', $16 = '0', $17 = '2026-02-25 15:30:40.197', $18 = '2026-02-25 15:30:39.949'
-
INSERT INTO T15 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;
Date: 2026-02-25 15:30:52 Duration: 0ms Database: postgres parameters: $1 = '2026-02-25 13:15:00', $2 = '123.8875', $3 = '123.8875', $4 = '123.8375', $5 = '123.8625', $6 = '182000000', $7 = '515840249645115300', $8 = '0', $9 = '2026-02-25 15:30:52.564', $10 = '2026-02-25 15:30:52.263', $11 = '123.8875', $12 = '123.8875', $13 = '123.8375', $14 = '123.8625', $15 = '182000000', $16 = '0', $17 = '2026-02-25 15:30:52.564', $18 = '2026-02-25 15:30:52.263'
11 252ms 3,125 0ms 0ms 0ms INSERT INTO T30 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;Times Reported Time consuming bind #11
Day Hour Count Duration Avg duration 15 3,125 252ms 0ms -
INSERT INTO T30 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;
Date: 2026-02-25 15:00:59 Duration: 0ms Database: postgres parameters: $1 = '2026-02-25 15:30:00', $2 = '1.369655', $3 = '1.36975', $4 = '1.36892', $5 = '1.369015', $6 = '1580', $7 = '515840249463916300', $8 = '0', $9 = '2026-02-25 15:00:59.933', $10 = '2026-02-25 15:00:59.897', $11 = '1.369655', $12 = '1.36975', $13 = '1.36892', $14 = '1.369015', $15 = '1580', $16 = '0', $17 = '2026-02-25 15:00:59.933', $18 = '2026-02-25 15:00:59.897'
-
INSERT INTO T30 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;
Date: 2026-02-25 15:31:37 Duration: 0ms Database: postgres parameters: $1 = '2026-02-25 15:00:00', $2 = '111.125', $3 = '111.1825', $4 = '111.1125', $5 = '111.149', $6 = '4192', $7 = '515840230426879300', $8 = '0', $9 = '2026-02-25 15:31:37.091', $10 = '2026-02-25 15:31:37.09', $11 = '111.125', $12 = '111.1825', $13 = '111.1125', $14 = '111.149', $15 = '4192', $16 = '0', $17 = '2026-02-25 15:31:37.091', $18 = '2026-02-25 15:31:37.09'
-
INSERT INTO T30 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;
Date: 2026-02-25 15:41:46 Duration: 0ms Database: postgres parameters: $1 = '2026-02-25 14:30:00', $2 = '26844.8', $3 = '26890.3', $4 = '26844.7', $5 = '26887.1', $6 = '2720', $7 = '515840247933633300', $8 = '0', $9 = '2026-02-25 15:41:46.706', $10 = '2026-02-25 15:41:46.617', $11 = '26844.8', $12 = '26890.3', $13 = '26844.7', $14 = '26887.1', $15 = '2720', $16 = '0', $17 = '2026-02-25 15:41:46.706', $18 = '2026-02-25 15:41:46.617'
12 175ms 2,058 0ms 1ms 0ms INSERT INTO T60 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;Times Reported Time consuming bind #12
Day Hour Count Duration Avg duration 15 2,058 175ms 0ms -
INSERT INTO T60 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;
Date: 2026-02-25 15:00:16 Duration: 1ms Database: postgres parameters: $1 = '2026-02-25 14:00:00', $2 = '1.61337', $3 = '1.614', $4 = '1.61311', $5 = '1.61336', $6 = '7723', $7 = '515840243897304300', $8 = '0', $9 = '2026-02-25 15:00:16.881', $10 = '2026-02-25 15:00:16.88', $11 = '1.61337', $12 = '1.614', $13 = '1.61311', $14 = '1.61336', $15 = '7723', $16 = '0', $17 = '2026-02-25 15:00:16.881', $18 = '2026-02-25 15:00:16.88'
-
INSERT INTO T60 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;
Date: 2026-02-25 15:06:00 Duration: 0ms Database: postgres parameters: $1 = '2026-02-25 14:00:00', $2 = '1.613465', $3 = '1.61408', $4 = '1.613215', $5 = '1.613475', $6 = '6930', $7 = '605679104145108300', $8 = '0', $9 = '2026-02-25 15:06:00.543', $10 = '2026-02-25 15:06:00.543', $11 = '1.613465', $12 = '1.61408', $13 = '1.613215', $14 = '1.613475', $15 = '6930', $16 = '0', $17 = '2026-02-25 15:06:00.543', $18 = '2026-02-25 15:06:00.543'
-
INSERT INTO T60 (pricedatetime, open, high, low, close, volume, symbolid, bsf, sastdatetimewritten, sastdatetimereceived) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10) ON CONFLICT (pricedatetime, symbolid) DO UPDATE SET open = $11, high = $12, low = $13, close = $14, volume = $15, bsf = $16, sastdatetimewritten = $17, sastdatetimereceived = $18;
Date: 2026-02-25 15:01:31 Duration: 0ms Database: postgres parameters: $1 = '2026-02-25 14:00:00', $2 = '90.465', $3 = '90.725', $4 = '90.135', $5 = '90.6', $6 = '3276', $7 = '515840230623790300', $8 = '0', $9 = '2026-02-25 15:01:31.282', $10 = '2026-02-25 15:01:31.281', $11 = '90.465', $12 = '90.725', $13 = '90.135', $14 = '90.6', $15 = '3276', $16 = '0', $17 = '2026-02-25 15:01:31.282', $18 = '2026-02-25 15:01:31.281'
13 112ms 107 0ms 2ms 1ms SELECT timegranularity FROM brokersymbollist bsl INNER JOIN symbols s ON bsl.symbolid = s.symbolid INNER JOIN downloadersymbolsettings dss on s.symbolid = dss.symbolid LEFT OUTER JOIN brokerinstrumentmapping bdfi ON bdfi.brokerid = $1 AND dss.datafeedinstrumentid = bdfi.datafeedinstrumentid WHERE s.nonliquid = 0 and s.deleted = 0 and dss.enabled = 1 AND s.symbol ILIKE $2 AND bsl.brokerid = $3 AND timegranularity >= 15 ORDER BY timegranularity LIMIT 1;Times Reported Time consuming bind #13
Day Hour Count Duration Avg duration 15 107 112ms 1ms -
SELECT timegranularity FROM brokersymbollist bsl INNER JOIN symbols s ON bsl.symbolid = s.symbolid INNER JOIN downloadersymbolsettings dss on s.symbolid = dss.symbolid LEFT OUTER JOIN brokerinstrumentmapping bdfi ON bdfi.brokerid = $1 AND dss.datafeedinstrumentid = bdfi.datafeedinstrumentid WHERE s.nonliquid = 0 and s.deleted = 0 and dss.enabled = 1 AND s.symbol ILIKE $2 AND bsl.brokerid = $3 AND timegranularity >= 15 ORDER BY timegranularity LIMIT 1;
Date: 2026-02-25 15:23:14 Duration: 2ms Database: postgres parameters: $1 = '689', $2 = 'XAUUSD', $3 = '689'
-
SELECT timegranularity FROM brokersymbollist bsl INNER JOIN symbols s ON bsl.symbolid = s.symbolid INNER JOIN downloadersymbolsettings dss on s.symbolid = dss.symbolid LEFT OUTER JOIN brokerinstrumentmapping bdfi ON bdfi.brokerid = $1 AND dss.datafeedinstrumentid = bdfi.datafeedinstrumentid WHERE s.nonliquid = 0 and s.deleted = 0 and dss.enabled = 1 AND s.symbol ILIKE $2 AND bsl.brokerid = $3 AND timegranularity >= 15 ORDER BY timegranularity LIMIT 1;
Date: 2026-02-25 15:46:50 Duration: 1ms Database: postgres parameters: $1 = '632', $2 = 'GBPUSD', $3 = '632'
-
SELECT timegranularity FROM brokersymbollist bsl INNER JOIN symbols s ON bsl.symbolid = s.symbolid INNER JOIN downloadersymbolsettings dss on s.symbolid = dss.symbolid LEFT OUTER JOIN brokerinstrumentmapping bdfi ON bdfi.brokerid = $1 AND dss.datafeedinstrumentid = bdfi.datafeedinstrumentid WHERE s.nonliquid = 0 and s.deleted = 0 and dss.enabled = 1 AND s.symbol ILIKE $2 AND bsl.brokerid = $3 AND timegranularity >= 15 ORDER BY timegranularity LIMIT 1;
Date: 2026-02-25 15:00:29 Duration: 1ms Database: postgres parameters: $1 = '558', $2 = 'AUDNZD', $3 = '558'
14 79ms 364 0ms 4ms 0ms /*server.CPResult*/ SELECT patternid, resy0, resy1, supporty0, supporty1, predictiontimeto, patternstarttime, s.symbolid, resx0, resx1, supportx0, supportx1, symbol, longname, shortname, timegranularity, patternendtime, pattern, a.direction, trendchange, patternlengthbars, patternquality, resultuid as uid, breakout, initialtrend, volumeincrease, symmetry as uniformity, predictionpricefrom, predictionpriceto, noise, exchange, breakout, dtt.absolutetimezoneoffset as tzOs, dtt.timezone as tz FROM autochartist_results a INNER JOIN downloadersymbolsettings dss on a.symbolid = dss.symbolid INNER JOIN datafeedstimetable dtt ON dss.classname = dtt.classname inner join symbols s on a.symbolid = s.symbolid inner join patterns p on p.patternname = a.pattern where resultuid = $1;Times Reported Time consuming bind #14
Day Hour Count Duration Avg duration 15 364 79ms 0ms -
/*server.CPResult*/ SELECT patternid, resy0, resy1, supporty0, supporty1, predictiontimeto, patternstarttime, s.symbolid, resx0, resx1, supportx0, supportx1, symbol, longname, shortname, timegranularity, patternendtime, pattern, a.direction, trendchange, patternlengthbars, patternquality, resultuid as uid, breakout, initialtrend, volumeincrease, symmetry as uniformity, predictionpricefrom, predictionpriceto, noise, exchange, breakout, dtt.absolutetimezoneoffset as tzOs, dtt.timezone as tz FROM autochartist_results a INNER JOIN downloadersymbolsettings dss on a.symbolid = dss.symbolid INNER JOIN datafeedstimetable dtt ON dss.classname = dtt.classname inner join symbols s on a.symbolid = s.symbolid inner join patterns p on p.patternname = a.pattern where resultuid = $1;
Date: 2026-02-25 15:20:25 Duration: 4ms Database: postgres parameters: $1 = '607739827702027301'
-
/*server.CPResult*/ SELECT patternid, resy0, resy1, supporty0, supporty1, predictiontimeto, patternstarttime, s.symbolid, resx0, resx1, supportx0, supportx1, symbol, longname, shortname, timegranularity, patternendtime, pattern, a.direction, trendchange, patternlengthbars, patternquality, resultuid as uid, breakout, initialtrend, volumeincrease, symmetry as uniformity, predictionpricefrom, predictionpriceto, noise, exchange, breakout, dtt.absolutetimezoneoffset as tzOs, dtt.timezone as tz FROM autochartist_results a INNER JOIN downloadersymbolsettings dss on a.symbolid = dss.symbolid INNER JOIN datafeedstimetable dtt ON dss.classname = dtt.classname inner join symbols s on a.symbolid = s.symbolid inner join patterns p on p.patternname = a.pattern where resultuid = $1;
Date: 2026-02-25 15:07:29 Duration: 4ms Database: postgres parameters: $1 = '607739947922961301'
-
/*server.CPResult*/ SELECT patternid, resy0, resy1, supporty0, supporty1, predictiontimeto, patternstarttime, s.symbolid, resx0, resx1, supportx0, supportx1, symbol, longname, shortname, timegranularity, patternendtime, pattern, a.direction, trendchange, patternlengthbars, patternquality, resultuid as uid, breakout, initialtrend, volumeincrease, symmetry as uniformity, predictionpricefrom, predictionpriceto, noise, exchange, breakout, dtt.absolutetimezoneoffset as tzOs, dtt.timezone as tz FROM autochartist_results a INNER JOIN downloadersymbolsettings dss on a.symbolid = dss.symbolid INNER JOIN datafeedstimetable dtt ON dss.classname = dtt.classname inner join symbols s on a.symbolid = s.symbolid inner join patterns p on p.patternname = a.pattern where resultuid = $1;
Date: 2026-02-25 15:37:13 Duration: 3ms Database: postgres parameters: $1 = '607740301494434301'
15 70ms 117 0ms 3ms 0ms SELECT NULL AS TABLE_CAT, n.nspname AS TABLE_SCHEM, c.relname AS TABLE_NAME, CASE n.nspname ~ '^pg_' OR n.nspname = 'information_schema' WHEN true THEN CASE WHEN n.nspname = 'pg_catalog' OR n.nspname = 'information_schema' THEN CASE c.relkind WHEN 'r' THEN 'SYSTEM TABLE' WHEN 'v' THEN 'SYSTEM VIEW' WHEN 'i' THEN 'SYSTEM INDEX' ELSE NULL END WHEN n.nspname = 'pg_toast' THEN CASE c.relkind WHEN 'r' THEN 'SYSTEM TOAST TABLE' WHEN 'i' THEN 'SYSTEM TOAST INDEX' ELSE NULL END ELSE CASE c.relkind WHEN 'r' THEN 'TEMPORARY TABLE' WHEN 'p' THEN 'TEMPORARY TABLE' WHEN 'i' THEN 'TEMPORARY INDEX' WHEN 'S' THEN 'TEMPORARY SEQUENCE' WHEN 'v' THEN 'TEMPORARY VIEW' ELSE NULL END END WHEN false THEN CASE c.relkind WHEN 'r' THEN 'TABLE' WHEN 'p' THEN 'PARTITIONED TABLE' WHEN 'i' THEN 'INDEX' WHEN 'S' THEN 'SEQUENCE' WHEN 'v' THEN 'VIEW' WHEN 'c' THEN 'TYPE' WHEN 'f' THEN 'FOREIGN TABLE' WHEN 'm' THEN 'MATERIALIZED VIEW' ELSE NULL END ELSE NULL END AS TABLE_TYPE, d.description AS REMARKS, '' as TYPE_CAT, '' as TYPE_SCHEM, '' as TYPE_NAME, '' AS SELF_REFERENCING_COL_NAME, '' AS REF_GENERATION FROM pg_catalog.pg_namespace n, pg_catalog.pg_class c LEFT JOIN pg_catalog.pg_description d ON (c.oid = d.objoid AND d.objsubid = 0) LEFT JOIN pg_catalog.pg_class dc ON (d.classoid = dc.oid AND dc.relname = 'pg_class') LEFT JOIN pg_catalog.pg_namespace dn ON (dn.oid = dc.relnamespace AND dn.nspname = 'pg_catalog') WHERE c.relnamespace = n.oid AND c.relname LIKE 'PROBABLYNOT' AND (false OR (c.relkind = 'r' AND n.nspname !~ '^pg_' AND n.nspname <> 'information_schema')) ORDER BY TABLE_TYPE, TABLE_SCHEM, TABLE_NAME;Times Reported Time consuming bind #15
Day Hour Count Duration Avg duration 15 117 70ms 0ms -
SELECT NULL AS TABLE_CAT, n.nspname AS TABLE_SCHEM, c.relname AS TABLE_NAME, CASE n.nspname ~ '^pg_' OR n.nspname = 'information_schema' WHEN true THEN CASE WHEN n.nspname = 'pg_catalog' OR n.nspname = 'information_schema' THEN CASE c.relkind WHEN 'r' THEN 'SYSTEM TABLE' WHEN 'v' THEN 'SYSTEM VIEW' WHEN 'i' THEN 'SYSTEM INDEX' ELSE NULL END WHEN n.nspname = 'pg_toast' THEN CASE c.relkind WHEN 'r' THEN 'SYSTEM TOAST TABLE' WHEN 'i' THEN 'SYSTEM TOAST INDEX' ELSE NULL END ELSE CASE c.relkind WHEN 'r' THEN 'TEMPORARY TABLE' WHEN 'p' THEN 'TEMPORARY TABLE' WHEN 'i' THEN 'TEMPORARY INDEX' WHEN 'S' THEN 'TEMPORARY SEQUENCE' WHEN 'v' THEN 'TEMPORARY VIEW' ELSE NULL END END WHEN false THEN CASE c.relkind WHEN 'r' THEN 'TABLE' WHEN 'p' THEN 'PARTITIONED TABLE' WHEN 'i' THEN 'INDEX' WHEN 'S' THEN 'SEQUENCE' WHEN 'v' THEN 'VIEW' WHEN 'c' THEN 'TYPE' WHEN 'f' THEN 'FOREIGN TABLE' WHEN 'm' THEN 'MATERIALIZED VIEW' ELSE NULL END ELSE NULL END AS TABLE_TYPE, d.description AS REMARKS, '' as TYPE_CAT, '' as TYPE_SCHEM, '' as TYPE_NAME, '' AS SELF_REFERENCING_COL_NAME, '' AS REF_GENERATION FROM pg_catalog.pg_namespace n, pg_catalog.pg_class c LEFT JOIN pg_catalog.pg_description d ON (c.oid = d.objoid AND d.objsubid = 0) LEFT JOIN pg_catalog.pg_class dc ON (d.classoid = dc.oid AND dc.relname = 'pg_class') LEFT JOIN pg_catalog.pg_namespace dn ON (dn.oid = dc.relnamespace AND dn.nspname = 'pg_catalog') WHERE c.relnamespace = n.oid AND c.relname LIKE 'PROBABLYNOT' AND (false OR (c.relkind = 'r' AND n.nspname !~ '^pg_' AND n.nspname <> 'information_schema')) ORDER BY TABLE_TYPE, TABLE_SCHEM, TABLE_NAME;
Date: 2026-02-25 15:13:17 Duration: 3ms Database: postgres
-
SELECT NULL AS TABLE_CAT, n.nspname AS TABLE_SCHEM, c.relname AS TABLE_NAME, CASE n.nspname ~ '^pg_' OR n.nspname = 'information_schema' WHEN true THEN CASE WHEN n.nspname = 'pg_catalog' OR n.nspname = 'information_schema' THEN CASE c.relkind WHEN 'r' THEN 'SYSTEM TABLE' WHEN 'v' THEN 'SYSTEM VIEW' WHEN 'i' THEN 'SYSTEM INDEX' ELSE NULL END WHEN n.nspname = 'pg_toast' THEN CASE c.relkind WHEN 'r' THEN 'SYSTEM TOAST TABLE' WHEN 'i' THEN 'SYSTEM TOAST INDEX' ELSE NULL END ELSE CASE c.relkind WHEN 'r' THEN 'TEMPORARY TABLE' WHEN 'p' THEN 'TEMPORARY TABLE' WHEN 'i' THEN 'TEMPORARY INDEX' WHEN 'S' THEN 'TEMPORARY SEQUENCE' WHEN 'v' THEN 'TEMPORARY VIEW' ELSE NULL END END WHEN false THEN CASE c.relkind WHEN 'r' THEN 'TABLE' WHEN 'p' THEN 'PARTITIONED TABLE' WHEN 'i' THEN 'INDEX' WHEN 'S' THEN 'SEQUENCE' WHEN 'v' THEN 'VIEW' WHEN 'c' THEN 'TYPE' WHEN 'f' THEN 'FOREIGN TABLE' WHEN 'm' THEN 'MATERIALIZED VIEW' ELSE NULL END ELSE NULL END AS TABLE_TYPE, d.description AS REMARKS, '' as TYPE_CAT, '' as TYPE_SCHEM, '' as TYPE_NAME, '' AS SELF_REFERENCING_COL_NAME, '' AS REF_GENERATION FROM pg_catalog.pg_namespace n, pg_catalog.pg_class c LEFT JOIN pg_catalog.pg_description d ON (c.oid = d.objoid AND d.objsubid = 0) LEFT JOIN pg_catalog.pg_class dc ON (d.classoid = dc.oid AND dc.relname = 'pg_class') LEFT JOIN pg_catalog.pg_namespace dn ON (dn.oid = dc.relnamespace AND dn.nspname = 'pg_catalog') WHERE c.relnamespace = n.oid AND c.relname LIKE 'PROBABLYNOT' AND (false OR (c.relkind = 'r' AND n.nspname !~ '^pg_' AND n.nspname <> 'information_schema')) ORDER BY TABLE_TYPE, TABLE_SCHEM, TABLE_NAME;
Date: 2026-02-25 15:13:18 Duration: 1ms Database: postgres
-
SELECT NULL AS TABLE_CAT, n.nspname AS TABLE_SCHEM, c.relname AS TABLE_NAME, CASE n.nspname ~ '^pg_' OR n.nspname = 'information_schema' WHEN true THEN CASE WHEN n.nspname = 'pg_catalog' OR n.nspname = 'information_schema' THEN CASE c.relkind WHEN 'r' THEN 'SYSTEM TABLE' WHEN 'v' THEN 'SYSTEM VIEW' WHEN 'i' THEN 'SYSTEM INDEX' ELSE NULL END WHEN n.nspname = 'pg_toast' THEN CASE c.relkind WHEN 'r' THEN 'SYSTEM TOAST TABLE' WHEN 'i' THEN 'SYSTEM TOAST INDEX' ELSE NULL END ELSE CASE c.relkind WHEN 'r' THEN 'TEMPORARY TABLE' WHEN 'p' THEN 'TEMPORARY TABLE' WHEN 'i' THEN 'TEMPORARY INDEX' WHEN 'S' THEN 'TEMPORARY SEQUENCE' WHEN 'v' THEN 'TEMPORARY VIEW' ELSE NULL END END WHEN false THEN CASE c.relkind WHEN 'r' THEN 'TABLE' WHEN 'p' THEN 'PARTITIONED TABLE' WHEN 'i' THEN 'INDEX' WHEN 'S' THEN 'SEQUENCE' WHEN 'v' THEN 'VIEW' WHEN 'c' THEN 'TYPE' WHEN 'f' THEN 'FOREIGN TABLE' WHEN 'm' THEN 'MATERIALIZED VIEW' ELSE NULL END ELSE NULL END AS TABLE_TYPE, d.description AS REMARKS, '' as TYPE_CAT, '' as TYPE_SCHEM, '' as TYPE_NAME, '' AS SELF_REFERENCING_COL_NAME, '' AS REF_GENERATION FROM pg_catalog.pg_namespace n, pg_catalog.pg_class c LEFT JOIN pg_catalog.pg_description d ON (c.oid = d.objoid AND d.objsubid = 0) LEFT JOIN pg_catalog.pg_class dc ON (d.classoid = dc.oid AND dc.relname = 'pg_class') LEFT JOIN pg_catalog.pg_namespace dn ON (dn.oid = dc.relnamespace AND dn.nspname = 'pg_catalog') WHERE c.relnamespace = n.oid AND c.relname LIKE 'PROBABLYNOT' AND (false OR (c.relkind = 'r' AND n.nspname !~ '^pg_' AND n.nspname <> 'information_schema')) ORDER BY TABLE_TYPE, TABLE_SCHEM, TABLE_NAME;
Date: 2026-02-25 15:13:18 Duration: 1ms Database: postgres
16 70ms 13 3ms 6ms 5ms SELECT DISTINCT ON (basegroupname, symbol) ;Times Reported Time consuming bind #16
Day Hour Count Duration Avg duration 15 13 70ms 5ms -
SELECT DISTINCT ON (basegroupname, symbol) ;
Date: 2026-02-25 15:53:55 Duration: 6ms Database: postgres parameters: $1 = '667', $2 = '667'
-
SELECT DISTINCT ON (basegroupname, symbol) ;
Date: 2026-02-25 15:46:09 Duration: 6ms Database: postgres parameters: $1 = '627', $2 = '627'
-
SELECT DISTINCT ON (basegroupname, symbol) ;
Date: 2026-02-25 15:17:24 Duration: 6ms Database: postgres parameters: $1 = '538', $2 = '538'
17 67ms 247 0ms 4ms 0ms /*server.KeyLevelResult*/ SELECT ResultUID AS ruid, s.symbolid AS sid, symbol AS sym, longname, shortname, Exchange AS e, timegranularity AS tg, a.PatternID AS pid, a.direction AS d, a.patternprice as pp, atbaridentified AS pet, CASE WHEN (x9 != '') THEN x9 WHEN (x8 != '') THEN x8 WHEN (x7 != '') THEN x7 WHEN (x6 != '') THEN x6 WHEN (x5 != '') THEN x5 WHEN (x4 != '') THEN x4 WHEN (x3 != '') THEN x3 WHEN (x2 != '') THEN x2 END AS pst, PatternPrice AS patp, x0, x1, x2, CASE WHEN (x3 != '') THEN x3 ELSE '1900-01-01' END as x3, CASE WHEN (x4 != '') THEN x4 ELSE '1900-01-01' END as x4, CASE WHEN (x5 != '') THEN x5 ELSE '1900-01-01' END as x5, CASE WHEN (x6 != '') THEN x6 ELSE '1900-01-01' END as x6, CASE WHEN (x7 != '') THEN x7 ELSE '1900-01-01' END as x7, CASE WHEN (x8 != '') THEN x8 ELSE '1900-01-01' END as x8, CASE WHEN (x9 != '') THEN x9 ELSE '1900-01-01' END as x9, errorMargin as erm, breakoutprice as pE, breakoutbars as be, breakout, atbaridentified as atBar, PatternLengthBars AS l, Bandwidth AS bw, QtyTP AS qtp, p.patternname as patternname, dtt.absolutetimezoneoffset as tzOs, dtt.timezone as tz, approachingtimestamp AS apt, approachingregion as apr, predictionpricefrom as ppf, predictionpriceto as ppt, predictiontimefrom as ptf, predictiontimebars as ptb FROM keylevels_results a INNER JOIN downloadersymbolsettings dss on a.symbolid = dss.symbolid INNER JOIN datafeedstimetable dtt ON dss.classname = dtt.classname inner join symbols s on a.symbolid = s.symbolid INNER JOIN hrspatterns p on a.patternid = p.patternid where resultuid = $1 and dtt.dayofweek = 3;Times Reported Time consuming bind #17
Day Hour Count Duration Avg duration 15 247 67ms 0ms -
/*server.KeyLevelResult*/ SELECT ResultUID AS ruid, s.symbolid AS sid, symbol AS sym, longname, shortname, Exchange AS e, timegranularity AS tg, a.PatternID AS pid, a.direction AS d, a.patternprice as pp, atbaridentified AS pet, CASE WHEN (x9 != '') THEN x9 WHEN (x8 != '') THEN x8 WHEN (x7 != '') THEN x7 WHEN (x6 != '') THEN x6 WHEN (x5 != '') THEN x5 WHEN (x4 != '') THEN x4 WHEN (x3 != '') THEN x3 WHEN (x2 != '') THEN x2 END AS pst, PatternPrice AS patp, x0, x1, x2, CASE WHEN (x3 != '') THEN x3 ELSE '1900-01-01' END as x3, CASE WHEN (x4 != '') THEN x4 ELSE '1900-01-01' END as x4, CASE WHEN (x5 != '') THEN x5 ELSE '1900-01-01' END as x5, CASE WHEN (x6 != '') THEN x6 ELSE '1900-01-01' END as x6, CASE WHEN (x7 != '') THEN x7 ELSE '1900-01-01' END as x7, CASE WHEN (x8 != '') THEN x8 ELSE '1900-01-01' END as x8, CASE WHEN (x9 != '') THEN x9 ELSE '1900-01-01' END as x9, errorMargin as erm, breakoutprice as pE, breakoutbars as be, breakout, atbaridentified as atBar, PatternLengthBars AS l, Bandwidth AS bw, QtyTP AS qtp, p.patternname as patternname, dtt.absolutetimezoneoffset as tzOs, dtt.timezone as tz, approachingtimestamp AS apt, approachingregion as apr, predictionpricefrom as ppf, predictionpriceto as ppt, predictiontimefrom as ptf, predictiontimebars as ptb FROM keylevels_results a INNER JOIN downloadersymbolsettings dss on a.symbolid = dss.symbolid INNER JOIN datafeedstimetable dtt ON dss.classname = dtt.classname inner join symbols s on a.symbolid = s.symbolid INNER JOIN hrspatterns p on a.patternid = p.patternid where resultuid = $1 and dtt.dayofweek = 3;
Date: 2026-02-25 15:20:44 Duration: 4ms Database: postgres parameters: $1 = '607739829724323303'
-
/*server.KeyLevelResult*/ SELECT ResultUID AS ruid, s.symbolid AS sid, symbol AS sym, longname, shortname, Exchange AS e, timegranularity AS tg, a.PatternID AS pid, a.direction AS d, a.patternprice as pp, atbaridentified AS pet, CASE WHEN (x9 != '') THEN x9 WHEN (x8 != '') THEN x8 WHEN (x7 != '') THEN x7 WHEN (x6 != '') THEN x6 WHEN (x5 != '') THEN x5 WHEN (x4 != '') THEN x4 WHEN (x3 != '') THEN x3 WHEN (x2 != '') THEN x2 END AS pst, PatternPrice AS patp, x0, x1, x2, CASE WHEN (x3 != '') THEN x3 ELSE '1900-01-01' END as x3, CASE WHEN (x4 != '') THEN x4 ELSE '1900-01-01' END as x4, CASE WHEN (x5 != '') THEN x5 ELSE '1900-01-01' END as x5, CASE WHEN (x6 != '') THEN x6 ELSE '1900-01-01' END as x6, CASE WHEN (x7 != '') THEN x7 ELSE '1900-01-01' END as x7, CASE WHEN (x8 != '') THEN x8 ELSE '1900-01-01' END as x8, CASE WHEN (x9 != '') THEN x9 ELSE '1900-01-01' END as x9, errorMargin as erm, breakoutprice as pE, breakoutbars as be, breakout, atbaridentified as atBar, PatternLengthBars AS l, Bandwidth AS bw, QtyTP AS qtp, p.patternname as patternname, dtt.absolutetimezoneoffset as tzOs, dtt.timezone as tz, approachingtimestamp AS apt, approachingregion as apr, predictionpricefrom as ppf, predictionpriceto as ppt, predictiontimefrom as ptf, predictiontimebars as ptb FROM keylevels_results a INNER JOIN downloadersymbolsettings dss on a.symbolid = dss.symbolid INNER JOIN datafeedstimetable dtt ON dss.classname = dtt.classname inner join symbols s on a.symbolid = s.symbolid INNER JOIN hrspatterns p on a.patternid = p.patternid where resultuid = $1 and dtt.dayofweek = 3;
Date: 2026-02-25 15:20:38 Duration: 2ms Database: postgres parameters: $1 = '607739830026782303'
-
/*server.KeyLevelResult*/ SELECT ResultUID AS ruid, s.symbolid AS sid, symbol AS sym, longname, shortname, Exchange AS e, timegranularity AS tg, a.PatternID AS pid, a.direction AS d, a.patternprice as pp, atbaridentified AS pet, CASE WHEN (x9 != '') THEN x9 WHEN (x8 != '') THEN x8 WHEN (x7 != '') THEN x7 WHEN (x6 != '') THEN x6 WHEN (x5 != '') THEN x5 WHEN (x4 != '') THEN x4 WHEN (x3 != '') THEN x3 WHEN (x2 != '') THEN x2 END AS pst, PatternPrice AS patp, x0, x1, x2, CASE WHEN (x3 != '') THEN x3 ELSE '1900-01-01' END as x3, CASE WHEN (x4 != '') THEN x4 ELSE '1900-01-01' END as x4, CASE WHEN (x5 != '') THEN x5 ELSE '1900-01-01' END as x5, CASE WHEN (x6 != '') THEN x6 ELSE '1900-01-01' END as x6, CASE WHEN (x7 != '') THEN x7 ELSE '1900-01-01' END as x7, CASE WHEN (x8 != '') THEN x8 ELSE '1900-01-01' END as x8, CASE WHEN (x9 != '') THEN x9 ELSE '1900-01-01' END as x9, errorMargin as erm, breakoutprice as pE, breakoutbars as be, breakout, atbaridentified as atBar, PatternLengthBars AS l, Bandwidth AS bw, QtyTP AS qtp, p.patternname as patternname, dtt.absolutetimezoneoffset as tzOs, dtt.timezone as tz, approachingtimestamp AS apt, approachingregion as apr, predictionpricefrom as ppf, predictionpriceto as ppt, predictiontimefrom as ptf, predictiontimebars as ptb FROM keylevels_results a INNER JOIN downloadersymbolsettings dss on a.symbolid = dss.symbolid INNER JOIN datafeedstimetable dtt ON dss.classname = dtt.classname inner join symbols s on a.symbolid = s.symbolid INNER JOIN hrspatterns p on a.patternid = p.patternid where resultuid = $1 and dtt.dayofweek = 3;
Date: 2026-02-25 15:37:13 Duration: 2ms Database: postgres parameters: $1 = '607740004575818303'
18 53ms 8 4ms 10ms 6ms WITH pre_symbols AS ( /* find relevant symbols */ ;Times Reported Time consuming bind #18
Day Hour Count Duration Avg duration 15 8 53ms 6ms -
WITH pre_symbols AS ( /* find relevant symbols */ ;
Date: 2026-02-25 15:13:17 Duration: 10ms Database: postgres parameters: $1 = '1018', $2 = 'ICMARKETS-AU-MT5', $3 = 'AAPL.NAS', $4 = 'ABBV.NYSE', $5 = 'AMCR.NYSE', $6 = 'AMP.NYSE', $7 = 'AMZN.NAS', $8 = 'ANZ.ASX', $9 = 'AUDJPY', $10 = 'AUDUSD', $11 = 'AUS200', $12 = 'BABA.NYSE', $13 = 'BIIB.NAS', $14 = 'BXB.ASX', $15 = 'CBA.ASX', $16 = 'CHINA50', $17 = 'CSL.ASX', $18 = 'DE30', $19 = 'ES35', $20 = 'EURCHF', $21 = 'EURGBP', $22 = 'EURUSD', $23 = 'F40', $24 = 'FMG.ASX', $25 = 'GBPJPY', $26 = 'GBPUSD', $27 = 'GOOG.NAS', $28 = 'HK50', $29 = 'IT40', $30 = 'JP225', $31 = 'KO.NYSE', $32 = 'MQG.ASX', $33 = 'MSFT.NAS', $34 = 'NAB.ASX', $35 = 'NFLX.NAS', $36 = 'PYPL.NAS', $37 = 'QBE.ASX', $38 = 'STOXX50', $39 = 'SUN.ASX', $40 = 'TCL.ASX', $41 = 'TLS.ASX', $42 = 'TSLA.NAS', $43 = 'UK100', $44 = 'UNH.NYSE', $45 = 'US2000', $46 = 'US30', $47 = 'US500', $48 = 'USDCAD', $49 = 'USDCHF', $50 = 'USDCNH', $51 = 'USDJPY', $52 = 'USTEC', $53 = 'WBC.ASX', $54 = 'WES.ASX', $55 = 'WOW.ASX', $56 = 'WPL.ASX', $57 = 'XAUEUR', $58 = 'XAUUSD', $59 = 'XBRUSD', $60 = 'XTIUSD', $61 = 'AAPL.NAS', $62 = 'ABBV.NYSE', $63 = 'AMCR.NYSE', $64 = 'AMP.NYSE', $65 = 'AMZN.NAS', $66 = 'ANZ.ASX', $67 = 'AUDJPY', $68 = 'AUDUSD', $69 = 'AUS200', $70 = 'BABA.NYSE', $71 = 'BIIB.NAS', $72 = 'BXB.ASX', $73 = 'CBA.ASX', $74 = 'CHINA50', $75 = 'CSL.ASX', $76 = 'DE30', $77 = 'ES35', $78 = 'EURCHF', $79 = 'EURGBP', $80 = 'EURUSD', $81 = 'F40', $82 = 'FMG.ASX', $83 = 'GBPJPY', $84 = 'GBPUSD', $85 = 'GOOG.NAS', $86 = 'HK50', $87 = 'IT40', $88 = 'JP225', $89 = 'KO.NYSE', $90 = 'MQG.ASX', $91 = 'MSFT.NAS', $92 = 'NAB.ASX', $93 = 'NFLX.NAS', $94 = 'PYPL.NAS', $95 = 'QBE.ASX', $96 = 'STOXX50', $97 = 'SUN.ASX', $98 = 'TCL.ASX', $99 = 'TLS.ASX', $100 = 'TSLA.NAS', $101 = 'UK100', $102 = 'UNH.NYSE', $103 = 'US2000', $104 = 'US30', $105 = 'US500', $106 = 'USDCAD', $107 = 'USDCHF', $108 = 'USDCNH', $109 = 'USDJPY', $110 = 'USTEC', $111 = 'WBC.ASX', $112 = 'WES.ASX', $113 = 'WOW.ASX', $114 = 'WPL.ASX', $115 = 'XAUEUR', $116 = 'XAUUSD', $117 = 'XBRUSD', $118 = 'XTIUSD', $119 = '5'
-
WITH pre_symbols AS ( /* find relevant symbols */ ;
Date: 2026-02-25 15:13:17 Duration: 10ms Database: postgres parameters: $1 = '1018', $2 = 'ICMARKETS-AU-MT5', $3 = 'AAPL.NAS', $4 = 'ABBV.NYSE', $5 = 'AMCR.NYSE', $6 = 'AMP.NYSE', $7 = 'AMZN.NAS', $8 = 'ANZ.ASX', $9 = 'AUDJPY', $10 = 'AUDUSD', $11 = 'AUS200', $12 = 'BABA.NYSE', $13 = 'BIIB.NAS', $14 = 'BXB.ASX', $15 = 'CBA.ASX', $16 = 'CHINA50', $17 = 'CSL.ASX', $18 = 'DE30', $19 = 'ES35', $20 = 'EURCHF', $21 = 'EURGBP', $22 = 'EURUSD', $23 = 'F40', $24 = 'FMG.ASX', $25 = 'GBPJPY', $26 = 'GBPUSD', $27 = 'GOOG.NAS', $28 = 'HK50', $29 = 'IT40', $30 = 'JP225', $31 = 'KO.NYSE', $32 = 'MQG.ASX', $33 = 'MSFT.NAS', $34 = 'NAB.ASX', $35 = 'NFLX.NAS', $36 = 'PYPL.NAS', $37 = 'QBE.ASX', $38 = 'STOXX50', $39 = 'SUN.ASX', $40 = 'TCL.ASX', $41 = 'TLS.ASX', $42 = 'TSLA.NAS', $43 = 'UK100', $44 = 'UNH.NYSE', $45 = 'US2000', $46 = 'US30', $47 = 'US500', $48 = 'USDCAD', $49 = 'USDCHF', $50 = 'USDCNH', $51 = 'USDJPY', $52 = 'USTEC', $53 = 'WBC.ASX', $54 = 'WES.ASX', $55 = 'WOW.ASX', $56 = 'WPL.ASX', $57 = 'XAUEUR', $58 = 'XAUUSD', $59 = 'XBRUSD', $60 = 'XTIUSD', $61 = 'AAPL.NAS', $62 = 'ABBV.NYSE', $63 = 'AMCR.NYSE', $64 = 'AMP.NYSE', $65 = 'AMZN.NAS', $66 = 'ANZ.ASX', $67 = 'AUDJPY', $68 = 'AUDUSD', $69 = 'AUS200', $70 = 'BABA.NYSE', $71 = 'BIIB.NAS', $72 = 'BXB.ASX', $73 = 'CBA.ASX', $74 = 'CHINA50', $75 = 'CSL.ASX', $76 = 'DE30', $77 = 'ES35', $78 = 'EURCHF', $79 = 'EURGBP', $80 = 'EURUSD', $81 = 'F40', $82 = 'FMG.ASX', $83 = 'GBPJPY', $84 = 'GBPUSD', $85 = 'GOOG.NAS', $86 = 'HK50', $87 = 'IT40', $88 = 'JP225', $89 = 'KO.NYSE', $90 = 'MQG.ASX', $91 = 'MSFT.NAS', $92 = 'NAB.ASX', $93 = 'NFLX.NAS', $94 = 'PYPL.NAS', $95 = 'QBE.ASX', $96 = 'STOXX50', $97 = 'SUN.ASX', $98 = 'TCL.ASX', $99 = 'TLS.ASX', $100 = 'TSLA.NAS', $101 = 'UK100', $102 = 'UNH.NYSE', $103 = 'US2000', $104 = 'US30', $105 = 'US500', $106 = 'USDCAD', $107 = 'USDCHF', $108 = 'USDCNH', $109 = 'USDJPY', $110 = 'USTEC', $111 = 'WBC.ASX', $112 = 'WES.ASX', $113 = 'WOW.ASX', $114 = 'WPL.ASX', $115 = 'XAUEUR', $116 = 'XAUUSD', $117 = 'XBRUSD', $118 = 'XTIUSD', $119 = '5'
-
WITH pre_symbols AS ( /* find relevant symbols */ ;
Date: 2026-02-25 15:13:17 Duration: 10ms Database: postgres parameters: $1 = '1018', $2 = 'ICMARKETS-AU-MT5', $3 = 'AAPL.NAS', $4 = 'ABBV.NYSE', $5 = 'AMCR.NYSE', $6 = 'AMP.NYSE', $7 = 'AMZN.NAS', $8 = 'ANZ.ASX', $9 = 'AUDJPY', $10 = 'AUDUSD', $11 = 'AUS200', $12 = 'BABA.NYSE', $13 = 'BIIB.NAS', $14 = 'BXB.ASX', $15 = 'CBA.ASX', $16 = 'CHINA50', $17 = 'CSL.ASX', $18 = 'DE30', $19 = 'ES35', $20 = 'EURCHF', $21 = 'EURGBP', $22 = 'EURUSD', $23 = 'F40', $24 = 'FMG.ASX', $25 = 'GBPJPY', $26 = 'GBPUSD', $27 = 'GOOG.NAS', $28 = 'HK50', $29 = 'IT40', $30 = 'JP225', $31 = 'KO.NYSE', $32 = 'MQG.ASX', $33 = 'MSFT.NAS', $34 = 'NAB.ASX', $35 = 'NFLX.NAS', $36 = 'PYPL.NAS', $37 = 'QBE.ASX', $38 = 'STOXX50', $39 = 'SUN.ASX', $40 = 'TCL.ASX', $41 = 'TLS.ASX', $42 = 'TSLA.NAS', $43 = 'UK100', $44 = 'UNH.NYSE', $45 = 'US2000', $46 = 'US30', $47 = 'US500', $48 = 'USDCAD', $49 = 'USDCHF', $50 = 'USDCNH', $51 = 'USDJPY', $52 = 'USTEC', $53 = 'WBC.ASX', $54 = 'WES.ASX', $55 = 'WOW.ASX', $56 = 'WPL.ASX', $57 = 'XAUEUR', $58 = 'XAUUSD', $59 = 'XBRUSD', $60 = 'XTIUSD', $61 = 'AAPL.NAS', $62 = 'ABBV.NYSE', $63 = 'AMCR.NYSE', $64 = 'AMP.NYSE', $65 = 'AMZN.NAS', $66 = 'ANZ.ASX', $67 = 'AUDJPY', $68 = 'AUDUSD', $69 = 'AUS200', $70 = 'BABA.NYSE', $71 = 'BIIB.NAS', $72 = 'BXB.ASX', $73 = 'CBA.ASX', $74 = 'CHINA50', $75 = 'CSL.ASX', $76 = 'DE30', $77 = 'ES35', $78 = 'EURCHF', $79 = 'EURGBP', $80 = 'EURUSD', $81 = 'F40', $82 = 'FMG.ASX', $83 = 'GBPJPY', $84 = 'GBPUSD', $85 = 'GOOG.NAS', $86 = 'HK50', $87 = 'IT40', $88 = 'JP225', $89 = 'KO.NYSE', $90 = 'MQG.ASX', $91 = 'MSFT.NAS', $92 = 'NAB.ASX', $93 = 'NFLX.NAS', $94 = 'PYPL.NAS', $95 = 'QBE.ASX', $96 = 'STOXX50', $97 = 'SUN.ASX', $98 = 'TCL.ASX', $99 = 'TLS.ASX', $100 = 'TSLA.NAS', $101 = 'UK100', $102 = 'UNH.NYSE', $103 = 'US2000', $104 = 'US30', $105 = 'US500', $106 = 'USDCAD', $107 = 'USDCHF', $108 = 'USDCNH', $109 = 'USDJPY', $110 = 'USTEC', $111 = 'WBC.ASX', $112 = 'WES.ASX', $113 = 'WOW.ASX', $114 = 'WPL.ASX', $115 = 'XAUEUR', $116 = 'XAUUSD', $117 = 'XBRUSD', $118 = 'XTIUSD', $119 = '5'
19 50ms 1 50ms 50ms 50ms with maxwhid as ( ;Times Reported Time consuming bind #19
Day Hour Count Duration Avg duration 15 1 50ms 50ms -
with maxwhid as ( ;
Date: 2026-02-25 15:13:24 Duration: 50ms Database: postgres parameters: $1 = '335', $2 = '621', $3 = '637', $4 = '642', $5 = '660', $6 = '666', $7 = '643', $8 = '630', $9 = '680', $10 = '641', $11 = '431', $12 = '622', $13 = '489', $14 = '529', $15 = '576', $16 = '665', $17 = '667', $18 = '558', $19 = '620', $20 = '125', $21 = '488', $22 = '567', $23 = '689', $24 = '700', $25 = '758', $26 = '763', $27 = '765', $28 = '817', $29 = '914', $30 = '972'
20 37ms 55 0ms 4ms 0ms /*server.FibonacciResult*/ SELECT ResultUID AS ruid, s.symbolid AS sid, symbol AS sym, Exchange AS e, longname as lo, shortname as sho, timegranularity AS tg, p.PatternID AS pid, Direction AS d, PatternStartTime AS pst, PatternEndTime AS pet, PatternStartPrice AS psp, PatternEndPrice AS pep, priceX as px, timeX as tx, priceA as pa, timeA as ta, priceB as pb, timeB as tb, priceC as pc, timeC as tc, priceD as pd, timeD as td, averagequality as aq, timequality as tq, 1 - errormargin as rq, 1 - noise as c, target10 as t10, target06 as t06, target16 as t16, target07 as t07, target12 as t12, target03 as t03, target05 as t05, PatternLengthBars AS l, temporarypattern as tp, Bandwidth AS bw, QtyTP AS qtp, p.patternname as patternname, dtt.absolutetimezoneoffset as tzOs, dtt.timezone as tz FROM Fibonacci_Results a INNER JOIN downloadersymbolsettings dss on a.symbolid = dss.symbolid INNER JOIN datafeedstimetable dtt ON dss.classname = dtt.classname inner join symbols s on a.symbolid = s.symbolid INNER JOIN fibonaccipatterns p on a.pattern = p.patternname where resultuid = $1 and dtt.dayofweek = 3;Times Reported Time consuming bind #20
Day Hour Count Duration Avg duration 15 55 37ms 0ms -
/*server.FibonacciResult*/ SELECT ResultUID AS ruid, s.symbolid AS sid, symbol AS sym, Exchange AS e, longname as lo, shortname as sho, timegranularity AS tg, p.PatternID AS pid, Direction AS d, PatternStartTime AS pst, PatternEndTime AS pet, PatternStartPrice AS psp, PatternEndPrice AS pep, priceX as px, timeX as tx, priceA as pa, timeA as ta, priceB as pb, timeB as tb, priceC as pc, timeC as tc, priceD as pd, timeD as td, averagequality as aq, timequality as tq, 1 - errormargin as rq, 1 - noise as c, target10 as t10, target06 as t06, target16 as t16, target07 as t07, target12 as t12, target03 as t03, target05 as t05, PatternLengthBars AS l, temporarypattern as tp, Bandwidth AS bw, QtyTP AS qtp, p.patternname as patternname, dtt.absolutetimezoneoffset as tzOs, dtt.timezone as tz FROM Fibonacci_Results a INNER JOIN downloadersymbolsettings dss on a.symbolid = dss.symbolid INNER JOIN datafeedstimetable dtt ON dss.classname = dtt.classname inner join symbols s on a.symbolid = s.symbolid INNER JOIN fibonaccipatterns p on a.pattern = p.patternname where resultuid = $1 and dtt.dayofweek = 3;
Date: 2026-02-25 15:06:10 Duration: 4ms Database: postgres parameters: $1 = '607739948036594302'
-
/*server.FibonacciResult*/ SELECT ResultUID AS ruid, s.symbolid AS sid, symbol AS sym, Exchange AS e, longname as lo, shortname as sho, timegranularity AS tg, p.PatternID AS pid, Direction AS d, PatternStartTime AS pst, PatternEndTime AS pet, PatternStartPrice AS psp, PatternEndPrice AS pep, priceX as px, timeX as tx, priceA as pa, timeA as ta, priceB as pb, timeB as tb, priceC as pc, timeC as tc, priceD as pd, timeD as td, averagequality as aq, timequality as tq, 1 - errormargin as rq, 1 - noise as c, target10 as t10, target06 as t06, target16 as t16, target07 as t07, target12 as t12, target03 as t03, target05 as t05, PatternLengthBars AS l, temporarypattern as tp, Bandwidth AS bw, QtyTP AS qtp, p.patternname as patternname, dtt.absolutetimezoneoffset as tzOs, dtt.timezone as tz FROM Fibonacci_Results a INNER JOIN downloadersymbolsettings dss on a.symbolid = dss.symbolid INNER JOIN datafeedstimetable dtt ON dss.classname = dtt.classname inner join symbols s on a.symbolid = s.symbolid INNER JOIN fibonaccipatterns p on a.pattern = p.patternname where resultuid = $1 and dtt.dayofweek = 3;
Date: 2026-02-25 15:35:22 Duration: 4ms Database: postgres parameters: $1 = '607740186072050302'
-
/*server.FibonacciResult*/ SELECT ResultUID AS ruid, s.symbolid AS sid, symbol AS sym, Exchange AS e, longname as lo, shortname as sho, timegranularity AS tg, p.PatternID AS pid, Direction AS d, PatternStartTime AS pst, PatternEndTime AS pet, PatternStartPrice AS psp, PatternEndPrice AS pep, priceX as px, timeX as tx, priceA as pa, timeA as ta, priceB as pb, timeB as tb, priceC as pc, timeC as tc, priceD as pd, timeD as td, averagequality as aq, timequality as tq, 1 - errormargin as rq, 1 - noise as c, target10 as t10, target06 as t06, target16 as t16, target07 as t07, target12 as t12, target03 as t03, target05 as t05, PatternLengthBars AS l, temporarypattern as tp, Bandwidth AS bw, QtyTP AS qtp, p.patternname as patternname, dtt.absolutetimezoneoffset as tzOs, dtt.timezone as tz FROM Fibonacci_Results a INNER JOIN downloadersymbolsettings dss on a.symbolid = dss.symbolid INNER JOIN datafeedstimetable dtt ON dss.classname = dtt.classname inner join symbols s on a.symbolid = s.symbolid INNER JOIN fibonaccipatterns p on a.pattern = p.patternname where resultuid = $1 and dtt.dayofweek = 3;
Date: 2026-02-25 15:42:17 Duration: 2ms Database: postgres parameters: $1 = '607739948092080302'
-
Events
Log levels
Key values
- 430,165 Log entries
Events distribution
Key values
- 0 PANIC entries
- 0 FATAL entries
- 406 ERROR entries
- 0 WARNING entries
Most Frequent Errors/Events
Key values
- 359 Max number of times the same event was reported
- 406 Total events found
Rank Times reported Error 1 359 ERROR: pg_stat_statements must be loaded via shared_preload_libraries
Times Reported Most Frequent Error / Event #1
Day Hour Count Feb 25 15 359 - ERROR: pg_stat_statements must be loaded via shared_preload_libraries
Statement: /* service='datadog-agent' */ SELECT COUNT(*) FROM pg_stat_statements(false)
Date: 2026-02-25 15:00:08
2 47 ERROR: schema "..." does not exist
Times Reported Most Frequent Error / Event #2
Day Hour Count Feb 25 15 47 - ERROR: schema "datadog" does not exist at character 38
Statement: /* service='datadog-agent' */ SELECT datadog.explain_statement($stmt$SELECT * FROM pg_stat_activity$stmt$)
Date: 2026-02-25 15:02:40