Dataset Viewer
Auto-converted to Parquet
Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
4
112
repo_url
stringlengths
33
141
action
stringclasses
3 values
title
stringlengths
1
970
labels
stringlengths
4
625
body
stringlengths
3
247k
index
stringclasses
9 values
text_combine
stringlengths
96
247k
label
stringclasses
2 values
text
stringlengths
96
218k
binary_label
int64
0
1
28,623
13,764,327,917
IssuesEvent
2020-10-07 11:54:57
ppy/osu
https://api.github.com/repos/ppy/osu
closed
osu! freezes within a beatmap
type:performance
Hi there, I just tested out osu lazer for a bit and came across some freezes in which the whole screen freezes for a second or more in the normal osu mode. I don't know what's causing the issue, maybe something with the database (?) as I had a few errors importing my beatmaps. Windows 10 (1709, x64) Intel Core i7 6800k Readon RX Vega 64 (driver: 18.3.4) osu version: 2018.423.0 using a guest account I hope those [logs](https://github.com/ppy/osu/files/1949392/logs.zip) help to fix the bug and get one step closer to a stable osu lazer release. ;) (I just put in all logs from the whole session... Performance logging features got enabled after I first noticed the problem)
True
osu! freezes within a beatmap - Hi there, I just tested out osu lazer for a bit and came across some freezes in which the whole screen freezes for a second or more in the normal osu mode. I don't know what's causing the issue, maybe something with the database (?) as I had a few errors importing my beatmaps. Windows 10 (1709, x64) Intel Core i7 6800k Readon RX Vega 64 (driver: 18.3.4) osu version: 2018.423.0 using a guest account I hope those [logs](https://github.com/ppy/osu/files/1949392/logs.zip) help to fix the bug and get one step closer to a stable osu lazer release. ;) (I just put in all logs from the whole session... Performance logging features got enabled after I first noticed the problem)
perf
osu freezes within a beatmap hi there i just tested out osu lazer for a bit and came across some freezes in which the whole screen freezes for a second or more in the normal osu mode i don t know what s causing the issue maybe something with the database as i had a few errors importing my beatmaps windows intel core readon rx vega driver osu version using a guest account i hope those help to fix the bug and get one step closer to a stable osu lazer release i just put in all logs from the whole session performance logging features got enabled after i first noticed the problem
1
506,648
14,669,807,730
IssuesEvent
2020-12-30 02:17:03
GrapheneOS/os_issue_tracker
https://api.github.com/repos/GrapheneOS/os_issue_tracker
closed
roll back changes to vndk ABI caused by churn from compiler hardening
enhancement low-priority upstream
This seems to have been triggered by enabling -ftrivial-auto-var-init=zero. It doesn't seem like anything is actually wrong but rather a lot of non-public things leak into the C++ ABI potentially due to them not using -fvisibility=hidden.
1.0
roll back changes to vndk ABI caused by churn from compiler hardening - This seems to have been triggered by enabling -ftrivial-auto-var-init=zero. It doesn't seem like anything is actually wrong but rather a lot of non-public things leak into the C++ ABI potentially due to them not using -fvisibility=hidden.
non_perf
roll back changes to vndk abi caused by churn from compiler hardening this seems to have been triggered by enabling ftrivial auto var init zero it doesn t seem like anything is actually wrong but rather a lot of non public things leak into the c abi potentially due to them not using fvisibility hidden
0
18,258
10,053,584,003
IssuesEvent
2019-07-21 18:00:21
raoulvdberge/refinedstorage
https://api.github.com/repos/raoulvdberge/refinedstorage
closed
Server crash-loop
Performance
#### Issue description: Server goes into crash-loop #### What happens: On boot, server hangs for a bit, then the watchdog service reboots the server for ticks taking longer than 60 seconds. #### What you expected to happen: No crashing #### Steps to reproduce: 1.boot server 2.wait a minute 3. server reboots ... #### Version (make sure you are on the latest version before reporting): - Minecraft: 1.12.2 - Forge: 14.23.5.2775 (and also 14.23.5.2768 tested both) - Refined Storage: 1.6.9 Does this issue occur on a server? [yes/no] yes #### If a (crash)log is relevant for this issue, link it here: https://pastebin.com/gi4KF9iC
True
Server crash-loop - #### Issue description: Server goes into crash-loop #### What happens: On boot, server hangs for a bit, then the watchdog service reboots the server for ticks taking longer than 60 seconds. #### What you expected to happen: No crashing #### Steps to reproduce: 1.boot server 2.wait a minute 3. server reboots ... #### Version (make sure you are on the latest version before reporting): - Minecraft: 1.12.2 - Forge: 14.23.5.2775 (and also 14.23.5.2768 tested both) - Refined Storage: 1.6.9 Does this issue occur on a server? [yes/no] yes #### If a (crash)log is relevant for this issue, link it here: https://pastebin.com/gi4KF9iC
perf
server crash loop issue description server goes into crash loop what happens on boot server hangs for a bit then the watchdog service reboots the server for ticks taking longer than seconds what you expected to happen no crashing steps to reproduce boot server wait a minute server reboots version make sure you are on the latest version before reporting minecraft forge and also tested both refined storage does this issue occur on a server yes if a crash log is relevant for this issue link it here
1
9,114
6,767,111,023
IssuesEvent
2017-10-26 01:10:36
ianstormtaylor/slate
https://api.github.com/repos/ianstormtaylor/slate
closed
optimize `state.toJSON` performance by being lazier
improvement ⚑ performance
Right now when we do `state.toJSON()`, we serialize all the potential properties of the state, and then delete the ones that shouldn't be included. This is obviously the slower way, we should add in properties if the options call for it instead.
True
optimize `state.toJSON` performance by being lazier - Right now when we do `state.toJSON()`, we serialize all the potential properties of the state, and then delete the ones that shouldn't be included. This is obviously the slower way, we should add in properties if the options call for it instead.
perf
optimize state tojson performance by being lazier right now when we do state tojson we serialize all the potential properties of the state and then delete the ones that shouldn t be included this is obviously the slower way we should add in properties if the options call for it instead
1
23,453
11,966,281,074
IssuesEvent
2020-04-06 02:52:11
Quarantine-Help/quarantine-hybrid-app
https://api.github.com/repos/Quarantine-Help/quarantine-hybrid-app
opened
Modularize the app and implement SelectivePreloadingStrategy
performance refactor
Group the page into modules and consolidate routing. The proposed hierarchy needs to be updated with the latest workflow changes before being implemented. ### Proposed Module Hierarchy **OnboardModule** * Landing pages * Registration pages for Volunteer & Quarantined **MainModule** * Maps page * Request creation pages * Request handling pages **MiscModule** * My Profile pages - edit/save ## Additional Context Go through the workflow and discuss on slack https://xd.adobe.com/spec/6a2c4d00-a356-4885-5e8d-a4379323a760-1c1d/grid/ **Refer** https://ionicframework.com/docs/angular/navigation#lazy-loading-routes https://ionicframework.com/blog/how-to-lazy-load-in-ionic-angular/
True
Modularize the app and implement SelectivePreloadingStrategy - Group the page into modules and consolidate routing. The proposed hierarchy needs to be updated with the latest workflow changes before being implemented. ### Proposed Module Hierarchy **OnboardModule** * Landing pages * Registration pages for Volunteer & Quarantined **MainModule** * Maps page * Request creation pages * Request handling pages **MiscModule** * My Profile pages - edit/save ## Additional Context Go through the workflow and discuss on slack https://xd.adobe.com/spec/6a2c4d00-a356-4885-5e8d-a4379323a760-1c1d/grid/ **Refer** https://ionicframework.com/docs/angular/navigation#lazy-loading-routes https://ionicframework.com/blog/how-to-lazy-load-in-ionic-angular/
perf
modularize the app and implement selectivepreloadingstrategy group the page into modules and consolidate routing the proposed hierarchy needs to be updated with the latest workflow changes before being implemented proposed module hierarchy onboardmodule landing pages registration pages for volunteer quarantined mainmodule maps page request creation pages request handling pages miscmodule my profile pages edit save additional context go through the workflow and discuss on slack refer
1
56,357
31,884,564,610
IssuesEvent
2023-09-16 19:48:15
dotnet/runtime
https://api.github.com/repos/dotnet/runtime
opened
Very slow in those type of calcs
tenet-performance
Net (all versions, 6, 7 and coming 8 are all very slow compared to other languages) I do not expect to be as fast as e.g. rust, but being 10x slower than slow python is a bit of a shame: https://programming-language-benchmarks.vercel.app/problem/edigits I'm not sure if this will work there, but this is a suggestion for possible optimisations. Hope that version 9 will do something with that.
True
Very slow in those type of calcs - Net (all versions, 6, 7 and coming 8 are all very slow compared to other languages) I do not expect to be as fast as e.g. rust, but being 10x slower than slow python is a bit of a shame: https://programming-language-benchmarks.vercel.app/problem/edigits I'm not sure if this will work there, but this is a suggestion for possible optimisations. Hope that version 9 will do something with that.
perf
very slow in those type of calcs net all versions and coming are all very slow compared to other languages i do not expect to be as fast as e g rust but being slower than slow python is a bit of a shame i m not sure if this will work there but this is a suggestion for possible optimisations hope that version will do something with that
1
26,568
13,055,871,729
IssuesEvent
2020-07-30 02:59:07
fused-effects/fused-effects
https://api.github.com/repos/fused-effects/fused-effects
closed
Manually-fused RWST carrier is significantly faster than StateC+ReaderC+WriterC
bug performance
Given [this RWSC carrier](https://gist.github.com/patrickt/0ea924b742bf675b3b1d47cf4091d720) and the following computation: ```haskell go :: ( Member (State Int) sig , Member (Writer String) sig , Member (Reader Bool) sig , Carrier sig m ) => Int -> m () go n = forM_ (take n (cycle names)) $ \str -> do let len = length str modify @Int (+ len) curr <- get @Int when (curr < n) $ do should <- ask let str' = if should then reverse str else str tell str' ``` interpreting it with RWSC is around twice as fast as StateC+ReaderC+WriterC: ``` benchmarked fused-effects/Separate carriers/10 time 480.1 ns (461.4 ns .. 501.3 ns) 0.980 R² (0.967 R² .. 0.990 R²) mean 515.3 ns (502.4 ns .. 537.7 ns) std dev 53.87 ns (39.49 ns .. 80.78 ns) variance introduced by outliers: 63% (severely inflated) benchmarked fused-effects/Separate carriers/100 time 4.781 μs (4.712 μs .. 4.879 μs) 0.995 R² (0.988 R² .. 0.999 R²) mean 4.767 μs (4.720 μs .. 4.844 μs) std dev 199.3 ns (136.2 ns .. 321.1 ns) variance introduced by outliers: 23% (moderately inflated) benchmarked fused-effects/Separate carriers/1000 time 50.35 μs (48.92 μs .. 51.66 μs) 0.992 R² (0.985 R² .. 0.997 R²) mean 49.53 μs (48.85 μs .. 50.44 μs) std dev 2.588 μs (1.973 μs .. 3.720 μs) variance introduced by outliers: 30% (moderately inflated) benchmarked fused-effects/RWST carrier/10 time 336.0 ns (317.5 ns .. 361.3 ns) 0.971 R² (0.951 R² .. 0.994 R²) mean 322.9 ns (317.0 ns .. 332.4 ns) std dev 25.15 ns (16.85 ns .. 36.78 ns) variance introduced by outliers: 50% (moderately inflated) benchmarked fused-effects/RWST carrier/100 time 2.664 μs (2.586 μs .. 2.723 μs) 0.993 R² (0.982 R² .. 0.999 R²) mean 2.928 μs (2.858 μs .. 3.105 μs) std dev 379.1 ns (137.0 ns .. 685.6 ns) variance introduced by outliers: 74% (severely inflated) benchmarked fused-effects/RWST carrier/1000 time 28.39 μs (28.01 μs .. 28.94 μs) 0.997 R² (0.995 R² .. 0.999 R²) mean 28.49 μs (28.30 μs .. 28.75 μs) std dev 753.7 ns (584.7 ns .. 1.040 μs) variance introduced by outliers: 11% (moderately inflated) ``` I would expect to see a small speedup using RWSC (`mtl` versions of similar code are around 10% faster with RWS). This is much too large and makes me think that we might not be getting enough fusion.
True
Manually-fused RWST carrier is significantly faster than StateC+ReaderC+WriterC - Given [this RWSC carrier](https://gist.github.com/patrickt/0ea924b742bf675b3b1d47cf4091d720) and the following computation: ```haskell go :: ( Member (State Int) sig , Member (Writer String) sig , Member (Reader Bool) sig , Carrier sig m ) => Int -> m () go n = forM_ (take n (cycle names)) $ \str -> do let len = length str modify @Int (+ len) curr <- get @Int when (curr < n) $ do should <- ask let str' = if should then reverse str else str tell str' ``` interpreting it with RWSC is around twice as fast as StateC+ReaderC+WriterC: ``` benchmarked fused-effects/Separate carriers/10 time 480.1 ns (461.4 ns .. 501.3 ns) 0.980 R² (0.967 R² .. 0.990 R²) mean 515.3 ns (502.4 ns .. 537.7 ns) std dev 53.87 ns (39.49 ns .. 80.78 ns) variance introduced by outliers: 63% (severely inflated) benchmarked fused-effects/Separate carriers/100 time 4.781 μs (4.712 μs .. 4.879 μs) 0.995 R² (0.988 R² .. 0.999 R²) mean 4.767 μs (4.720 μs .. 4.844 μs) std dev 199.3 ns (136.2 ns .. 321.1 ns) variance introduced by outliers: 23% (moderately inflated) benchmarked fused-effects/Separate carriers/1000 time 50.35 μs (48.92 μs .. 51.66 μs) 0.992 R² (0.985 R² .. 0.997 R²) mean 49.53 μs (48.85 μs .. 50.44 μs) std dev 2.588 μs (1.973 μs .. 3.720 μs) variance introduced by outliers: 30% (moderately inflated) benchmarked fused-effects/RWST carrier/10 time 336.0 ns (317.5 ns .. 361.3 ns) 0.971 R² (0.951 R² .. 0.994 R²) mean 322.9 ns (317.0 ns .. 332.4 ns) std dev 25.15 ns (16.85 ns .. 36.78 ns) variance introduced by outliers: 50% (moderately inflated) benchmarked fused-effects/RWST carrier/100 time 2.664 μs (2.586 μs .. 2.723 μs) 0.993 R² (0.982 R² .. 0.999 R²) mean 2.928 μs (2.858 μs .. 3.105 μs) std dev 379.1 ns (137.0 ns .. 685.6 ns) variance introduced by outliers: 74% (severely inflated) benchmarked fused-effects/RWST carrier/1000 time 28.39 μs (28.01 μs .. 28.94 μs) 0.997 R² (0.995 R² .. 0.999 R²) mean 28.49 μs (28.30 μs .. 28.75 μs) std dev 753.7 ns (584.7 ns .. 1.040 μs) variance introduced by outliers: 11% (moderately inflated) ``` I would expect to see a small speedup using RWSC (`mtl` versions of similar code are around 10% faster with RWS). This is much too large and makes me think that we might not be getting enough fusion.
perf
manually fused rwst carrier is significantly faster than statec readerc writerc given and the following computation haskell go member state int sig member writer string sig member reader bool sig carrier sig m int m go n form take n cycle names str do let len length str modify int len curr get int when curr n do should ask let str if should then reverse str else str tell str interpreting it with rwsc is around twice as fast as statec readerc writerc benchmarked fused effects separate carriers time ns ns ns r² r² r² mean ns ns ns std dev ns ns ns variance introduced by outliers severely inflated benchmarked fused effects separate carriers time μs μs μs r² r² r² mean μs μs μs std dev ns ns ns variance introduced by outliers moderately inflated benchmarked fused effects separate carriers time μs μs μs r² r² r² mean μs μs μs std dev μs μs μs variance introduced by outliers moderately inflated benchmarked fused effects rwst carrier time ns ns ns r² r² r² mean ns ns ns std dev ns ns ns variance introduced by outliers moderately inflated benchmarked fused effects rwst carrier time μs μs μs r² r² r² mean μs μs μs std dev ns ns ns variance introduced by outliers severely inflated benchmarked fused effects rwst carrier time μs μs μs r² r² r² mean μs μs μs std dev ns ns μs variance introduced by outliers moderately inflated i would expect to see a small speedup using rwsc mtl versions of similar code are around faster with rws this is much too large and makes me think that we might not be getting enough fusion
1
285,518
8,761,701,155
IssuesEvent
2018-12-16 20:05:49
FSPNet/Orion
https://api.github.com/repos/FSPNet/Orion
opened
HTTP Method shouldn't be GET
🐛 Bug 🔖 Version/1.0 🚨 Priority/P0
**Describe the bug** HTTP Method shouldn't be GET, it should be POST, **Expected behavior** Change some GET routes to POST. like 'warband', 'factorio', **Environment (please complete the following information):** - Orion version(s): 1.0.2
1.0
HTTP Method shouldn't be GET - **Describe the bug** HTTP Method shouldn't be GET, it should be POST, **Expected behavior** Change some GET routes to POST. like 'warband', 'factorio', **Environment (please complete the following information):** - Orion version(s): 1.0.2
non_perf
http method shouldn t be get describe the bug http method shouldn t be get it should be post expected behavior change some get routes to post like warband factorio environment please complete the following information orion version s
0
197,997
14,953,083,623
IssuesEvent
2021-01-26 16:16:22
pints-team/pints
https://api.github.com/repos/pints-team/pints
closed
Add value-based (numerical) tests for all samplers / optimisers
unit-testing
E.g. - Seed - Run 100 iterations - Check that there's sufficient change within those iterations (and reduce n if possible) - Store output, either in CSV or in code - Compare This would be _in addition to_ functional testing, and would be slightly annoying because you'd need to update the stored results any time you made changes. But probably still good to have to check the impact of e.g. refactoring Thoughts @ben18785 @fcooper8472 @martinjrobins @DavAug @rcw5890 ?
1.0
Add value-based (numerical) tests for all samplers / optimisers - E.g. - Seed - Run 100 iterations - Check that there's sufficient change within those iterations (and reduce n if possible) - Store output, either in CSV or in code - Compare This would be _in addition to_ functional testing, and would be slightly annoying because you'd need to update the stored results any time you made changes. But probably still good to have to check the impact of e.g. refactoring Thoughts @ben18785 @fcooper8472 @martinjrobins @DavAug @rcw5890 ?
non_perf
add value based numerical tests for all samplers optimisers e g seed run iterations check that there s sufficient change within those iterations and reduce n if possible store output either in csv or in code compare this would be in addition to functional testing and would be slightly annoying because you d need to update the stored results any time you made changes but probably still good to have to check the impact of e g refactoring thoughts martinjrobins davaug
0
37,990
18,871,891,643
IssuesEvent
2021-11-13 10:23:46
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
reopened
sql: don't touch ranges unnecessarily during limited scans
C-performance A-sql-execution T-sql-queries
A user saw a case where selecting all rows from a (small) partitioned table was significantly faster than selecting one row using LIMIT 1. The reason was that the first region was also the farthest away, and it only had one row. There is a single TableReader planned in that region (because of the limit); the kv fetcher requests two keys in this case, and the scan ends up going to a range on another region. In general, we need to fetch one more key so that we're sure we got all keys for a row (if there are multiple column families). But we could get the same signal if we knew that we hit the end of a range. ~Having a KV API that allows stopping a scan at the end of the range would be useful here. CC @tbg @andreimatei who have been thinking about the APIs between KV and SQL.~ EDIT (@erikgrinaker): The KV API is available as of #70763.
True
sql: don't touch ranges unnecessarily during limited scans - A user saw a case where selecting all rows from a (small) partitioned table was significantly faster than selecting one row using LIMIT 1. The reason was that the first region was also the farthest away, and it only had one row. There is a single TableReader planned in that region (because of the limit); the kv fetcher requests two keys in this case, and the scan ends up going to a range on another region. In general, we need to fetch one more key so that we're sure we got all keys for a row (if there are multiple column families). But we could get the same signal if we knew that we hit the end of a range. ~Having a KV API that allows stopping a scan at the end of the range would be useful here. CC @tbg @andreimatei who have been thinking about the APIs between KV and SQL.~ EDIT (@erikgrinaker): The KV API is available as of #70763.
perf
sql don t touch ranges unnecessarily during limited scans a user saw a case where selecting all rows from a small partitioned table was significantly faster than selecting one row using limit the reason was that the first region was also the farthest away and it only had one row there is a single tablereader planned in that region because of the limit the kv fetcher requests two keys in this case and the scan ends up going to a range on another region in general we need to fetch one more key so that we re sure we got all keys for a row if there are multiple column families but we could get the same signal if we knew that we hit the end of a range having a kv api that allows stopping a scan at the end of the range would be useful here cc tbg andreimatei who have been thinking about the apis between kv and sql edit erikgrinaker the kv api is available as of
1
23,460
11,887,072,937
IssuesEvent
2020-03-28 00:02:55
microsoft/vscode-cpptools
https://api.github.com/repos/microsoft/vscode-cpptools
closed
Compiler path with spaces produces error
Feature: Configuration Language Service bug quick fix regression
Related topic https://community.platformio.org/t/platform-io-compiler-error/12684 1) We provide a full path to the compiler using `compilerPath` option 2) This a path option, so we do not do any modifications because arguments are passed to `compilerArgs` Yes, we can escape `compilerPath` by default but it looks like a bug. Thanks! /cc @valeros @sean-mcmanus
1.0
Compiler path with spaces produces error - Related topic https://community.platformio.org/t/platform-io-compiler-error/12684 1) We provide a full path to the compiler using `compilerPath` option 2) This a path option, so we do not do any modifications because arguments are passed to `compilerArgs` Yes, we can escape `compilerPath` by default but it looks like a bug. Thanks! /cc @valeros @sean-mcmanus
non_perf
compiler path with spaces produces error related topic we provide a full path to the compiler using compilerpath option this a path option so we do not do any modifications because arguments are passed to compilerargs yes we can escape compilerpath by default but it looks like a bug thanks cc valeros sean mcmanus
0
12,872
8,029,252,919
IssuesEvent
2018-07-27 15:25:07
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
opened
storage: Raft not committing new entries incrementally
A-core-replication C-performance
A cluster that got into a weird state has revealed something that doesn't appear to be working correctly in raft. The cluster was running 2.0.4 with the patch in https://github.com/cockroachdb/cockroach/issues/27804#issuecomment-406635478 A range had gotten into a state with only two replicas, and one of them had been down for a long time. The live node was left as leader and continually added to its raft log. After the downed follower came back online, that multi-GB raft log had to be copied to the follower before new progress could be made (including establishing a new lease for the range or adding a third replica). This process went extremely slowly, taking 30 hours in one case to transfer a few GB. While this was happening, the range status page showed that the follower's Last Index was increasing steadily, but the Commit and Applied indexes remained constant until the follower caught up completely. This is unexpected; the Commit and Applied indexes should increase as the follower progresses, tracking only a few message round trips behind the Last index. Note that raft leadership was stable during this time (term number was only 141), so it was not the case that elections were being called frequently (which could slow down the leader and prevent it from committing entries as it processes MsgAppResps).
True
storage: Raft not committing new entries incrementally - A cluster that got into a weird state has revealed something that doesn't appear to be working correctly in raft. The cluster was running 2.0.4 with the patch in https://github.com/cockroachdb/cockroach/issues/27804#issuecomment-406635478 A range had gotten into a state with only two replicas, and one of them had been down for a long time. The live node was left as leader and continually added to its raft log. After the downed follower came back online, that multi-GB raft log had to be copied to the follower before new progress could be made (including establishing a new lease for the range or adding a third replica). This process went extremely slowly, taking 30 hours in one case to transfer a few GB. While this was happening, the range status page showed that the follower's Last Index was increasing steadily, but the Commit and Applied indexes remained constant until the follower caught up completely. This is unexpected; the Commit and Applied indexes should increase as the follower progresses, tracking only a few message round trips behind the Last index. Note that raft leadership was stable during this time (term number was only 141), so it was not the case that elections were being called frequently (which could slow down the leader and prevent it from committing entries as it processes MsgAppResps).
perf
storage raft not committing new entries incrementally a cluster that got into a weird state has revealed something that doesn t appear to be working correctly in raft the cluster was running with the patch in a range had gotten into a state with only two replicas and one of them had been down for a long time the live node was left as leader and continually added to its raft log after the downed follower came back online that multi gb raft log had to be copied to the follower before new progress could be made including establishing a new lease for the range or adding a third replica this process went extremely slowly taking hours in one case to transfer a few gb while this was happening the range status page showed that the follower s last index was increasing steadily but the commit and applied indexes remained constant until the follower caught up completely this is unexpected the commit and applied indexes should increase as the follower progresses tracking only a few message round trips behind the last index note that raft leadership was stable during this time term number was only so it was not the case that elections were being called frequently which could slow down the leader and prevent it from committing entries as it processes msgappresps
1
176,051
13,625,085,484
IssuesEvent
2020-09-24 09:00:10
dart-lang/sdk
https://api.github.com/repos/dart-lang/sdk
closed
[Test] Test runner doesn't seem to honor weak/strong flag
area-test
See detailed transcript below, but in summary: Running test.py on a test with `--nnbd=weak` passed as an argument results in this command line: ``` DART_CONFIGURATION=ReleaseX64 xcodebuild/ReleaseX64/dart --enable-experiment=non-nullable --ignore-unrecognized-flags --packages=/Users/leafp/src/dart-repo/sdk/.packages /Users/leafp/src/dart-repo/sdk/tests/language/nnbd/normalization/generic_function_type_object_normalization_test.dart ``` Note that this command line does not set the mode to weak mode. Running with `--nnbd=strong` results in the same command line. cc @munificent @sortie ``` leafp-macbookpro:sdk leafp$ python tools/test.py -c dartk -m release --enable-experiment=non-nullable --nnbd=weak tests/language/nnbd/normalization/generic_function_type_object_normalization_test.dart Test configuration: custom configuration(architecture: x64, compiler: dartk, mode: release, runtime: vm, system: mac, nnbd: weak, enable-experiment: [non-nullable]) Suites tested: language FAILED: dartk-vm release_x64 language/nnbd/normalization/generic_function_type_object_normalization_test Expected: Pass Actual: RuntimeError --- Command "vm" (took 602ms): DART_CONFIGURATION=ReleaseX64 xcodebuild/ReleaseX64/dart --enable-experiment=non-nullable --ignore-unrecognized-flags --packages=/Users/leafp/src/dart-repo/sdk/.packages /Users/leafp/src/dart-repo/sdk/tests/language/nnbd/normalization/generic_function_type_object_normalization_test.dart exit code: 255 stderr: Unhandled exception: Expect.notEquals(unexpected: <<R0 extends Future<Never>, R1 extends Never, R2 extends Null>() => Null>, actual:<<R0 extends Future<Never>, R1 extends Never, R2 extends Null>() => Null>) fails. #0 Expect._fail (package:expect/expect.dart:685:5) #1 Expect.notEquals (package:expect/expect.dart:306:5) #2 checkNotEquals2 (file:///Users/leafp/src/dart-repo/sdk/tests/language/nnbd/normalization/type_builder.dart:136:10) #3 checkTypeNotEquals2 (file:///Users/leafp/src/dart-repo/sdk/tests/language/nnbd/normalization/generic_function_type_object_normalization_test.dart:22:3) #4 neverBoundTests (file:///Users/leafp/src/dart-repo/sdk/tests/language/nnbd/normalization/generic_function_type_object_normalization_test.dart:175:5) #5 main (file:///Users/leafp/src/dart-repo/sdk/tests/language/nnbd/normalization/generic_function_type_object_normalization_test.dart:199:3) #6 _startIsolate.<anonymous closure> (dart:isolate-patch/isolate_patch.dart:301:19) #7 _RawReceivePortImpl._handleMessage (dart:isolate-patch/isolate_patch.dart:168:12) --- Re-run this test: python tools/test.py -m release -c dartk --nnbd weak --enable-experiment non-nullable language/nnbd/normalization/generic_function_type_object_normalization_test [00:00 | 100% | + 0 | - 1] === 0 tests passed, 1 failed === ```
1.0
[Test] Test runner doesn't seem to honor weak/strong flag - See detailed transcript below, but in summary: Running test.py on a test with `--nnbd=weak` passed as an argument results in this command line: ``` DART_CONFIGURATION=ReleaseX64 xcodebuild/ReleaseX64/dart --enable-experiment=non-nullable --ignore-unrecognized-flags --packages=/Users/leafp/src/dart-repo/sdk/.packages /Users/leafp/src/dart-repo/sdk/tests/language/nnbd/normalization/generic_function_type_object_normalization_test.dart ``` Note that this command line does not set the mode to weak mode. Running with `--nnbd=strong` results in the same command line. cc @munificent @sortie ``` leafp-macbookpro:sdk leafp$ python tools/test.py -c dartk -m release --enable-experiment=non-nullable --nnbd=weak tests/language/nnbd/normalization/generic_function_type_object_normalization_test.dart Test configuration: custom configuration(architecture: x64, compiler: dartk, mode: release, runtime: vm, system: mac, nnbd: weak, enable-experiment: [non-nullable]) Suites tested: language FAILED: dartk-vm release_x64 language/nnbd/normalization/generic_function_type_object_normalization_test Expected: Pass Actual: RuntimeError --- Command "vm" (took 602ms): DART_CONFIGURATION=ReleaseX64 xcodebuild/ReleaseX64/dart --enable-experiment=non-nullable --ignore-unrecognized-flags --packages=/Users/leafp/src/dart-repo/sdk/.packages /Users/leafp/src/dart-repo/sdk/tests/language/nnbd/normalization/generic_function_type_object_normalization_test.dart exit code: 255 stderr: Unhandled exception: Expect.notEquals(unexpected: <<R0 extends Future<Never>, R1 extends Never, R2 extends Null>() => Null>, actual:<<R0 extends Future<Never>, R1 extends Never, R2 extends Null>() => Null>) fails. #0 Expect._fail (package:expect/expect.dart:685:5) #1 Expect.notEquals (package:expect/expect.dart:306:5) #2 checkNotEquals2 (file:///Users/leafp/src/dart-repo/sdk/tests/language/nnbd/normalization/type_builder.dart:136:10) #3 checkTypeNotEquals2 (file:///Users/leafp/src/dart-repo/sdk/tests/language/nnbd/normalization/generic_function_type_object_normalization_test.dart:22:3) #4 neverBoundTests (file:///Users/leafp/src/dart-repo/sdk/tests/language/nnbd/normalization/generic_function_type_object_normalization_test.dart:175:5) #5 main (file:///Users/leafp/src/dart-repo/sdk/tests/language/nnbd/normalization/generic_function_type_object_normalization_test.dart:199:3) #6 _startIsolate.<anonymous closure> (dart:isolate-patch/isolate_patch.dart:301:19) #7 _RawReceivePortImpl._handleMessage (dart:isolate-patch/isolate_patch.dart:168:12) --- Re-run this test: python tools/test.py -m release -c dartk --nnbd weak --enable-experiment non-nullable language/nnbd/normalization/generic_function_type_object_normalization_test [00:00 | 100% | + 0 | - 1] === 0 tests passed, 1 failed === ```
non_perf
test runner doesn t seem to honor weak strong flag see detailed transcript below but in summary running test py on a test with nnbd weak passed as an argument results in this command line dart configuration xcodebuild dart enable experiment non nullable ignore unrecognized flags packages users leafp src dart repo sdk packages users leafp src dart repo sdk tests language nnbd normalization generic function type object normalization test dart note that this command line does not set the mode to weak mode running with nnbd strong results in the same command line cc munificent sortie leafp macbookpro sdk leafp python tools test py c dartk m release enable experiment non nullable nnbd weak tests language nnbd normalization generic function type object normalization test dart test configuration custom configuration architecture compiler dartk mode release runtime vm system mac nnbd weak enable experiment suites tested language failed dartk vm release language nnbd normalization generic function type object normalization test expected pass actual runtimeerror command vm took dart configuration xcodebuild dart enable experiment non nullable ignore unrecognized flags packages users leafp src dart repo sdk packages users leafp src dart repo sdk tests language nnbd normalization generic function type object normalization test dart exit code stderr unhandled exception expect notequals unexpected extends never extends null null actual extends never extends null null fails expect fail package expect expect dart expect notequals package expect expect dart file users leafp src dart repo sdk tests language nnbd normalization type builder dart file users leafp src dart repo sdk tests language nnbd normalization generic function type object normalization test dart neverboundtests file users leafp src dart repo sdk tests language nnbd normalization generic function type object normalization test dart main file users leafp src dart repo sdk tests language nnbd normalization generic function type object normalization test dart startisolate dart isolate patch isolate patch dart rawreceiveportimpl handlemessage dart isolate patch isolate patch dart re run this test python tools test py m release c dartk nnbd weak enable experiment non nullable language nnbd normalization generic function type object normalization test tests passed failed
0
23,646
12,056,063,823
IssuesEvent
2020-04-15 13:56:08
NREL/EnergyPlus
https://api.github.com/repos/NREL/EnergyPlus
opened
Request to change SQL data type for Tabular Data from text to value
Performance
Issue overview -------------- Interface developer notes performance problems accessing SQLite database because the data type for tabular data are strings instead of a value. Heavy use of custom and predefined tabular data makes for large database files. Interface developer feels that they can improve performance during post processing if the data types were changed to "value." I recall that this was tried, by @kbenne, way back when SQLite content was greatly expanded. But I cannot remember what the issue was. Seems like occasionally some special values wouldn't translate okay. Perhaps with an increased focus on performance this could be revisited. ![79130266-4a9d8a80-7d6c-11ea-9fd9-c3bf8a96414b](https://user-images.githubusercontent.com/8754769/79345300-f080fe80-7efe-11ea-813f-dc7df4e7216c.png) ### Details Some additional details for this issue (if relevant): - Platform (Operating system, version) - Version of EnergyPlus (if using an intermediate build, include SHA) - Unmethours link or helpdesk ticket number ### Checklist Add to this list or remove from it as applicable. This is a simple templated set of guidelines. - [ ] Defect file added (list location of defect file here) - [ ] Ticket added to Pivotal for defect (development team task) - [ ] Pull request created (the pull request will have additional tasks related to reviewing changes that fix this defect)
True
Request to change SQL data type for Tabular Data from text to value - Issue overview -------------- Interface developer notes performance problems accessing SQLite database because the data type for tabular data are strings instead of a value. Heavy use of custom and predefined tabular data makes for large database files. Interface developer feels that they can improve performance during post processing if the data types were changed to "value." I recall that this was tried, by @kbenne, way back when SQLite content was greatly expanded. But I cannot remember what the issue was. Seems like occasionally some special values wouldn't translate okay. Perhaps with an increased focus on performance this could be revisited. ![79130266-4a9d8a80-7d6c-11ea-9fd9-c3bf8a96414b](https://user-images.githubusercontent.com/8754769/79345300-f080fe80-7efe-11ea-813f-dc7df4e7216c.png) ### Details Some additional details for this issue (if relevant): - Platform (Operating system, version) - Version of EnergyPlus (if using an intermediate build, include SHA) - Unmethours link or helpdesk ticket number ### Checklist Add to this list or remove from it as applicable. This is a simple templated set of guidelines. - [ ] Defect file added (list location of defect file here) - [ ] Ticket added to Pivotal for defect (development team task) - [ ] Pull request created (the pull request will have additional tasks related to reviewing changes that fix this defect)
perf
request to change sql data type for tabular data from text to value issue overview interface developer notes performance problems accessing sqlite database because the data type for tabular data are strings instead of a value heavy use of custom and predefined tabular data makes for large database files interface developer feels that they can improve performance during post processing if the data types were changed to value i recall that this was tried by kbenne way back when sqlite content was greatly expanded but i cannot remember what the issue was seems like occasionally some special values wouldn t translate okay perhaps with an increased focus on performance this could be revisited details some additional details for this issue if relevant platform operating system version version of energyplus if using an intermediate build include sha unmethours link or helpdesk ticket number checklist add to this list or remove from it as applicable this is a simple templated set of guidelines defect file added list location of defect file here ticket added to pivotal for defect development team task pull request created the pull request will have additional tasks related to reviewing changes that fix this defect
1
316,214
23,619,773,455
IssuesEvent
2022-08-24 19:19:46
Kong/gateway-operator
https://api.github.com/repos/Kong/gateway-operator
opened
Document Supported Gateway Topologies
documentation area/kep area/scalability
### Is there an existing issue for this? - [X] I have searched the existing issues ### Problem Statement Currently in our early alpha stage we support only a single monolithic `Gateway` in that there's only one `DataPlane` behind it, and only one "instance" (`Pod`, to be precise) behind that. We do however have notions of other features which come with some topology considerations: - [ ] hybrid mode https://github.com/Kong/gateway-operator/issues/229 - [ ] horizontal pod scaling https://github.com/Kong/gateway-operator/issues/170 https://github.com/Kong/gateway-operator/issues/171 The purpose of this task is to document the topologies that we expect to support when we go into beta within the [relevant KEP](https://github.com/Kong/gateway-operator/blob/main/keps/0001-managed-gateways.md) to build consensus about this, and for posterity. ### Proposed Solution _No response_ ### Additional information _No response_ ### Acceptance Criteria _No response_
1.0
Document Supported Gateway Topologies - ### Is there an existing issue for this? - [X] I have searched the existing issues ### Problem Statement Currently in our early alpha stage we support only a single monolithic `Gateway` in that there's only one `DataPlane` behind it, and only one "instance" (`Pod`, to be precise) behind that. We do however have notions of other features which come with some topology considerations: - [ ] hybrid mode https://github.com/Kong/gateway-operator/issues/229 - [ ] horizontal pod scaling https://github.com/Kong/gateway-operator/issues/170 https://github.com/Kong/gateway-operator/issues/171 The purpose of this task is to document the topologies that we expect to support when we go into beta within the [relevant KEP](https://github.com/Kong/gateway-operator/blob/main/keps/0001-managed-gateways.md) to build consensus about this, and for posterity. ### Proposed Solution _No response_ ### Additional information _No response_ ### Acceptance Criteria _No response_
non_perf
document supported gateway topologies is there an existing issue for this i have searched the existing issues problem statement currently in our early alpha stage we support only a single monolithic gateway in that there s only one dataplane behind it and only one instance pod to be precise behind that we do however have notions of other features which come with some topology considerations hybrid mode horizontal pod scaling the purpose of this task is to document the topologies that we expect to support when we go into beta within the to build consensus about this and for posterity proposed solution no response additional information no response acceptance criteria no response
0
829,767
31,897,169,510
IssuesEvent
2023-09-18 03:39:59
wso2/api-manager
https://api.github.com/repos/wso2/api-manager
closed
[4.2.0] Upgrade REST API version in Docs
Priority/Highest Component/APIM
This issue is used to track the REST API upgrade in Docs. Major Rest API versions are modified as below publisher : v3--> v4 (major version: v4 , latest version : v4) admin : v3 --> v4 (major version: v4 , latest version : v4) devportal : v2 --> (major version: v3 , latest version : v3)
1.0
[4.2.0] Upgrade REST API version in Docs - This issue is used to track the REST API upgrade in Docs. Major Rest API versions are modified as below publisher : v3--> v4 (major version: v4 , latest version : v4) admin : v3 --> v4 (major version: v4 , latest version : v4) devportal : v2 --> (major version: v3 , latest version : v3)
non_perf
upgrade rest api version in docs this issue is used to track the rest api upgrade in docs major rest api versions are modified as below publisher major version latest version admin major version latest version devportal major version latest version
0
48,373
25,498,732,103
IssuesEvent
2022-11-28 00:22:20
cessen/ropey
https://api.github.com/repos/cessen/ropey
closed
Make Lines iterator more efficient
performance
Currently the `Lines` iterator is roughly equivalent to just calling `Rope::line()` repeatedly with an incrementing index. This is O(log N) for each call to `Lines::next()`, and also is just generally less efficient than it needs to be. This is not only sub-optimal, but also stands out compared to the other iterators which are all O(1) and very fast. It should be possible to also make `Lines` O(1) and just generally more efficient.
True
Make Lines iterator more efficient - Currently the `Lines` iterator is roughly equivalent to just calling `Rope::line()` repeatedly with an incrementing index. This is O(log N) for each call to `Lines::next()`, and also is just generally less efficient than it needs to be. This is not only sub-optimal, but also stands out compared to the other iterators which are all O(1) and very fast. It should be possible to also make `Lines` O(1) and just generally more efficient.
perf
make lines iterator more efficient currently the lines iterator is roughly equivalent to just calling rope line repeatedly with an incrementing index this is o log n for each call to lines next and also is just generally less efficient than it needs to be this is not only sub optimal but also stands out compared to the other iterators which are all o and very fast it should be possible to also make lines o and just generally more efficient
1
24,158
12,226,403,415
IssuesEvent
2020-05-03 10:45:17
returntocorp/semgrep
https://api.github.com/repos/returntocorp/semgrep
closed
Investigate slow semgrep perf
performance
Scanning over https://github.com/sobolevn/python-code-disasters has my CPU pinned at 100% for 2 hours. ``` docker run --rm -v $(pwd):/home/repo returntocorp/sgrep:0.4.9 --json --config=https://sgrep.live/c/r/r2c --skip-pattern-validation ``` Adding as a performance investigation / something to add to our perf regression suite.
True
Investigate slow semgrep perf - Scanning over https://github.com/sobolevn/python-code-disasters has my CPU pinned at 100% for 2 hours. ``` docker run --rm -v $(pwd):/home/repo returntocorp/sgrep:0.4.9 --json --config=https://sgrep.live/c/r/r2c --skip-pattern-validation ``` Adding as a performance investigation / something to add to our perf regression suite.
perf
investigate slow semgrep perf scanning over has my cpu pinned at for hours docker run rm v pwd home repo returntocorp sgrep json config skip pattern validation adding as a performance investigation something to add to our perf regression suite
1
40,875
21,259,429,633
IssuesEvent
2022-04-13 01:21:19
reclosedev/requests-cache
https://api.github.com/repos/reclosedev/requests-cache
opened
Filesystem backend: Add option to store response content in a separate file
enhancement performance
Related to #407. This would help with a few use cases that have come up a few of times now: * Efficiently storing large response contents (without the serialization bottleneck) * Caching responses that primarily contain media or other files you want to access outsize of requests-cache
True
Filesystem backend: Add option to store response content in a separate file - Related to #407. This would help with a few use cases that have come up a few of times now: * Efficiently storing large response contents (without the serialization bottleneck) * Caching responses that primarily contain media or other files you want to access outsize of requests-cache
perf
filesystem backend add option to store response content in a separate file related to this would help with a few use cases that have come up a few of times now efficiently storing large response contents without the serialization bottleneck caching responses that primarily contain media or other files you want to access outsize of requests cache
1
22,410
11,595,707,394
IssuesEvent
2020-02-24 17:30:18
qbittorrent/qBittorrent
https://api.github.com/repos/qbittorrent/qBittorrent
closed
stutters
Duplicate Performance
**Please provide the following information** ### qBittorrent version and Operating System Os: Windows 7 64 Bit Ultimate Qbittorrent 4.1.1 64 Bit ### If on linux, libtorrent and Qt version N/A ### What is the problem the setup menu stops every time I open it. Or, when downloading 1 torrent, it will still be blocked. Videos: https://drive.google.com/open?id=11KTLTG_jUyCCNUyZ9cwcHPZ9IKpD3AfV ### What is the expected behavior Not to stumble. (no stutters!) ### Steps to reproduce 1. Go to settings. 2. click on any of the menus. ### Extra info(if any) 7Gb ram Q6600 processor 64Gb SSD qbittorrent installed and 1Tb Western Digital 7200rpm HDD.
True
stutters - **Please provide the following information** ### qBittorrent version and Operating System Os: Windows 7 64 Bit Ultimate Qbittorrent 4.1.1 64 Bit ### If on linux, libtorrent and Qt version N/A ### What is the problem the setup menu stops every time I open it. Or, when downloading 1 torrent, it will still be blocked. Videos: https://drive.google.com/open?id=11KTLTG_jUyCCNUyZ9cwcHPZ9IKpD3AfV ### What is the expected behavior Not to stumble. (no stutters!) ### Steps to reproduce 1. Go to settings. 2. click on any of the menus. ### Extra info(if any) 7Gb ram Q6600 processor 64Gb SSD qbittorrent installed and 1Tb Western Digital 7200rpm HDD.
perf
stutters please provide the following information qbittorrent version and operating system os windows bit ultimate qbittorrent bit if on linux libtorrent and qt version n a what is the problem the setup menu stops every time i open it or when downloading torrent it will still be blocked videos what is the expected behavior not to stumble no stutters steps to reproduce go to settings click on any of the menus extra info if any ram processor ssd qbittorrent installed and western digital hdd
1
56,665
8,109,213,931
IssuesEvent
2018-08-14 06:36:54
emotion-js/emotion
https://api.github.com/repos/emotion-js/emotion
closed
Browser Support?
documentation question stale
Just curious if there's a spot I wasn't able to find in the docs that lists what browsers are supported by emotion. I just deployed a component that uses it and have a torrent of errors from IE11 saying that it doesn't understand `WeakMap` - is this an oversight, recommended to be polyfilled independently, or intended not to support IE11? Whatever the case, I'd be happy to help out with a PR if that would be useful, just want to make sure I understand the intent here!
1.0
Browser Support? - Just curious if there's a spot I wasn't able to find in the docs that lists what browsers are supported by emotion. I just deployed a component that uses it and have a torrent of errors from IE11 saying that it doesn't understand `WeakMap` - is this an oversight, recommended to be polyfilled independently, or intended not to support IE11? Whatever the case, I'd be happy to help out with a PR if that would be useful, just want to make sure I understand the intent here!
non_perf
browser support just curious if there s a spot i wasn t able to find in the docs that lists what browsers are supported by emotion i just deployed a component that uses it and have a torrent of errors from saying that it doesn t understand weakmap is this an oversight recommended to be polyfilled independently or intended not to support whatever the case i d be happy to help out with a pr if that would be useful just want to make sure i understand the intent here
0
365,907
10,799,585,405
IssuesEvent
2019-11-06 12:31:23
DFO-Ocean-Navigator/Ocean-Data-Map-Project
https://api.github.com/repos/DFO-Ocean-Navigator/Ocean-Data-Map-Project
closed
Enable OpenLayers zoom slider
Javascript New Feature Priority: Low
I'm thinking the version with the slider in between the zoom buttons. ![image](https://user-images.githubusercontent.com/5572045/64077766-f5fe4100-ccad-11e9-97a4-f2e14634f54c.png) Example code: https://openlayers.org/en/latest/examples/zoomslider.html
1.0
Enable OpenLayers zoom slider - I'm thinking the version with the slider in between the zoom buttons. ![image](https://user-images.githubusercontent.com/5572045/64077766-f5fe4100-ccad-11e9-97a4-f2e14634f54c.png) Example code: https://openlayers.org/en/latest/examples/zoomslider.html
non_perf
enable openlayers zoom slider i m thinking the version with the slider in between the zoom buttons example code
0
651,023
21,448,104,383
IssuesEvent
2022-04-25 08:37:06
space-wizards/space-station-14
https://api.github.com/repos/space-wizards/space-station-14
closed
Implement hardsuits properly
Priority: 2-Before Release Issue: Feature Request Difficulty: 2-Medium
<!-- To automatically tag this issue, add the uppercase label(s) surrounded by brackets below, for example: [LABEL] --> ## Description <!-- Explain your issue in detail, including the steps to reproduce it if applicable. Issues without proper explanation are liable to be closed by maintainers.--> Currently hardsuits are two pieces, a helmet and a torso. They are functional in-game currently as they protect against depressurization through PressureProtection.cs. However the hardsuit should be one piece. A torso that when donned puts a "helmet" icon in your hotbar that toggles the helmet of the suit on or off. When the helmet is toggled on, the suit should draw from an oxygen tank on the players belt, hand, back or "suit-slot". What should happen when a player dons a hardsuit: 1. A helmet hotbar icon appears that toggles the suits helmet on/off. 2. A suit-slot should appear as a separate slot on the users hotbar. This can be used for hooking stuff like tanks, tools and other gadgets. 3. When toggled off, the user should be able to eat, drink, and wear hats like normal. However, a worn hat will prevent the hardsuit helmet from closing when pressed. 4. When toggled on, the suit should draw from an oxygen tank on the players belt, hand, back or "suit-slot". It also should provide another hotbar icon for the flashlight on the hardsuit. The flashlight powercell will be swappable by screwdrivering the hardsuit and removing the cell. I think that hardsuits should be a rarer find in SS14. I also think the order you put a hardsuit on should effect some stuff, like if you put the hardsuit on over your belt, your belt would be inaccessible, but if you put your hardsuit on and then your belt, you'd be able to access it.
1.0
Implement hardsuits properly - <!-- To automatically tag this issue, add the uppercase label(s) surrounded by brackets below, for example: [LABEL] --> ## Description <!-- Explain your issue in detail, including the steps to reproduce it if applicable. Issues without proper explanation are liable to be closed by maintainers.--> Currently hardsuits are two pieces, a helmet and a torso. They are functional in-game currently as they protect against depressurization through PressureProtection.cs. However the hardsuit should be one piece. A torso that when donned puts a "helmet" icon in your hotbar that toggles the helmet of the suit on or off. When the helmet is toggled on, the suit should draw from an oxygen tank on the players belt, hand, back or "suit-slot". What should happen when a player dons a hardsuit: 1. A helmet hotbar icon appears that toggles the suits helmet on/off. 2. A suit-slot should appear as a separate slot on the users hotbar. This can be used for hooking stuff like tanks, tools and other gadgets. 3. When toggled off, the user should be able to eat, drink, and wear hats like normal. However, a worn hat will prevent the hardsuit helmet from closing when pressed. 4. When toggled on, the suit should draw from an oxygen tank on the players belt, hand, back or "suit-slot". It also should provide another hotbar icon for the flashlight on the hardsuit. The flashlight powercell will be swappable by screwdrivering the hardsuit and removing the cell. I think that hardsuits should be a rarer find in SS14. I also think the order you put a hardsuit on should effect some stuff, like if you put the hardsuit on over your belt, your belt would be inaccessible, but if you put your hardsuit on and then your belt, you'd be able to access it.
non_perf
implement hardsuits properly description currently hardsuits are two pieces a helmet and a torso they are functional in game currently as they protect against depressurization through pressureprotection cs however the hardsuit should be one piece a torso that when donned puts a helmet icon in your hotbar that toggles the helmet of the suit on or off when the helmet is toggled on the suit should draw from an oxygen tank on the players belt hand back or suit slot what should happen when a player dons a hardsuit a helmet hotbar icon appears that toggles the suits helmet on off a suit slot should appear as a separate slot on the users hotbar this can be used for hooking stuff like tanks tools and other gadgets when toggled off the user should be able to eat drink and wear hats like normal however a worn hat will prevent the hardsuit helmet from closing when pressed when toggled on the suit should draw from an oxygen tank on the players belt hand back or suit slot it also should provide another hotbar icon for the flashlight on the hardsuit the flashlight powercell will be swappable by screwdrivering the hardsuit and removing the cell i think that hardsuits should be a rarer find in i also think the order you put a hardsuit on should effect some stuff like if you put the hardsuit on over your belt your belt would be inaccessible but if you put your hardsuit on and then your belt you d be able to access it
0
12,064
7,775,270,400
IssuesEvent
2018-06-05 01:48:00
deeplearning4j/deeplearning4j
https://api.github.com/repos/deeplearning4j/deeplearning4j
opened
DL4J: Benchmarks, resnet50: can't run batch size 16, can run batch 32
Bug DL4J Performance
I believe this is related to how CuDNN is configured... this particular model is set to ```ConvolutionLayer.AlgoMode.PREFER_FASTEST```; I suspect it's the CuDNN mode internally that is the reason (i.e., batch size 32 uses a different mode that requires less memory). Now, CuDNN (or at least some of the more recent versions?) does support specifying a maximum workspace size. We may be able to inspect the amount of available memory, and have CuDNN base it's algorithm selection on that. Alternatively, if we detect an OOM, we might be able to enforce use of a less memory-intensive algorithm rather than failing outright.
True
DL4J: Benchmarks, resnet50: can't run batch size 16, can run batch 32 - I believe this is related to how CuDNN is configured... this particular model is set to ```ConvolutionLayer.AlgoMode.PREFER_FASTEST```; I suspect it's the CuDNN mode internally that is the reason (i.e., batch size 32 uses a different mode that requires less memory). Now, CuDNN (or at least some of the more recent versions?) does support specifying a maximum workspace size. We may be able to inspect the amount of available memory, and have CuDNN base it's algorithm selection on that. Alternatively, if we detect an OOM, we might be able to enforce use of a less memory-intensive algorithm rather than failing outright.
perf
benchmarks can t run batch size can run batch i believe this is related to how cudnn is configured this particular model is set to convolutionlayer algomode prefer fastest i suspect it s the cudnn mode internally that is the reason i e batch size uses a different mode that requires less memory now cudnn or at least some of the more recent versions does support specifying a maximum workspace size we may be able to inspect the amount of available memory and have cudnn base it s algorithm selection on that alternatively if we detect an oom we might be able to enforce use of a less memory intensive algorithm rather than failing outright
1
7,323
5,970,677,253
IssuesEvent
2017-05-30 23:29:30
mozilla/thimble.mozilla.org
https://api.github.com/repos/mozilla/thimble.mozilla.org
closed
Publish server - Update knexjs
Performance publish.webmaker.org
We should update knex.js to the latest version as their changelog indicates some perf fixes as well as a switch from `pool2` to `generic-pool` as their connection pooling interface (I don't know if there will be significant implications for this but I assume better connection management). It is a bit of a random fix but is still probably worth doing. cc @cadecairos
True
Publish server - Update knexjs - We should update knex.js to the latest version as their changelog indicates some perf fixes as well as a switch from `pool2` to `generic-pool` as their connection pooling interface (I don't know if there will be significant implications for this but I assume better connection management). It is a bit of a random fix but is still probably worth doing. cc @cadecairos
perf
publish server update knexjs we should update knex js to the latest version as their changelog indicates some perf fixes as well as a switch from to generic pool as their connection pooling interface i don t know if there will be significant implications for this but i assume better connection management it is a bit of a random fix but is still probably worth doing cc cadecairos
1
600,774
18,356,285,077
IssuesEvent
2021-10-08 18:41:18
vtdangg/fa21-cse110-lab3
https://api.github.com/repos/vtdangg/fa21-cse110-lab3
opened
Use CSS Selectors
enhancement high priority collaborate
## What is the purpose fo the new feature or addition? To style the HTML elements from the meeting minutes. ## A clear and concise description of what the addition is and what it does. Each selector to be used will target a different identifier on the HTML element.
1.0
Use CSS Selectors - ## What is the purpose fo the new feature or addition? To style the HTML elements from the meeting minutes. ## A clear and concise description of what the addition is and what it does. Each selector to be used will target a different identifier on the HTML element.
non_perf
use css selectors what is the purpose fo the new feature or addition to style the html elements from the meeting minutes a clear and concise description of what the addition is and what it does each selector to be used will target a different identifier on the html element
0
13,468
8,228,076,532
IssuesEvent
2018-09-07 02:54:55
deeplearning4j/deeplearning4j
https://api.github.com/repos/deeplearning4j/deeplearning4j
opened
DL4J: MLP Profiling
DL4J Performance
ERROR: type should be string, got "\r\nhttps://gist.github.com/AlexDBlack/7fa542887d5e7933fc2c866819d9e1ac\r\n\r\nTracing results: (tracing adds per-method overhead: 5400 ms per epoch average)\r\n\r\n- Updater: 10500ms (41%)\r\n- Dropout, forward pass: 1649ms (6%) - mainly RNG op\r\n- Gemm, forward pass: 1423ms (6%)\r\n- Score calculation (output layer) is 2200ms (9%)\r\n\r\nNote also that for profiling (less accurate for small method calls), updater is 53% of runtime.\r\n\r\nNote that this is AMSGrapd updater, so one of the more complex ones (IIRC it has 3x parameters as state).\r\n\r\nAnyway, I see two main areas for improvement here:\r\n1. Dropout (known issue; there's multiple github issues open about it)\r\n2. Native updaters (mainly for better memory access patterns - iterate over arrays once, rather than N times)\r\n\r\n\r\n[Perf-2018-09-07.zip](https://github.com/deeplearning4j/deeplearning4j/files/2359313/Perf-2018-09-07.zip)\r\n\r\n\r\n"
True
DL4J: MLP Profiling - https://gist.github.com/AlexDBlack/7fa542887d5e7933fc2c866819d9e1ac Tracing results: (tracing adds per-method overhead: 5400 ms per epoch average) - Updater: 10500ms (41%) - Dropout, forward pass: 1649ms (6%) - mainly RNG op - Gemm, forward pass: 1423ms (6%) - Score calculation (output layer) is 2200ms (9%) Note also that for profiling (less accurate for small method calls), updater is 53% of runtime. Note that this is AMSGrapd updater, so one of the more complex ones (IIRC it has 3x parameters as state). Anyway, I see two main areas for improvement here: 1. Dropout (known issue; there's multiple github issues open about it) 2. Native updaters (mainly for better memory access patterns - iterate over arrays once, rather than N times) [Perf-2018-09-07.zip](https://github.com/deeplearning4j/deeplearning4j/files/2359313/Perf-2018-09-07.zip)
perf
mlp profiling tracing results tracing adds per method overhead ms per epoch average updater dropout forward pass mainly rng op gemm forward pass score calculation output layer is note also that for profiling less accurate for small method calls updater is of runtime note that this is amsgrapd updater so one of the more complex ones iirc it has parameters as state anyway i see two main areas for improvement here dropout known issue there s multiple github issues open about it native updaters mainly for better memory access patterns iterate over arrays once rather than n times
1
13,171
8,135,211,436
IssuesEvent
2018-08-20 01:10:03
OctopusDeploy/Issues
https://api.github.com/repos/OctopusDeploy/Issues
closed
As more tasks are queued, starting new tasks takes longer
area/performance
The more tasks that are queued, the longer it takes to start a new task. This starts being noticeable at the 500 queued task mark. Consequently the server has trouble reaching it's task cap if the tasks are relatively short.
True
As more tasks are queued, starting new tasks takes longer - The more tasks that are queued, the longer it takes to start a new task. This starts being noticeable at the 500 queued task mark. Consequently the server has trouble reaching it's task cap if the tasks are relatively short.
perf
as more tasks are queued starting new tasks takes longer the more tasks that are queued the longer it takes to start a new task this starts being noticeable at the queued task mark consequently the server has trouble reaching it s task cap if the tasks are relatively short
1
23,677
12,061,071,834
IssuesEvent
2020-04-15 22:42:30
microsoft/MixedRealityToolkit-Unity
https://api.github.com/repos/microsoft/MixedRealityToolkit-Unity
closed
Add Unity profiler markers to key MRTK code to assist in performance understanding / optimization
Feature Request Performance
MRTK does a lot of work on the behalf of applications. That work has a cost associated. Adding profiler markers can help the MRTK team better optimize critical code paths. This will also allow customers to better understand the costs associated with MRTK and to best optimize their applications. These markers are only active in development builds, per Unity documentation and would take the form of: ```c# private static readonly ProfilerMarker marker = new ProfilerMarker("[MRTK] class.method - optional note"); Method() { using (marker.Auto()) { ... } } ``` This feature should be implemented primarily in inner loop code for each of the core systems and providers. Note: some systems do not have inner loop code. Core Systems - ~~Boundary~~ - No inner loop functionality. - [X] Camera (#7654) - [x] Diagnostics (#7652) - [x] Visual Profiler - [x] Input (#7590) - [x] Unity input controllers - [x] OpenVR - [x] Windows Mixed Reality - [x] Windows Voice - [x] Pointers - [x] Scene (#7658) - [x] Spatial Awareness (#7649, #7654) - [x] Windows Mixed Reality Mesh Observer - [x] Spatial Object Mesh Observer - [x] Teleport (#7653) Extensions (#7661) - [x] Hand Physics - [x] Scene Transition - [x] Tracking Lost Documentation updates (#7671) - [x] Data providers - [x] Input - [x] Spatial awareness - [x] Performance
True
Add Unity profiler markers to key MRTK code to assist in performance understanding / optimization - MRTK does a lot of work on the behalf of applications. That work has a cost associated. Adding profiler markers can help the MRTK team better optimize critical code paths. This will also allow customers to better understand the costs associated with MRTK and to best optimize their applications. These markers are only active in development builds, per Unity documentation and would take the form of: ```c# private static readonly ProfilerMarker marker = new ProfilerMarker("[MRTK] class.method - optional note"); Method() { using (marker.Auto()) { ... } } ``` This feature should be implemented primarily in inner loop code for each of the core systems and providers. Note: some systems do not have inner loop code. Core Systems - ~~Boundary~~ - No inner loop functionality. - [X] Camera (#7654) - [x] Diagnostics (#7652) - [x] Visual Profiler - [x] Input (#7590) - [x] Unity input controllers - [x] OpenVR - [x] Windows Mixed Reality - [x] Windows Voice - [x] Pointers - [x] Scene (#7658) - [x] Spatial Awareness (#7649, #7654) - [x] Windows Mixed Reality Mesh Observer - [x] Spatial Object Mesh Observer - [x] Teleport (#7653) Extensions (#7661) - [x] Hand Physics - [x] Scene Transition - [x] Tracking Lost Documentation updates (#7671) - [x] Data providers - [x] Input - [x] Spatial awareness - [x] Performance
perf
add unity profiler markers to key mrtk code to assist in performance understanding optimization mrtk does a lot of work on the behalf of applications that work has a cost associated adding profiler markers can help the mrtk team better optimize critical code paths this will also allow customers to better understand the costs associated with mrtk and to best optimize their applications these markers are only active in development builds per unity documentation and would take the form of c private static readonly profilermarker marker new profilermarker class method optional note method using marker auto this feature should be implemented primarily in inner loop code for each of the core systems and providers note some systems do not have inner loop code core systems boundary no inner loop functionality camera diagnostics visual profiler input unity input controllers openvr windows mixed reality windows voice pointers scene spatial awareness windows mixed reality mesh observer spatial object mesh observer teleport extensions hand physics scene transition tracking lost documentation updates data providers input spatial awareness performance
1
8,749
2,611,542,789
IssuesEvent
2015-02-27 06:11:31
chrsmith/hedgewars
https://api.github.com/repos/chrsmith/hedgewars
opened
siPointType (LuaAPI: constant for SendStat) doesn’t work when used the first time in game.
auto-migrated Priority-Medium Type-Defect
``` What steps will reproduce the problem? 1. Start one of the new target practice missions 2. Finish it or lose by failing (doesn’t matter) 3. See the ranking. Is it the word “kills” or “points”? 4. Start the mission again 5. Repeat step 3 What is the expected output? What do you see instead? At steps 3 and 5, I want to see the word “points”. At step 3, I actually see the word “kills”, at step 5 “points. That’s weird. :/ What version of the product are you using? On what operating system? r2f19ff0ded73 on GNU/Linux. Please provide any additional information below. I fear there may be other stats-screen related issues. ``` Original issue reported on code.google.com by `[email protected]` on 15 Dec 2014 at 9:11
1.0
siPointType (LuaAPI: constant for SendStat) doesn’t work when used the first time in game. - ``` What steps will reproduce the problem? 1. Start one of the new target practice missions 2. Finish it or lose by failing (doesn’t matter) 3. See the ranking. Is it the word “kills” or “points”? 4. Start the mission again 5. Repeat step 3 What is the expected output? What do you see instead? At steps 3 and 5, I want to see the word “points”. At step 3, I actually see the word “kills”, at step 5 “points. That’s weird. :/ What version of the product are you using? On what operating system? r2f19ff0ded73 on GNU/Linux. Please provide any additional information below. I fear there may be other stats-screen related issues. ``` Original issue reported on code.google.com by `[email protected]` on 15 Dec 2014 at 9:11
non_perf
sipointtype luaapi constant for sendstat doesn’t work when used the first time in game what steps will reproduce the problem start one of the new target practice missions finish it or lose by failing doesn’t matter see the ranking is it the word “kills” or “points” start the mission again repeat step what is the expected output what do you see instead at steps and i want to see the word “points” at step i actually see the word “kills” at step “points that’s weird what version of the product are you using on what operating system on gnu linux please provide any additional information below i fear there may be other stats screen related issues original issue reported on code google com by almikes aol com on dec at
0
23,292
11,902,321,336
IssuesEvent
2020-03-30 13:48:29
scalableminds/webknossos
https://api.github.com/repos/scalableminds/webknossos
closed
The comment tab can be quite slow when there are lots of comments
frontend performance
`getDerivedStateFromProps` is a major performance bottleneck when there are lots of trees with comments. Even clicking on one comment can take up to 7 seconds until something happens. The culprit is most likely the sorting in `getDerivedStateFromProps` which could be easily cached (and maybe optimized in another way, as well).
True
The comment tab can be quite slow when there are lots of comments - `getDerivedStateFromProps` is a major performance bottleneck when there are lots of trees with comments. Even clicking on one comment can take up to 7 seconds until something happens. The culprit is most likely the sorting in `getDerivedStateFromProps` which could be easily cached (and maybe optimized in another way, as well).
perf
the comment tab can be quite slow when there are lots of comments getderivedstatefromprops is a major performance bottleneck when there are lots of trees with comments even clicking on one comment can take up to seconds until something happens the culprit is most likely the sorting in getderivedstatefromprops which could be easily cached and maybe optimized in another way as well
1
78,694
15,051,594,451
IssuesEvent
2021-02-03 14:17:32
joomla/joomla-cms
https://api.github.com/repos/joomla/joomla-cms
closed
[4.0] UnknownAssetException: There is no "bootstrap.dropdown" asset of a "script" type in the registry.
No Code Attached Yet
### Steps to reproduce the issue I updated my copy of J4 from git this morning (past git update was from Jan 26) and now I get this error when trying to access the Administrator. ### Expected result Administrator opens. ### Actual result Get this error. Below is the call stack. Is there a database update that needs to be done for Bootstrap 5 such that a git pull is insufficient to update? Joomla\CMS\WebAsset\Exception\ UnknownAssetException in /var/www/html/libraries/src/WebAsset/WebAssetRegistry.php (line 132) WebAssetRegistry->get() in /var/www/html/libraries/src/WebAsset/WebAssetManager.php (line 257) WebAssetManager->useAsset() in /var/www/html/libraries/src/WebAsset/WebAssetManager.php (line 181) WebAssetManager->__call() in /var/www/html/libraries/src/HTML/Helpers/Bootstrap.php (line 232) Bootstrap::dropdown() in /var/www/html/libraries/src/HTML/HTMLHelper.php (line 322) HTMLHelper::call() in /var/www/html/libraries/src/HTML/HTMLHelper.php (line 154) HTMLHelper::_() in /var/www/html/administrator/modules/mod_user/tmpl/default.php (line 19) require('/var/www/html/administrator/modules/mod_user/tmpl/default.php') in /var/www/html/administrator/modules/mod_user/mod_user.php (line 16) include('/var/www/html/administrator/modules/mod_user/mod_user.php') in /var/www/html/libraries/src/Dispatcher/ModuleDispatcher.php (line 54) ModuleDispatcher::Joomla\CMS\Dispatcher\{closure}() in /var/www/html/libraries/src/Dispatcher/ModuleDispatcher.php (line 57) ModuleDispatcher->dispatch() in /var/www/html/libraries/src/Helper/ModuleHelper.php (line 293) ModuleHelper::renderRawModule() in /var/www/html/libraries/src/Helper/ModuleHelper.php (line 166) ModuleHelper::renderModule() in /var/www/html/libraries/src/Document/Renderer/Html/ModuleRenderer.php (line 97) ModuleRenderer->render() in /var/www/html/libraries/src/Document/Renderer/Html/ModulesRenderer.php (line 48) ModulesRenderer->render() in /var/www/html/libraries/src/Document/HtmlDocument.php (line 589) HtmlDocument->getBuffer() in /var/www/html/libraries/src/Document/HtmlDocument.php (line 895) HtmlDocument->_renderTemplate() in /var/www/html/libraries/src/Document/HtmlDocument.php (line 660) HtmlDocument->render() in /var/www/html/libraries/src/Document/ErrorDocument.php (line 140) ErrorDocument->render() in /var/www/html/libraries/src/Error/Renderer/HtmlRenderer.php (line 76) HtmlRenderer->render() in /var/www/html/libraries/src/Exception/ExceptionHandler.php (line 128) ExceptionHandler::render() in /var/www/html/libraries/src/Exception/ExceptionHandler.php (line 71) ExceptionHandler::handleException() in /var/www/html/libraries/src/Application/CMSApplication.php (line 299) CMSApplication->execute() in /var/www/html/administrator/includes/app.php (line 63) require_once('/var/www/html/administrator/includes/app.php') in /var/www/html/administrator/index.php (line 32) ### System information (as much as possible) ### Additional comments
1.0
[4.0] UnknownAssetException: There is no "bootstrap.dropdown" asset of a "script" type in the registry. - ### Steps to reproduce the issue I updated my copy of J4 from git this morning (past git update was from Jan 26) and now I get this error when trying to access the Administrator. ### Expected result Administrator opens. ### Actual result Get this error. Below is the call stack. Is there a database update that needs to be done for Bootstrap 5 such that a git pull is insufficient to update? Joomla\CMS\WebAsset\Exception\ UnknownAssetException in /var/www/html/libraries/src/WebAsset/WebAssetRegistry.php (line 132) WebAssetRegistry->get() in /var/www/html/libraries/src/WebAsset/WebAssetManager.php (line 257) WebAssetManager->useAsset() in /var/www/html/libraries/src/WebAsset/WebAssetManager.php (line 181) WebAssetManager->__call() in /var/www/html/libraries/src/HTML/Helpers/Bootstrap.php (line 232) Bootstrap::dropdown() in /var/www/html/libraries/src/HTML/HTMLHelper.php (line 322) HTMLHelper::call() in /var/www/html/libraries/src/HTML/HTMLHelper.php (line 154) HTMLHelper::_() in /var/www/html/administrator/modules/mod_user/tmpl/default.php (line 19) require('/var/www/html/administrator/modules/mod_user/tmpl/default.php') in /var/www/html/administrator/modules/mod_user/mod_user.php (line 16) include('/var/www/html/administrator/modules/mod_user/mod_user.php') in /var/www/html/libraries/src/Dispatcher/ModuleDispatcher.php (line 54) ModuleDispatcher::Joomla\CMS\Dispatcher\{closure}() in /var/www/html/libraries/src/Dispatcher/ModuleDispatcher.php (line 57) ModuleDispatcher->dispatch() in /var/www/html/libraries/src/Helper/ModuleHelper.php (line 293) ModuleHelper::renderRawModule() in /var/www/html/libraries/src/Helper/ModuleHelper.php (line 166) ModuleHelper::renderModule() in /var/www/html/libraries/src/Document/Renderer/Html/ModuleRenderer.php (line 97) ModuleRenderer->render() in /var/www/html/libraries/src/Document/Renderer/Html/ModulesRenderer.php (line 48) ModulesRenderer->render() in /var/www/html/libraries/src/Document/HtmlDocument.php (line 589) HtmlDocument->getBuffer() in /var/www/html/libraries/src/Document/HtmlDocument.php (line 895) HtmlDocument->_renderTemplate() in /var/www/html/libraries/src/Document/HtmlDocument.php (line 660) HtmlDocument->render() in /var/www/html/libraries/src/Document/ErrorDocument.php (line 140) ErrorDocument->render() in /var/www/html/libraries/src/Error/Renderer/HtmlRenderer.php (line 76) HtmlRenderer->render() in /var/www/html/libraries/src/Exception/ExceptionHandler.php (line 128) ExceptionHandler::render() in /var/www/html/libraries/src/Exception/ExceptionHandler.php (line 71) ExceptionHandler::handleException() in /var/www/html/libraries/src/Application/CMSApplication.php (line 299) CMSApplication->execute() in /var/www/html/administrator/includes/app.php (line 63) require_once('/var/www/html/administrator/includes/app.php') in /var/www/html/administrator/index.php (line 32) ### System information (as much as possible) ### Additional comments
non_perf
unknownassetexception there is no bootstrap dropdown asset of a script type in the registry steps to reproduce the issue i updated my copy of from git this morning past git update was from jan and now i get this error when trying to access the administrator expected result administrator opens actual result get this error below is the call stack is there a database update that needs to be done for bootstrap such that a git pull is insufficient to update joomla cms webasset exception unknownassetexception in var www html libraries src webasset webassetregistry php line webassetregistry get in var www html libraries src webasset webassetmanager php line webassetmanager useasset in var www html libraries src webasset webassetmanager php line webassetmanager call in var www html libraries src html helpers bootstrap php line bootstrap dropdown in var www html libraries src html htmlhelper php line htmlhelper call in var www html libraries src html htmlhelper php line htmlhelper in var www html administrator modules mod user tmpl default php line require var www html administrator modules mod user tmpl default php in var www html administrator modules mod user mod user php line include var www html administrator modules mod user mod user php in var www html libraries src dispatcher moduledispatcher php line moduledispatcher joomla cms dispatcher closure in var www html libraries src dispatcher moduledispatcher php line moduledispatcher dispatch in var www html libraries src helper modulehelper php line modulehelper renderrawmodule in var www html libraries src helper modulehelper php line modulehelper rendermodule in var www html libraries src document renderer html modulerenderer php line modulerenderer render in var www html libraries src document renderer html modulesrenderer php line modulesrenderer render in var www html libraries src document htmldocument php line htmldocument getbuffer in var www html libraries src document htmldocument php line htmldocument rendertemplate in var www html libraries src document htmldocument php line htmldocument render in var www html libraries src document errordocument php line errordocument render in var www html libraries src error renderer htmlrenderer php line htmlrenderer render in var www html libraries src exception exceptionhandler php line exceptionhandler render in var www html libraries src exception exceptionhandler php line exceptionhandler handleexception in var www html libraries src application cmsapplication php line cmsapplication execute in var www html administrator includes app php line require once var www html administrator includes app php in var www html administrator index php line system information as much as possible additional comments
0
69,574
17,767,908,845
IssuesEvent
2021-08-30 09:55:13
srodrigo/anime-suupu
https://api.github.com/repos/srodrigo/anime-suupu
closed
Upgrade to node 14.17
build
This removes some warnings then doing npm install. ``` #10 1.887 npm WARN EBADENGINE Unsupported engine { #10 1.887 npm WARN EBADENGINE package: '@jest/[email protected]', #10 1.887 npm WARN EBADENGINE required: { node: '^10.13.0 || ^12.13.0 || ^14.15.0 || >=15.0.0' }, #10 1.887 npm WARN EBADENGINE current: { node: 'v14.7.0', npm: '7.19.1' } #10 1.887 npm WARN EBADENGINE } ```
1.0
Upgrade to node 14.17 - This removes some warnings then doing npm install. ``` #10 1.887 npm WARN EBADENGINE Unsupported engine { #10 1.887 npm WARN EBADENGINE package: '@jest/[email protected]', #10 1.887 npm WARN EBADENGINE required: { node: '^10.13.0 || ^12.13.0 || ^14.15.0 || >=15.0.0' }, #10 1.887 npm WARN EBADENGINE current: { node: 'v14.7.0', npm: '7.19.1' } #10 1.887 npm WARN EBADENGINE } ```
non_perf
upgrade to node this removes some warnings then doing npm install npm warn ebadengine unsupported engine npm warn ebadengine package jest console npm warn ebadengine required node npm warn ebadengine current node npm npm warn ebadengine
0
195,256
22,295,916,439
IssuesEvent
2022-06-13 01:32:21
n-devs/uiWebView
https://api.github.com/repos/n-devs/uiWebView
opened
CVE-2022-25851 (High) detected in jpeg-js-0.3.5.tgz
security vulnerability
## CVE-2022-25851 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jpeg-js-0.3.5.tgz</b></p></summary> <p>A pure javascript JPEG encoder and decoder</p> <p>Library home page: <a href="https://registry.npmjs.org/jpeg-js/-/jpeg-js-0.3.5.tgz">https://registry.npmjs.org/jpeg-js/-/jpeg-js-0.3.5.tgz</a></p> <p>Path to dependency file: /uiWebView/package.json</p> <p>Path to vulnerable library: /node_modules/jpeg-js/package.json</p> <p> Dependency Hierarchy: - get-pixels-3.3.2.tgz (Root Library) - :x: **jpeg-js-0.3.5.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://api.github.com/repos/n-psk/uiWebView/commits/c2829975424625f178515c9822baef2dafbce81c">c2829975424625f178515c9822baef2dafbce81c</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The package jpeg-js before 0.4.4 are vulnerable to Denial of Service (DoS) where a particular piece of input will cause to enter an infinite loop and never return. <p>Publish Date: 2022-06-10 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-25851>CVE-2022-25851</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Release Date: 2022-06-10</p> <p>Fix Resolution: jpeg-js - 0.4.4</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2022-25851 (High) detected in jpeg-js-0.3.5.tgz - ## CVE-2022-25851 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jpeg-js-0.3.5.tgz</b></p></summary> <p>A pure javascript JPEG encoder and decoder</p> <p>Library home page: <a href="https://registry.npmjs.org/jpeg-js/-/jpeg-js-0.3.5.tgz">https://registry.npmjs.org/jpeg-js/-/jpeg-js-0.3.5.tgz</a></p> <p>Path to dependency file: /uiWebView/package.json</p> <p>Path to vulnerable library: /node_modules/jpeg-js/package.json</p> <p> Dependency Hierarchy: - get-pixels-3.3.2.tgz (Root Library) - :x: **jpeg-js-0.3.5.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://api.github.com/repos/n-psk/uiWebView/commits/c2829975424625f178515c9822baef2dafbce81c">c2829975424625f178515c9822baef2dafbce81c</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The package jpeg-js before 0.4.4 are vulnerable to Denial of Service (DoS) where a particular piece of input will cause to enter an infinite loop and never return. <p>Publish Date: 2022-06-10 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-25851>CVE-2022-25851</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Release Date: 2022-06-10</p> <p>Fix Resolution: jpeg-js - 0.4.4</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_perf
cve high detected in jpeg js tgz cve high severity vulnerability vulnerable library jpeg js tgz a pure javascript jpeg encoder and decoder library home page a href path to dependency file uiwebview package json path to vulnerable library node modules jpeg js package json dependency hierarchy get pixels tgz root library x jpeg js tgz vulnerable library found in head commit a href vulnerability details the package jpeg js before are vulnerable to denial of service dos where a particular piece of input will cause to enter an infinite loop and never return publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version release date fix resolution jpeg js step up your open source security game with mend
0
54,219
29,868,003,707
IssuesEvent
2023-06-20 06:21:21
eclipse-ee4j/jersey
https://api.github.com/repos/eclipse-ee4j/jersey
closed
Having @QueryParam without @DefaultValue throws expensive IllegalArgumentException for every missing querystring parameter
performance
When performance testing our new Jersey 2.38 application, i noticed that some endpoints were much slower than before (when using Apache Wink). I narrowed it down with Yourkit and noticed that `IllegalArgumentException` is thrown for every querystring parameter that's not present in the request querystring. When adding a `@DefaultValue`, this doesn't occur though. See relevant code here: https://github.com/eclipse-ee4j/jersey/blob/2.38/core-common/src/main/java/org/glassfish/jersey/internal/inject/ParamConverters.java#L63 Throwing exceptions to catch "regular" application flow is generally a bad habit and decreases performance of the application (https://www.baeldung.com/java-exceptions-performance). Is this intended behavior? E.g. when specifying a `@QueryParam` without `@DefaultValue`, does that automatically make it a required querystring parameter? The JAX-RS specification doesn't seem to reflect this, having an optional querystring parameter without a default value seems legit and should just return null. It becomes more noticable when you have endpoints with a lot of optional querystring parameters. More info can be found here https://stackoverflow.com/a/35625547/3032647. I would suggest that the relevant method could also just return null in this case: ``` @Override public T fromString(final String value) { if (value == null) { //throw new IllegalArgumentException(LocalizationMessages.METHOD_PARAMETER_CANNOT_BE_NULL("value")); return null; } try { return _fromString(value); } catch (final InvocationTargetException ex) { // if the value is an empty string, return null if (value.isEmpty()) { return null; } final Throwable cause = ex.getCause(); if (cause instanceof WebApplicationException) { throw (WebApplicationException) cause; } else { throw new ExtractorException(cause); } } catch (final Exception ex) { throw new ProcessingException(ex); } } ``` The same goes for querystring parameters of different types, like int. These throw a NumberFormatException.
True
Having @QueryParam without @DefaultValue throws expensive IllegalArgumentException for every missing querystring parameter - When performance testing our new Jersey 2.38 application, i noticed that some endpoints were much slower than before (when using Apache Wink). I narrowed it down with Yourkit and noticed that `IllegalArgumentException` is thrown for every querystring parameter that's not present in the request querystring. When adding a `@DefaultValue`, this doesn't occur though. See relevant code here: https://github.com/eclipse-ee4j/jersey/blob/2.38/core-common/src/main/java/org/glassfish/jersey/internal/inject/ParamConverters.java#L63 Throwing exceptions to catch "regular" application flow is generally a bad habit and decreases performance of the application (https://www.baeldung.com/java-exceptions-performance). Is this intended behavior? E.g. when specifying a `@QueryParam` without `@DefaultValue`, does that automatically make it a required querystring parameter? The JAX-RS specification doesn't seem to reflect this, having an optional querystring parameter without a default value seems legit and should just return null. It becomes more noticable when you have endpoints with a lot of optional querystring parameters. More info can be found here https://stackoverflow.com/a/35625547/3032647. I would suggest that the relevant method could also just return null in this case: ``` @Override public T fromString(final String value) { if (value == null) { //throw new IllegalArgumentException(LocalizationMessages.METHOD_PARAMETER_CANNOT_BE_NULL("value")); return null; } try { return _fromString(value); } catch (final InvocationTargetException ex) { // if the value is an empty string, return null if (value.isEmpty()) { return null; } final Throwable cause = ex.getCause(); if (cause instanceof WebApplicationException) { throw (WebApplicationException) cause; } else { throw new ExtractorException(cause); } } catch (final Exception ex) { throw new ProcessingException(ex); } } ``` The same goes for querystring parameters of different types, like int. These throw a NumberFormatException.
perf
having queryparam without defaultvalue throws expensive illegalargumentexception for every missing querystring parameter when performance testing our new jersey application i noticed that some endpoints were much slower than before when using apache wink i narrowed it down with yourkit and noticed that illegalargumentexception is thrown for every querystring parameter that s not present in the request querystring when adding a defaultvalue this doesn t occur though see relevant code here throwing exceptions to catch regular application flow is generally a bad habit and decreases performance of the application is this intended behavior e g when specifying a queryparam without defaultvalue does that automatically make it a required querystring parameter the jax rs specification doesn t seem to reflect this having an optional querystring parameter without a default value seems legit and should just return null it becomes more noticable when you have endpoints with a lot of optional querystring parameters more info can be found here i would suggest that the relevant method could also just return null in this case override public t fromstring final string value if value null throw new illegalargumentexception localizationmessages method parameter cannot be null value return null try return fromstring value catch final invocationtargetexception ex if the value is an empty string return null if value isempty return null final throwable cause ex getcause if cause instanceof webapplicationexception throw webapplicationexception cause else throw new extractorexception cause catch final exception ex throw new processingexception ex the same goes for querystring parameters of different types like int these throw a numberformatexception
1
6,545
5,514,507,125
IssuesEvent
2017-03-17 15:17:35
catapult-project/catapult
https://api.github.com/repos/catapult-project/catapult
opened
Dashboard - Investigate memcache calls in stored_object.
Hotlist:Perf Dashboard Performance Perf Dashboard
NDB is supposed to managed memcache for you, unless you explicitly change the functionality so the extra memcache calls in here seem redundant. Possibly remove them?
True
Dashboard - Investigate memcache calls in stored_object. - NDB is supposed to managed memcache for you, unless you explicitly change the functionality so the extra memcache calls in here seem redundant. Possibly remove them?
perf
dashboard investigate memcache calls in stored object ndb is supposed to managed memcache for you unless you explicitly change the functionality so the extra memcache calls in here seem redundant possibly remove them
1
403,733
27,432,327,275
IssuesEvent
2023-03-02 03:02:25
amishpatel0423/Ticket_tracker
https://api.github.com/repos/amishpatel0423/Ticket_tracker
opened
Account creation, login and logout
documentation
- Research and implement user authentication and authorization using a library in node.js. - Create routes for handling account creation and login. - Test the account creation and login functionality using Mocha - Create login sign-up pages using bootstrap.
1.0
Account creation, login and logout - - Research and implement user authentication and authorization using a library in node.js. - Create routes for handling account creation and login. - Test the account creation and login functionality using Mocha - Create login sign-up pages using bootstrap.
non_perf
account creation login and logout research and implement user authentication and authorization using a library in node js create routes for handling account creation and login test the account creation and login functionality using mocha create login sign up pages using bootstrap
0
20,162
10,616,409,247
IssuesEvent
2019-10-12 11:30:03
coq/coq
https://api.github.com/repos/coq/coq
closed
master uses 1.8x+ the memory as 8.9.1
kind: performance kind: regression part: vernac
<!-- Thank you for reporting a bug to Coq! --> #### Description of the problem Coq master seems to use almost twice as much memory as 8.9.1 when processing `Kami/CompileVerifiable.v`. This means we cannot run coq master in our CI, and thus cannot add kami to coq CI (#10306, and afaik this is the only remaining blocker for that). build [log](https://builds.sr.ht/~andres_tries_srht_github/job/83001) and [script](https://builds.sr.ht/api/jobs/83001/manifest) (failing after 2GB RAM runs out) I confirmed this locally, seeing 281178 minor page faults from 8.9.1 and 530762 from master, putting it just slightly over 2GB. ``` git clone --recursive [email protected]:sifive/RiscvSpecFormal.git cd RiscvSpecFormal make /usr/bin/time -v coqc -q -Q coq-record-update/src RecordUpdate -Q bbv/theories bbv -Q Kami Kami -Q FpuKami FpuKami -Q ProcKami ProcKami -Q StdLibKami StdLibKami Kami/CompileVerifiable.v ``` @tjmach @vmurali #### Coq Version e9c42c26d1fc653d1411fa2fe41b12bffa8ae992 vs 8.9.1
True
master uses 1.8x+ the memory as 8.9.1 - <!-- Thank you for reporting a bug to Coq! --> #### Description of the problem Coq master seems to use almost twice as much memory as 8.9.1 when processing `Kami/CompileVerifiable.v`. This means we cannot run coq master in our CI, and thus cannot add kami to coq CI (#10306, and afaik this is the only remaining blocker for that). build [log](https://builds.sr.ht/~andres_tries_srht_github/job/83001) and [script](https://builds.sr.ht/api/jobs/83001/manifest) (failing after 2GB RAM runs out) I confirmed this locally, seeing 281178 minor page faults from 8.9.1 and 530762 from master, putting it just slightly over 2GB. ``` git clone --recursive [email protected]:sifive/RiscvSpecFormal.git cd RiscvSpecFormal make /usr/bin/time -v coqc -q -Q coq-record-update/src RecordUpdate -Q bbv/theories bbv -Q Kami Kami -Q FpuKami FpuKami -Q ProcKami ProcKami -Q StdLibKami StdLibKami Kami/CompileVerifiable.v ``` @tjmach @vmurali #### Coq Version e9c42c26d1fc653d1411fa2fe41b12bffa8ae992 vs 8.9.1
perf
master uses the memory as description of the problem coq master seems to use almost twice as much memory as when processing kami compileverifiable v this means we cannot run coq master in our ci and thus cannot add kami to coq ci and afaik this is the only remaining blocker for that build and failing after ram runs out i confirmed this locally seeing minor page faults from and from master putting it just slightly over git clone recursive git github com sifive riscvspecformal git cd riscvspecformal make usr bin time v coqc q q coq record update src recordupdate q bbv theories bbv q kami kami q fpukami fpukami q prockami prockami q stdlibkami stdlibkami kami compileverifiable v tjmach vmurali coq version vs
1
54,213
29,865,514,488
IssuesEvent
2023-06-20 03:13:34
NuGet/Home
https://api.github.com/repos/NuGet/Home
opened
GetContentFileFolderRelativeToFramework allocates too much
Type:Bug Tenet:Performance
GetContentFileFolderRelativeToFramework allocates too many strings, enumerators and arrays. It can be made to allocate only the output string.
True
GetContentFileFolderRelativeToFramework allocates too much - GetContentFileFolderRelativeToFramework allocates too many strings, enumerators and arrays. It can be made to allocate only the output string.
perf
getcontentfilefolderrelativetoframework allocates too much getcontentfilefolderrelativetoframework allocates too many strings enumerators and arrays it can be made to allocate only the output string
1
23,844
16,618,506,307
IssuesEvent
2021-06-02 20:10:30
yt-project/yt
https://api.github.com/repos/yt-project/yt
closed
CI: failures on the main branch (?)
blocker infrastructure tests: running tests
### Bug report **Bug summary** Jenkins CI is apparently broken as some 44 failures are showing up on different PRs (#3234, #3236, #3270, #3295, to name a few I could identify) see for instance : PR #3295 https://tests.yt-project.org/job/yt_py38_git/3236/#showFailuresLink The errors seem to be related to incorrect field access, though it's not clear to me what change may have caused this. @Xarthisius, do you know if this could be a result of changes on the server, rather than on the main branch of the repo ?
1.0
CI: failures on the main branch (?) - ### Bug report **Bug summary** Jenkins CI is apparently broken as some 44 failures are showing up on different PRs (#3234, #3236, #3270, #3295, to name a few I could identify) see for instance : PR #3295 https://tests.yt-project.org/job/yt_py38_git/3236/#showFailuresLink The errors seem to be related to incorrect field access, though it's not clear to me what change may have caused this. @Xarthisius, do you know if this could be a result of changes on the server, rather than on the main branch of the repo ?
non_perf
ci failures on the main branch bug report bug summary jenkins ci is apparently broken as some failures are showing up on different prs to name a few i could identify see for instance pr the errors seem to be related to incorrect field access though it s not clear to me what change may have caused this xarthisius do you know if this could be a result of changes on the server rather than on the main branch of the repo
0
13,076
8,101,225,624
IssuesEvent
2018-08-12 11:06:18
phpstan/phpstan
https://api.github.com/repos/phpstan/phpstan
closed
phpstan becomes slow in 0.10.x
performance
### Summary of a problem or a feature request Exuding the file ([libraries/classes/Controllers/Server/ServerVariablesController.php](https://github.com/phpmyadmin/phpmyadmin/blob/2334119d8a6f25f40fde966fb6252ffe169d3bd6/libraries/classes/Controllers/Server/ServerVariablesController.php#L392)), the analysis takes 5 minutes. More than 45 minutes if not excluded (I had to stop the process because it is unacceptable). Using command : ```bash starttime=$(date +"%s") ; ./vendor/bin/phpstan analyse ./ --configuration=phpstan.neon --level=1 --memory-limit=1G --debug ; echo seconds=$(($(date +"%s")-$starttime)) ``` https://github.com/phpmyadmin/phpmyadmin/blob/master/libraries/classes/Relation.php took 6 min. 6x more on dev-master than with 0.10.1 https://github.com/phpmyadmin/phpmyadmin/blob/master/libraries/classes/Tracking.php hangs for 1 min and 10 seconds approximately. https://github.com/phpmyadmin/phpmyadmin/blob/2334119d8a6f25f40fde966fb6252ffe169d3bd6/libraries/classes/Controllers/Server/ServerVariablesController.php#L392 took forever ... ### Expected Run as fast as 0.9.x ### Actual Takes forever or take more than 10 minutes (TravisCI stops after 10 minutes running)
True
phpstan becomes slow in 0.10.x - ### Summary of a problem or a feature request Exuding the file ([libraries/classes/Controllers/Server/ServerVariablesController.php](https://github.com/phpmyadmin/phpmyadmin/blob/2334119d8a6f25f40fde966fb6252ffe169d3bd6/libraries/classes/Controllers/Server/ServerVariablesController.php#L392)), the analysis takes 5 minutes. More than 45 minutes if not excluded (I had to stop the process because it is unacceptable). Using command : ```bash starttime=$(date +"%s") ; ./vendor/bin/phpstan analyse ./ --configuration=phpstan.neon --level=1 --memory-limit=1G --debug ; echo seconds=$(($(date +"%s")-$starttime)) ``` https://github.com/phpmyadmin/phpmyadmin/blob/master/libraries/classes/Relation.php took 6 min. 6x more on dev-master than with 0.10.1 https://github.com/phpmyadmin/phpmyadmin/blob/master/libraries/classes/Tracking.php hangs for 1 min and 10 seconds approximately. https://github.com/phpmyadmin/phpmyadmin/blob/2334119d8a6f25f40fde966fb6252ffe169d3bd6/libraries/classes/Controllers/Server/ServerVariablesController.php#L392 took forever ... ### Expected Run as fast as 0.9.x ### Actual Takes forever or take more than 10 minutes (TravisCI stops after 10 minutes running)
perf
phpstan becomes slow in x summary of a problem or a feature request exuding the file the analysis takes minutes more than minutes if not excluded i had to stop the process because it is unacceptable using command bash starttime date s vendor bin phpstan analyse configuration phpstan neon level memory limit debug echo seconds date s starttime took min more on dev master than with hangs for min and seconds approximately took forever expected run as fast as x actual takes forever or take more than minutes travisci stops after minutes running
1
25,705
12,709,500,556
IssuesEvent
2020-06-23 12:28:07
unoplatform/uno
https://api.github.com/repos/unoplatform/uno
closed
Layouts with lots of buttons are taking too much time to load
area/wasm kind/bug kind/performance triage/needs-information
## Current behavior A page that have a great number of control takes too much to load. ## Expected behavior The page should take the same amount of time to load as in the UWP solution. ## How to reproduce it (as minimally and precisely as possible) Repo link: https://github.com/JhonH3avy/ControlLoadingPerformance.git ## Environment Nuget Package: - Uno.UI: 1.46.199-dev2445 Affected platform(s): - [ ] iOS - [ ] Android - [x] WebAssembly - [ ] Windows - [ ] Build tasks Visual Studio - [ ] 2017 (version: ) - [x] 2019 (version:16.2.1) - [ ] for Mac (version: ) Relevant plugins - [x] Resharper (version:183.0) ## Anything else we need to know? The repo has four pages that have different amounts of controls, the more controls the page have the time to load the page grows apparently linearly.
True
Layouts with lots of buttons are taking too much time to load - ## Current behavior A page that have a great number of control takes too much to load. ## Expected behavior The page should take the same amount of time to load as in the UWP solution. ## How to reproduce it (as minimally and precisely as possible) Repo link: https://github.com/JhonH3avy/ControlLoadingPerformance.git ## Environment Nuget Package: - Uno.UI: 1.46.199-dev2445 Affected platform(s): - [ ] iOS - [ ] Android - [x] WebAssembly - [ ] Windows - [ ] Build tasks Visual Studio - [ ] 2017 (version: ) - [x] 2019 (version:16.2.1) - [ ] for Mac (version: ) Relevant plugins - [x] Resharper (version:183.0) ## Anything else we need to know? The repo has four pages that have different amounts of controls, the more controls the page have the time to load the page grows apparently linearly.
perf
layouts with lots of buttons are taking too much time to load current behavior a page that have a great number of control takes too much to load expected behavior the page should take the same amount of time to load as in the uwp solution how to reproduce it as minimally and precisely as possible repo link environment nuget package uno ui affected platform s ios android webassembly windows build tasks visual studio version version for mac version relevant plugins resharper version anything else we need to know the repo has four pages that have different amounts of controls the more controls the page have the time to load the page grows apparently linearly
1
36,014
17,391,977,748
IssuesEvent
2021-08-02 08:35:55
NVIDIA/TensorRT
https://api.github.com/repos/NVIDIA/TensorRT
closed
converted bart model is slower than the original one during inference time
Component: ONNX Runtime: Performance triaged
hi there, I have a project to use facebook bart for news summerization. In order to make the inference faster, we are trying to convert part of the model to tensorrt and then interegerated into the original model. Via this repo, I have successfully converted facebook bart decoder layers to tensorrt model, and successfully integerated, however, the total inference time of generated tokens of the new bart model(i.e. the model integerated with converted tensorrt decoder layer) is 2 times slower than the original one, so, I tried to find why, and finally I found that the new bart model itself is faster than the original one, see code below, line1 is faster than before after changing with new bart model, but is became much slower after line2, line1: outputs = self(model_inputs, return_dict=True) line2: next_token_logits = outputs.logits[:, -1, :] line3: next_token_logits = self.adjust_logits_during_generation( line4: next_token_logits, cur_len=cur_len, max_length=max_length) below you can find the comparing speed of new bart model and original one (corresponding to comparing results of code line1 above), <img width="292" alt="1" src="https://user-images.githubusercontent.com/13851442/106444599-edb8bf80-64b8-11eb-8a55-a76e342b1447.PNG"> below you can find the comparing speed of new bart model and original one(corresponding to comparing results of code after line2 above) <img width="353" alt="2" src="https://user-images.githubusercontent.com/13851442/106444940-615acc80-64b9-11eb-914b-fc6df0f849ee.PNG"> Does anyone knows why it became slow after line1 code above?
True
converted bart model is slower than the original one during inference time - hi there, I have a project to use facebook bart for news summerization. In order to make the inference faster, we are trying to convert part of the model to tensorrt and then interegerated into the original model. Via this repo, I have successfully converted facebook bart decoder layers to tensorrt model, and successfully integerated, however, the total inference time of generated tokens of the new bart model(i.e. the model integerated with converted tensorrt decoder layer) is 2 times slower than the original one, so, I tried to find why, and finally I found that the new bart model itself is faster than the original one, see code below, line1 is faster than before after changing with new bart model, but is became much slower after line2, line1: outputs = self(model_inputs, return_dict=True) line2: next_token_logits = outputs.logits[:, -1, :] line3: next_token_logits = self.adjust_logits_during_generation( line4: next_token_logits, cur_len=cur_len, max_length=max_length) below you can find the comparing speed of new bart model and original one (corresponding to comparing results of code line1 above), <img width="292" alt="1" src="https://user-images.githubusercontent.com/13851442/106444599-edb8bf80-64b8-11eb-8a55-a76e342b1447.PNG"> below you can find the comparing speed of new bart model and original one(corresponding to comparing results of code after line2 above) <img width="353" alt="2" src="https://user-images.githubusercontent.com/13851442/106444940-615acc80-64b9-11eb-914b-fc6df0f849ee.PNG"> Does anyone knows why it became slow after line1 code above?
perf
converted bart model is slower than the original one during inference time hi there i have a project to use facebook bart for news summerization in order to make the inference faster we are trying to convert part of the model to tensorrt and then interegerated into the original model via this repo i have successfully converted facebook bart decoder layers to tensorrt model and successfully integerated however the total inference time of generated tokens of the new bart model i e the model integerated with converted tensorrt decoder layer is times slower than the original one so i tried to find why and finally i found that the new bart model itself is faster than the original one see code below is faster than before after changing with new bart model but is became much slower after outputs self model inputs return dict true next token logits outputs logits next token logits self adjust logits during generation next token logits cur len cur len max length max length below you can find the comparing speed of new bart model and original one corresponding to comparing results of code above img width alt src below you can find the comparing speed of new bart model and original one corresponding to comparing results of code after above img width alt src does anyone knows why it became slow after code above
1
48,249
25,451,172,388
IssuesEvent
2022-11-24 10:33:42
topling/toplingdb
https://api.github.com/repos/topling/toplingdb
closed
FindFileInRange: devirtualize comparator and add prefix cache
performance
## This was a PR to upstream https://github.com/facebook/rocksdb/pull/10646 ## Copied from https://github.com/facebook/rocksdb/pull/10646 This PR is based on https://github.com/facebook/rocksdb/pull/10645. If comparator is BytewiseComparator or ReverseBytewiseComparator: devirtualize comparator: specialize the impl by direct call memcmp add prefix cache: narrow the search range by prefix cache, then find by comparator ## Relevant commits 3304b5b5d586f3abbd80a65eed700b543c461c2c 0f98a93ebb4c968b56fbc300cd1bae5d35bc54ce 9ae79bdea0d6af9c63126a53441167aee76b2301 271a43d6a255bec75fbe0973bfa87373ee165be7 0ae017a2c491b6093c23a8ea9d3e09c51e33e169
True
FindFileInRange: devirtualize comparator and add prefix cache - ## This was a PR to upstream https://github.com/facebook/rocksdb/pull/10646 ## Copied from https://github.com/facebook/rocksdb/pull/10646 This PR is based on https://github.com/facebook/rocksdb/pull/10645. If comparator is BytewiseComparator or ReverseBytewiseComparator: devirtualize comparator: specialize the impl by direct call memcmp add prefix cache: narrow the search range by prefix cache, then find by comparator ## Relevant commits 3304b5b5d586f3abbd80a65eed700b543c461c2c 0f98a93ebb4c968b56fbc300cd1bae5d35bc54ce 9ae79bdea0d6af9c63126a53441167aee76b2301 271a43d6a255bec75fbe0973bfa87373ee165be7 0ae017a2c491b6093c23a8ea9d3e09c51e33e169
perf
findfileinrange devirtualize comparator and add prefix cache this was a pr to upstream copied from this pr is based on if comparator is bytewisecomparator or reversebytewisecomparator devirtualize comparator specialize the impl by direct call memcmp add prefix cache narrow the search range by prefix cache then find by comparator relevant commits
1
41,349
21,647,405,864
IssuesEvent
2022-05-06 04:51:45
JuliaData/TypedTables.jl
https://api.github.com/repos/JuliaData/TypedTables.jl
closed
How to deal with latency with large number of columns?
performance
Currently I have a column type that is lazy. It represents ~GB of stuff that needs to be read and decompressed on the fly and cached (by chunk). Turns out I can construct `Table` nicely and the laziness works. However, sometimes we have 1000+ columns, in this case the compiler struggles a lot. Is it possible to have a less-typed but same interfaced `Table`?
True
How to deal with latency with large number of columns? - Currently I have a column type that is lazy. It represents ~GB of stuff that needs to be read and decompressed on the fly and cached (by chunk). Turns out I can construct `Table` nicely and the laziness works. However, sometimes we have 1000+ columns, in this case the compiler struggles a lot. Is it possible to have a less-typed but same interfaced `Table`?
perf
how to deal with latency with large number of columns currently i have a column type that is lazy it represents gb of stuff that needs to be read and decompressed on the fly and cached by chunk turns out i can construct table nicely and the laziness works however sometimes we have columns in this case the compiler struggles a lot is it possible to have a less typed but same interfaced table
1
18,051
9,986,346,257
IssuesEvent
2019-07-10 18:52:05
tensorflow/tensorflow
https://api.github.com/repos/tensorflow/tensorflow
closed
TF2 - apparent memory leak when running dataset ops eagerly
2.0.0-beta0 comp:data type:bug/performance
**System information** - Have I written custom code: yes - OS Platform and Distribution (e.g., Linux Ubuntu 16.04): OSX - TensorFlow installed from (source or binary): 2.0.0beta - TensorFlow version (use command below): v1.12.1-3259-gf59745a381 2.0.0-beta0 - Python version: 3.6.8 **Describe the current behavior** When using the function `tf.autograph.to_graph`, I see a memory leak which I don't see if I use the annotation `@tf.function` **Describe the expected behavior** There should not be a memory leak. **Code to reproduce the issue** ```python import os import psutil import numpy as np import tensorflow as tf process = psutil.Process(os.getpid()) # @tf.function def train_epoch(model, p_data): for real_inputs in p_data: model * real_inputs train_epoch = tf.autograph.to_graph(train_epoch) data = np.random.normal(0., 1., [10000, 2]) p_data = tf.data.Dataset.from_tensor_slices(data).batch(32) model = tf.Variable([1., 1.], dtype=tf.float64) for i in range(5000): train_epoch(model, p_data) if i % 50 == 0: print(process.memory_info().rss) ```
True
TF2 - apparent memory leak when running dataset ops eagerly - **System information** - Have I written custom code: yes - OS Platform and Distribution (e.g., Linux Ubuntu 16.04): OSX - TensorFlow installed from (source or binary): 2.0.0beta - TensorFlow version (use command below): v1.12.1-3259-gf59745a381 2.0.0-beta0 - Python version: 3.6.8 **Describe the current behavior** When using the function `tf.autograph.to_graph`, I see a memory leak which I don't see if I use the annotation `@tf.function` **Describe the expected behavior** There should not be a memory leak. **Code to reproduce the issue** ```python import os import psutil import numpy as np import tensorflow as tf process = psutil.Process(os.getpid()) # @tf.function def train_epoch(model, p_data): for real_inputs in p_data: model * real_inputs train_epoch = tf.autograph.to_graph(train_epoch) data = np.random.normal(0., 1., [10000, 2]) p_data = tf.data.Dataset.from_tensor_slices(data).batch(32) model = tf.Variable([1., 1.], dtype=tf.float64) for i in range(5000): train_epoch(model, p_data) if i % 50 == 0: print(process.memory_info().rss) ```
perf
apparent memory leak when running dataset ops eagerly system information have i written custom code yes os platform and distribution e g linux ubuntu osx tensorflow installed from source or binary tensorflow version use command below python version describe the current behavior when using the function tf autograph to graph i see a memory leak which i don t see if i use the annotation tf function describe the expected behavior there should not be a memory leak code to reproduce the issue python import os import psutil import numpy as np import tensorflow as tf process psutil process os getpid tf function def train epoch model p data for real inputs in p data model real inputs train epoch tf autograph to graph train epoch data np random normal p data tf data dataset from tensor slices data batch model tf variable dtype tf for i in range train epoch model p data if i print process memory info rss
1
18,575
13,046,744,677
IssuesEvent
2020-07-29 09:30:21
OpenRA/OpenRA
https://api.github.com/repos/OpenRA/OpenRA
closed
macOS: suppress shortcut cmd+Q
Idea/Wishlist OS: MacOS X Usability
The default key combination for terminating applications under macOS is cmd+Q. In OpenRA, the key combination cmd+<number key> combines units into a group. This can lead to a premature end of the game. In the hotkey settings the key combination for cmd+Q can be assigned, but it will always terminate the application. The key combination should be intercepted to prevent the application from being terminated without asking. Alternatively, a dialog should be displayed asking if you really want to quit the application.
True
macOS: suppress shortcut cmd+Q - The default key combination for terminating applications under macOS is cmd+Q. In OpenRA, the key combination cmd+<number key> combines units into a group. This can lead to a premature end of the game. In the hotkey settings the key combination for cmd+Q can be assigned, but it will always terminate the application. The key combination should be intercepted to prevent the application from being terminated without asking. Alternatively, a dialog should be displayed asking if you really want to quit the application.
non_perf
macos suppress shortcut cmd q the default key combination for terminating applications under macos is cmd q in openra the key combination cmd combines units into a group this can lead to a premature end of the game in the hotkey settings the key combination for cmd q can be assigned but it will always terminate the application the key combination should be intercepted to prevent the application from being terminated without asking alternatively a dialog should be displayed asking if you really want to quit the application
0
77,195
3,506,270,780
IssuesEvent
2016-01-08 05:10:03
OregonCore/OregonCore
https://api.github.com/repos/OregonCore/OregonCore
closed
Talent's Bug Important ! (BB #257)
duplicate migrated Priority: Medium Type: Bug
This issue was migrated from bitbucket. **Original Reporter:** **Original Date:** 07.08.2010 05:12:57 GMT+0000 **Original Priority:** major **Original Type:** bug **Original State:** duplicate **Direct Link:** https://bitbucket.org/oregon/oregoncore/issues/257 <hr> when Get talent point for example 0/3, 1/3 , 2/3 , 3/3 here if Again hit the talent buttom Can Get 1/3 , 2/3 , 3/3 This is a big Problem it is Not God For this project !
1.0
Talent's Bug Important ! (BB #257) - This issue was migrated from bitbucket. **Original Reporter:** **Original Date:** 07.08.2010 05:12:57 GMT+0000 **Original Priority:** major **Original Type:** bug **Original State:** duplicate **Direct Link:** https://bitbucket.org/oregon/oregoncore/issues/257 <hr> when Get talent point for example 0/3, 1/3 , 2/3 , 3/3 here if Again hit the talent buttom Can Get 1/3 , 2/3 , 3/3 This is a big Problem it is Not God For this project !
non_perf
talent s bug important bb this issue was migrated from bitbucket original reporter original date gmt original priority major original type bug original state duplicate direct link when get talent point for example here if again hit the talent buttom can get this is a big problem it is not god for this project
0
102,656
12,814,349,540
IssuesEvent
2020-07-04 18:13:13
magento/adobe-stock-integration
https://api.github.com/repos/magento/adobe-stock-integration
closed
[Spike] Investigate meta data extracting
Priority: P1 Progress: PR created requires technical design
As a Merchant I want Magento Media Storage to parse image meta data that is available in the file and expose it on image details page so that I can manage it and use for filtering images **Additional context** https://github.com/magento/adobe-stock-integration/issues/1183 # Open questions: - is there any required set of fields that exist in image meta data? Like title, description, keywords. Or any fields can be removed/added to the file according to standard? - is there a way in a meta data file to define what fields can be edited by user in the application? e.g. Camera data, image format, creation and modification date can not be edited
1.0
[Spike] Investigate meta data extracting - As a Merchant I want Magento Media Storage to parse image meta data that is available in the file and expose it on image details page so that I can manage it and use for filtering images **Additional context** https://github.com/magento/adobe-stock-integration/issues/1183 # Open questions: - is there any required set of fields that exist in image meta data? Like title, description, keywords. Or any fields can be removed/added to the file according to standard? - is there a way in a meta data file to define what fields can be edited by user in the application? e.g. Camera data, image format, creation and modification date can not be edited
non_perf
investigate meta data extracting as a merchant i want magento media storage to parse image meta data that is available in the file and expose it on image details page so that i can manage it and use for filtering images additional context open questions is there any required set of fields that exist in image meta data like title description keywords or any fields can be removed added to the file according to standard is there a way in a meta data file to define what fields can be edited by user in the application e g camera data image format creation and modification date can not be edited
0
6,579
5,533,877,465
IssuesEvent
2017-03-21 14:19:49
wojtpl2/ExtendedXmlSerializer
https://api.github.com/repos/wojtpl2/ExtendedXmlSerializer
opened
Performance Issues
help wanted performance
Well, double whammy. While implementing the feature for #56, I found out two issues. 1. We have a performance creep between 4f015846a5fb899eb5865e432b095722b273d0c1 and the current commit (b61e39944009dc07fa9124e7de7ad2df6179ae92). Here is the performance from the first: ``` Method | Mean | StdDev | ---------------------------------- |----------- |---------- | SerializationClassWithPrimitive | 34.9980 us | 0.0149 us | DeserializationClassWithPrimitive | 44.8417 us | 0.0299 us | ``` And here it is currently: ``` Method | Mean | StdDev | ---------------------------------- |----------- |---------- | SerializationClassWithPrimitive | 37.0960 us | 0.0456 us | DeserializationClassWithPrimitive | 48.5083 us | 0.0949 us | ``` I took some time to see where it could be taking place, but could not see anything obvious. I did find one area that I fixed, but it is still too slow. To be honest, I am a little burnt out on fixing the performance, so I am definitely open to any assistance here. It has easily consumed 40% of my time on this project, if not more. Secondly -- and probably worse -- it appears that the way in which we were testing the original `XmlSerializer` was not accurate. I have updated the tests so [that they are doing the same thing](https://github.com/wojtpl2/ExtendedXmlSerializer/blob/v2.0.0/test/ExtendedXmlSerializer.Tests.Performance/Benchmarks.cs#L112-L151), and here is the new results: ``` Method | Mean | StdDev | ---------------------------------- |----------- |---------- | SerializationClassWithPrimitive | 40.8919 us | 0.2061 us | DeserializationClassWithPrimitive | 57.7255 us | 0.0529 us | ``` This is from ~62/60 on my machine. So, a considerable jump. Just so you know, I start a new work project on April 3rd, so I will not be able to help out here much after that. I hope to have all the outstanding issues complete by then. Although I am not so sure about this one. If you want to help out and look at this issue, please feel free to do so.
True
Performance Issues - Well, double whammy. While implementing the feature for #56, I found out two issues. 1. We have a performance creep between 4f015846a5fb899eb5865e432b095722b273d0c1 and the current commit (b61e39944009dc07fa9124e7de7ad2df6179ae92). Here is the performance from the first: ``` Method | Mean | StdDev | ---------------------------------- |----------- |---------- | SerializationClassWithPrimitive | 34.9980 us | 0.0149 us | DeserializationClassWithPrimitive | 44.8417 us | 0.0299 us | ``` And here it is currently: ``` Method | Mean | StdDev | ---------------------------------- |----------- |---------- | SerializationClassWithPrimitive | 37.0960 us | 0.0456 us | DeserializationClassWithPrimitive | 48.5083 us | 0.0949 us | ``` I took some time to see where it could be taking place, but could not see anything obvious. I did find one area that I fixed, but it is still too slow. To be honest, I am a little burnt out on fixing the performance, so I am definitely open to any assistance here. It has easily consumed 40% of my time on this project, if not more. Secondly -- and probably worse -- it appears that the way in which we were testing the original `XmlSerializer` was not accurate. I have updated the tests so [that they are doing the same thing](https://github.com/wojtpl2/ExtendedXmlSerializer/blob/v2.0.0/test/ExtendedXmlSerializer.Tests.Performance/Benchmarks.cs#L112-L151), and here is the new results: ``` Method | Mean | StdDev | ---------------------------------- |----------- |---------- | SerializationClassWithPrimitive | 40.8919 us | 0.2061 us | DeserializationClassWithPrimitive | 57.7255 us | 0.0529 us | ``` This is from ~62/60 on my machine. So, a considerable jump. Just so you know, I start a new work project on April 3rd, so I will not be able to help out here much after that. I hope to have all the outstanding issues complete by then. Although I am not so sure about this one. If you want to help out and look at this issue, please feel free to do so.
perf
performance issues well double whammy while implementing the feature for i found out two issues we have a performance creep between and the current commit here is the performance from the first method mean stddev serializationclasswithprimitive us us deserializationclasswithprimitive us us and here it is currently method mean stddev serializationclasswithprimitive us us deserializationclasswithprimitive us us i took some time to see where it could be taking place but could not see anything obvious i did find one area that i fixed but it is still too slow to be honest i am a little burnt out on fixing the performance so i am definitely open to any assistance here it has easily consumed of my time on this project if not more secondly and probably worse it appears that the way in which we were testing the original xmlserializer was not accurate i have updated the tests so and here is the new results method mean stddev serializationclasswithprimitive us us deserializationclasswithprimitive us us this is from on my machine so a considerable jump just so you know i start a new work project on april so i will not be able to help out here much after that i hope to have all the outstanding issues complete by then although i am not so sure about this one if you want to help out and look at this issue please feel free to do so
1
44,315
23,551,478,880
IssuesEvent
2022-08-21 22:10:03
neovim/neovim
https://api.github.com/repos/neovim/neovim
closed
Expanding wildcards in directories with several self-referential links is very, very slow
performance bug-vim filesystem
### Neovim version (nvim -v) 0.6.1 ### Vim (not Nvim) behaves the same? yes, 8.2.3995 ### Operating system/version PoP OS 22.04 ### Terminal name/version GNOME Terminal 3.44.0 using VTE 0.68.0 +BIDI +GNUTLS +ICU +SYSTEMD ### $TERM environment variable xterm-256color ### Installation apt ### How to reproduce the issue 1. In a UNIX shell, create a directory with several self-referential links: ``` $ mkdir foo $ cd foo $ ln -s ./ link1 $ ln -s ./ link2 $ ln -s ./ link3 $ ln -s ./ link4 $ ln -s ./ link5 $ ls link1 link2 link3 link4 link5 ``` 2. Open `nvim` in that directory: ``` $ vim --clean ``` 3. Expand `**`: ```vim :echom expand('**') ``` 4. Press enter ### Expected behavior Neovim should expand the wildcard within a reasonable amount of time. ### Actual behavior Neovim attempts to expand `**` for a very, very long time (I haven't let it run long enough to finish).
True
Expanding wildcards in directories with several self-referential links is very, very slow - ### Neovim version (nvim -v) 0.6.1 ### Vim (not Nvim) behaves the same? yes, 8.2.3995 ### Operating system/version PoP OS 22.04 ### Terminal name/version GNOME Terminal 3.44.0 using VTE 0.68.0 +BIDI +GNUTLS +ICU +SYSTEMD ### $TERM environment variable xterm-256color ### Installation apt ### How to reproduce the issue 1. In a UNIX shell, create a directory with several self-referential links: ``` $ mkdir foo $ cd foo $ ln -s ./ link1 $ ln -s ./ link2 $ ln -s ./ link3 $ ln -s ./ link4 $ ln -s ./ link5 $ ls link1 link2 link3 link4 link5 ``` 2. Open `nvim` in that directory: ``` $ vim --clean ``` 3. Expand `**`: ```vim :echom expand('**') ``` 4. Press enter ### Expected behavior Neovim should expand the wildcard within a reasonable amount of time. ### Actual behavior Neovim attempts to expand `**` for a very, very long time (I haven't let it run long enough to finish).
perf
expanding wildcards in directories with several self referential links is very very slow neovim version nvim v vim not nvim behaves the same yes operating system version pop os terminal name version gnome terminal using vte bidi gnutls icu systemd term environment variable xterm installation apt how to reproduce the issue in a unix shell create a directory with several self referential links mkdir foo cd foo ln s ln s ln s ln s ln s ls open nvim in that directory vim clean expand vim echom expand press enter expected behavior neovim should expand the wildcard within a reasonable amount of time actual behavior neovim attempts to expand for a very very long time i haven t let it run long enough to finish
1
31,545
14,988,488,580
IssuesEvent
2021-01-29 01:22:33
GaloisInc/crucible
https://api.github.com/repos/GaloisInc/crucible
opened
crux-mir performance regression
MIR crux performance
`crux-mir`'s `symb_eval/scalar/test1.rs` test case has gotten slower: on my machine, it takes around 2m30s to run in commit 9bb4b78b0412698046837d8057b99adf6ab81459 (when `crux-mir` was merged into this repository), and now takes around 3m30s as of 09532645133d39cda983bca67c2dbb91d05e4373 (current `master`). `git bisect` blames ced74d4a5e4d7eb01fdb9f3d6ada382a365fc22f "Add a configuration option that controls if the online backend methods for maintaining a connection with a solver are enabled" from #570. On ced74d4a5e4d7eb01fdb9f3d6ada382a365fc22f the test runs in about 3m20s, while on 12a06e3d24d25b2016b7491d7c282dcfd634380b (its immediate ancestor) it runs in 2m25s. I've been timing the test using the following command: ```sh time cabal v2-run -- crux-mir --assert-false-on-error -s z3 test/symb_eval/scalar/test1.rs ``` (After building first, so the `time` doesn't include the time spent building) Note that the two commits in question (ced74d4a5e4d7eb01fdb9f3d6ada382a365fc22f and 12a06e3d24d25b2016b7491d7c282dcfd634380b) don't build as-is, due to changes on Hackage. Both require changing `what4.cabal`'s bound on the `versions` dependency from `versions >= 3.5.2` to `versions >= 3.5.2 && < 4.0` (otherwise you'll get a type error about `Versions.VChunk` and `Versions.VUnit`). Furthermore, ced74d4a5e4d7eb01fdb9f3d6ada382a365fc22f requires the `crux-mir` build fix from d18fd5074433dad46c9917b159b8b9aaf6d8da24 - I applied it by running `git checkout d18fd50 -- crux-mir` in the top-level `crucible` directory.
True
crux-mir performance regression - `crux-mir`'s `symb_eval/scalar/test1.rs` test case has gotten slower: on my machine, it takes around 2m30s to run in commit 9bb4b78b0412698046837d8057b99adf6ab81459 (when `crux-mir` was merged into this repository), and now takes around 3m30s as of 09532645133d39cda983bca67c2dbb91d05e4373 (current `master`). `git bisect` blames ced74d4a5e4d7eb01fdb9f3d6ada382a365fc22f "Add a configuration option that controls if the online backend methods for maintaining a connection with a solver are enabled" from #570. On ced74d4a5e4d7eb01fdb9f3d6ada382a365fc22f the test runs in about 3m20s, while on 12a06e3d24d25b2016b7491d7c282dcfd634380b (its immediate ancestor) it runs in 2m25s. I've been timing the test using the following command: ```sh time cabal v2-run -- crux-mir --assert-false-on-error -s z3 test/symb_eval/scalar/test1.rs ``` (After building first, so the `time` doesn't include the time spent building) Note that the two commits in question (ced74d4a5e4d7eb01fdb9f3d6ada382a365fc22f and 12a06e3d24d25b2016b7491d7c282dcfd634380b) don't build as-is, due to changes on Hackage. Both require changing `what4.cabal`'s bound on the `versions` dependency from `versions >= 3.5.2` to `versions >= 3.5.2 && < 4.0` (otherwise you'll get a type error about `Versions.VChunk` and `Versions.VUnit`). Furthermore, ced74d4a5e4d7eb01fdb9f3d6ada382a365fc22f requires the `crux-mir` build fix from d18fd5074433dad46c9917b159b8b9aaf6d8da24 - I applied it by running `git checkout d18fd50 -- crux-mir` in the top-level `crucible` directory.
perf
crux mir performance regression crux mir s symb eval scalar rs test case has gotten slower on my machine it takes around to run in commit when crux mir was merged into this repository and now takes around as of current master git bisect blames add a configuration option that controls if the online backend methods for maintaining a connection with a solver are enabled from on the test runs in about while on its immediate ancestor it runs in i ve been timing the test using the following command sh time cabal run crux mir assert false on error s test symb eval scalar rs after building first so the time doesn t include the time spent building note that the two commits in question and don t build as is due to changes on hackage both require changing cabal s bound on the versions dependency from versions to versions otherwise you ll get a type error about versions vchunk and versions vunit furthermore requires the crux mir build fix from i applied it by running git checkout crux mir in the top level crucible directory
1
29,396
14,108,654,560
IssuesEvent
2020-11-06 18:11:06
nvm-sh/nvm
https://api.github.com/repos/nvm-sh/nvm
closed
init-nvm.sh - slow (mostly while executing "npm")
performance
<!-- Thank you for being interested in nvm! Please help us by filling out the following form if you‘re having trouble. If you have a feature request, or some other question, please feel free to clear out the form. Thanks! --> #### Operating system and version: #### `nvm debug` output: <details> <!-- do not delete the following blank line --> ```sh nvm --version: v0.35.0 $SHELL: /bin/bash $SHLVL: 1 ${HOME}: /home/kolorafa ${NVM_DIR}: '${HOME}/.nvm' ${PATH}: ${NVM_DIR}/versions/node/v8.17.0/bin:/usr/local/sbin:/usr/local/bin:/usr/bin:/opt/android-sdk/tools:/opt/android-sdk/tools/bin:/opt/COMODO:/opt/cuda/bin:/var/lib/flatpak/exports/bin:/usr/lib/jvm/default/bin:/usr/lib32/jvm/default/bin:/usr/bin/site_perl:/usr/bin/vendor_perl:/usr/bin/core_perl:/var/lib/snapd/snap/bin:/opt/xtensa-lx106-elf-gcc/bin:${HOME}/.cargo/bin:${HOME}/.local/bin $PREFIX: '' ${NPM_CONFIG_PREFIX}: '' $NVM_NODEJS_ORG_MIRROR: '' $NVM_IOJS_ORG_MIRROR: '' shell version: 'GNU bash, wersja 5.0.17(1)-release (x86_64-pc-linux-gnu)' uname -a: 'Linux 5.7.2-arch1-1 #1 SMP PREEMPT Wed, 10 Jun 2020 20:36:24 +0000 x86_64 GNU/Linux' OS version: Antergos Linux () curl: curl jest /usr/bin/curl, curl 7.70.0 (x86_64-pc-linux-gnu) libcurl/7.70.0 OpenSSL/1.1.1g zlib/1.2.11 libidn2/2.3.0 libpsl/0.21.0 (+libidn2/2.2.0) libssh2/1.9.0 nghttp2/1.41.0 wget: wget jest /usr/bin/wget, GNU Wget 1.20.3 zbudowany na systemie linux-gnu. ls: nie ma dostępu do 'git': Nie ma takiego pliku ani katalogu git: git jest /usr/bin/git, git version 2.27.0 ls: nie ma dostępu do 'grep': Nie ma takiego pliku ani katalogu grep: grep jest aliasem do grep --colour=auto', grep (GNU grep) 3.4 ls: nie ma dostępu do 'awk': Nie ma takiego pliku ani katalogu awk: awk jest /usr/bin/awk, GNU Awk 5.1.0, API: 3.0 (GNU MPFR 4.0.2, GNU MP 6.2.0) ls: nie ma dostępu do 'sed': Nie ma takiego pliku ani katalogu sed: sed jest /usr/bin/sed, sed (GNU sed) 4.8 ls: nie ma dostępu do 'cut': Nie ma takiego pliku ani katalogu cut: cut jest /usr/bin/cut, cut (GNU coreutils) 8.32 ls: nie ma dostępu do 'basename': Nie ma takiego pliku ani katalogu basename: basename jest /usr/bin/basename, basename (GNU coreutils) 8.32 ls: nie ma dostępu do 'rm': Nie ma takiego pliku ani katalogu rm: rm jest /usr/bin/rm, rm (GNU coreutils) 8.32 ls: nie ma dostępu do 'mkdir': Nie ma takiego pliku ani katalogu mkdir: mkdir jest /usr/bin/mkdir, mkdir (GNU coreutils) 8.32 ls: nie ma dostępu do 'xargs': Nie ma takiego pliku ani katalogu xargs: xargs jest /usr/bin/xargs, xargs (GNU findutils) 4.7.0 nvm current: v8.17.0 which node: ${NVM_DIR}/versions/node/v8.17.0/bin/node which iojs: which: no iojs in (${NVM_DIR}/versions/node/v8.17.0/bin:/usr/local/sbin:/usr/local/bin:/usr/bin:/opt/android-sdk/tools:/opt/android-sdk/tools/bin:/opt/COMODO:/opt/cuda/bin:/var/lib/flatpak/exports/bin:/usr/lib/jvm/default/bin:/usr/lib32/jvm/default/bin:/usr/bin/site_perl:/usr/bin/vendor_perl:/usr/bin/core_perl:/var/lib/snapd/snap/bin:/opt/xtensa-lx106-elf-gcc/bin:${HOME}/.cargo/bin:${HOME}/.local/bin) which npm: ${NVM_DIR}/versions/node/v8.17.0/bin/npm npm config get prefix: ${NVM_DIR}/versions/node/v8.17.0 npm root -g: ${NVM_DIR}/versions/node/v8.17.0/lib/node_modules ``` </details> #### `nvm ls` output: <details> <!-- do not delete the following blank line --> ```sh -> v8.17.0 v12.16.1 system default -> 8 (-> v8.17.0) node -> stable (-> v12.16.1) (default) stable -> 12.16 (-> v12.16.1) (default) iojs -> N/A (default) unstable -> N/A (default) lts/* -> lts/erbium (-> v12.16.1) lts/argon -> v4.9.1 (-> N/A) lts/boron -> v6.17.1 (-> N/A) lts/carbon -> v8.17.0 lts/dubnium -> v10.19.0 (-> N/A) lts/erbium -> v12.16.1 ``` </details> #### How did you install `nvm`? Arch - nvm AUR #### What steps did you perform? Open new gnome terminal by shortcut #### What happened? Terminal opened but bash didn't show for 20s (when HD was slowed down by copy) While HD moderatly used - bash show in 1-2s ( https://youtu.be/Ie4Nnml55-g ) While HD was very busy and terminal load in 20s, i used different already open terminal and while doing ps i found that it's stuck at "npm" command: ``` kolorafa 3024410 0.0 0.0 15844 9960 pts/9 Ss 08:07 0:00 | \_ bash kolorafa 3026045 0.0 0.0 13008 4632 pts/9 R+ 08:09 0:00 | | \_ ps auxf kolorafa 3025222 0.0 0.0 14364 8504 pts/11 Ss+ 08:08 0:00 | \_ bash kolorafa 3025440 0.4 0.1 1036508 39904 pts/11 Dl+ 08:08 0:00 | | \_ npm kolorafa 3025697 0.1 0.0 14364 8348 pts/13 Ss+ 08:09 0:00 | \_ bash kolorafa 3025914 0.8 0.1 1036512 40236 pts/13 Dl+ 08:09 0:00 | \_ npm ``` #### What did you expect to happen? NVM loads without slowing down anything ;) #### Is there anything in any of your profile files that modifies the `PATH`? Don't think anything related
True
init-nvm.sh - slow (mostly while executing "npm") - <!-- Thank you for being interested in nvm! Please help us by filling out the following form if you‘re having trouble. If you have a feature request, or some other question, please feel free to clear out the form. Thanks! --> #### Operating system and version: #### `nvm debug` output: <details> <!-- do not delete the following blank line --> ```sh nvm --version: v0.35.0 $SHELL: /bin/bash $SHLVL: 1 ${HOME}: /home/kolorafa ${NVM_DIR}: '${HOME}/.nvm' ${PATH}: ${NVM_DIR}/versions/node/v8.17.0/bin:/usr/local/sbin:/usr/local/bin:/usr/bin:/opt/android-sdk/tools:/opt/android-sdk/tools/bin:/opt/COMODO:/opt/cuda/bin:/var/lib/flatpak/exports/bin:/usr/lib/jvm/default/bin:/usr/lib32/jvm/default/bin:/usr/bin/site_perl:/usr/bin/vendor_perl:/usr/bin/core_perl:/var/lib/snapd/snap/bin:/opt/xtensa-lx106-elf-gcc/bin:${HOME}/.cargo/bin:${HOME}/.local/bin $PREFIX: '' ${NPM_CONFIG_PREFIX}: '' $NVM_NODEJS_ORG_MIRROR: '' $NVM_IOJS_ORG_MIRROR: '' shell version: 'GNU bash, wersja 5.0.17(1)-release (x86_64-pc-linux-gnu)' uname -a: 'Linux 5.7.2-arch1-1 #1 SMP PREEMPT Wed, 10 Jun 2020 20:36:24 +0000 x86_64 GNU/Linux' OS version: Antergos Linux () curl: curl jest /usr/bin/curl, curl 7.70.0 (x86_64-pc-linux-gnu) libcurl/7.70.0 OpenSSL/1.1.1g zlib/1.2.11 libidn2/2.3.0 libpsl/0.21.0 (+libidn2/2.2.0) libssh2/1.9.0 nghttp2/1.41.0 wget: wget jest /usr/bin/wget, GNU Wget 1.20.3 zbudowany na systemie linux-gnu. ls: nie ma dostępu do 'git': Nie ma takiego pliku ani katalogu git: git jest /usr/bin/git, git version 2.27.0 ls: nie ma dostępu do 'grep': Nie ma takiego pliku ani katalogu grep: grep jest aliasem do grep --colour=auto', grep (GNU grep) 3.4 ls: nie ma dostępu do 'awk': Nie ma takiego pliku ani katalogu awk: awk jest /usr/bin/awk, GNU Awk 5.1.0, API: 3.0 (GNU MPFR 4.0.2, GNU MP 6.2.0) ls: nie ma dostępu do 'sed': Nie ma takiego pliku ani katalogu sed: sed jest /usr/bin/sed, sed (GNU sed) 4.8 ls: nie ma dostępu do 'cut': Nie ma takiego pliku ani katalogu cut: cut jest /usr/bin/cut, cut (GNU coreutils) 8.32 ls: nie ma dostępu do 'basename': Nie ma takiego pliku ani katalogu basename: basename jest /usr/bin/basename, basename (GNU coreutils) 8.32 ls: nie ma dostępu do 'rm': Nie ma takiego pliku ani katalogu rm: rm jest /usr/bin/rm, rm (GNU coreutils) 8.32 ls: nie ma dostępu do 'mkdir': Nie ma takiego pliku ani katalogu mkdir: mkdir jest /usr/bin/mkdir, mkdir (GNU coreutils) 8.32 ls: nie ma dostępu do 'xargs': Nie ma takiego pliku ani katalogu xargs: xargs jest /usr/bin/xargs, xargs (GNU findutils) 4.7.0 nvm current: v8.17.0 which node: ${NVM_DIR}/versions/node/v8.17.0/bin/node which iojs: which: no iojs in (${NVM_DIR}/versions/node/v8.17.0/bin:/usr/local/sbin:/usr/local/bin:/usr/bin:/opt/android-sdk/tools:/opt/android-sdk/tools/bin:/opt/COMODO:/opt/cuda/bin:/var/lib/flatpak/exports/bin:/usr/lib/jvm/default/bin:/usr/lib32/jvm/default/bin:/usr/bin/site_perl:/usr/bin/vendor_perl:/usr/bin/core_perl:/var/lib/snapd/snap/bin:/opt/xtensa-lx106-elf-gcc/bin:${HOME}/.cargo/bin:${HOME}/.local/bin) which npm: ${NVM_DIR}/versions/node/v8.17.0/bin/npm npm config get prefix: ${NVM_DIR}/versions/node/v8.17.0 npm root -g: ${NVM_DIR}/versions/node/v8.17.0/lib/node_modules ``` </details> #### `nvm ls` output: <details> <!-- do not delete the following blank line --> ```sh -> v8.17.0 v12.16.1 system default -> 8 (-> v8.17.0) node -> stable (-> v12.16.1) (default) stable -> 12.16 (-> v12.16.1) (default) iojs -> N/A (default) unstable -> N/A (default) lts/* -> lts/erbium (-> v12.16.1) lts/argon -> v4.9.1 (-> N/A) lts/boron -> v6.17.1 (-> N/A) lts/carbon -> v8.17.0 lts/dubnium -> v10.19.0 (-> N/A) lts/erbium -> v12.16.1 ``` </details> #### How did you install `nvm`? Arch - nvm AUR #### What steps did you perform? Open new gnome terminal by shortcut #### What happened? Terminal opened but bash didn't show for 20s (when HD was slowed down by copy) While HD moderatly used - bash show in 1-2s ( https://youtu.be/Ie4Nnml55-g ) While HD was very busy and terminal load in 20s, i used different already open terminal and while doing ps i found that it's stuck at "npm" command: ``` kolorafa 3024410 0.0 0.0 15844 9960 pts/9 Ss 08:07 0:00 | \_ bash kolorafa 3026045 0.0 0.0 13008 4632 pts/9 R+ 08:09 0:00 | | \_ ps auxf kolorafa 3025222 0.0 0.0 14364 8504 pts/11 Ss+ 08:08 0:00 | \_ bash kolorafa 3025440 0.4 0.1 1036508 39904 pts/11 Dl+ 08:08 0:00 | | \_ npm kolorafa 3025697 0.1 0.0 14364 8348 pts/13 Ss+ 08:09 0:00 | \_ bash kolorafa 3025914 0.8 0.1 1036512 40236 pts/13 Dl+ 08:09 0:00 | \_ npm ``` #### What did you expect to happen? NVM loads without slowing down anything ;) #### Is there anything in any of your profile files that modifies the `PATH`? Don't think anything related
perf
init nvm sh slow mostly while executing npm operating system and version nvm debug output sh nvm version shell bin bash shlvl home home kolorafa nvm dir home nvm path nvm dir versions node bin usr local sbin usr local bin usr bin opt android sdk tools opt android sdk tools bin opt comodo opt cuda bin var lib flatpak exports bin usr lib jvm default bin usr jvm default bin usr bin site perl usr bin vendor perl usr bin core perl var lib snapd snap bin opt xtensa elf gcc bin home cargo bin home local bin prefix npm config prefix nvm nodejs org mirror nvm iojs org mirror shell version gnu bash wersja release pc linux gnu uname a linux smp preempt wed jun gnu linux os version antergos linux curl curl jest usr bin curl curl pc linux gnu libcurl openssl zlib libpsl wget wget jest usr bin wget gnu wget zbudowany na systemie linux gnu ls nie ma dostępu do git nie ma takiego pliku ani katalogu git git jest usr bin git git version ls nie ma dostępu do grep nie ma takiego pliku ani katalogu grep grep jest aliasem do grep colour auto grep gnu grep ls nie ma dostępu do awk nie ma takiego pliku ani katalogu awk awk jest usr bin awk gnu awk api gnu mpfr gnu mp ls nie ma dostępu do sed nie ma takiego pliku ani katalogu sed sed jest usr bin sed sed gnu sed ls nie ma dostępu do cut nie ma takiego pliku ani katalogu cut cut jest usr bin cut cut gnu coreutils ls nie ma dostępu do basename nie ma takiego pliku ani katalogu basename basename jest usr bin basename basename gnu coreutils ls nie ma dostępu do rm nie ma takiego pliku ani katalogu rm rm jest usr bin rm rm gnu coreutils ls nie ma dostępu do mkdir nie ma takiego pliku ani katalogu mkdir mkdir jest usr bin mkdir mkdir gnu coreutils ls nie ma dostępu do xargs nie ma takiego pliku ani katalogu xargs xargs jest usr bin xargs xargs gnu findutils nvm current which node nvm dir versions node bin node which iojs which no iojs in nvm dir versions node bin usr local sbin usr local bin usr bin opt android sdk tools opt android sdk tools bin opt comodo opt cuda bin var lib flatpak exports bin usr lib jvm default bin usr jvm default bin usr bin site perl usr bin vendor perl usr bin core perl var lib snapd snap bin opt xtensa elf gcc bin home cargo bin home local bin which npm nvm dir versions node bin npm npm config get prefix nvm dir versions node npm root g nvm dir versions node lib node modules nvm ls output sh system default node stable default stable default iojs n a default unstable n a default lts lts erbium lts argon n a lts boron n a lts carbon lts dubnium n a lts erbium how did you install nvm arch nvm aur what steps did you perform open new gnome terminal by shortcut what happened terminal opened but bash didn t show for when hd was slowed down by copy while hd moderatly used bash show in while hd was very busy and terminal load in i used different already open terminal and while doing ps i found that it s stuck at npm command kolorafa pts ss bash kolorafa pts r ps auxf kolorafa pts ss bash kolorafa pts dl npm kolorafa pts ss bash kolorafa pts dl npm what did you expect to happen nvm loads without slowing down anything is there anything in any of your profile files that modifies the path don t think anything related
1
81,369
23,449,062,922
IssuesEvent
2022-08-15 23:21:53
trilinos/Trilinos
https://api.github.com/repos/trilinos/Trilinos
closed
Many packages erroneously using abs dir CMAKE_PREFIX_PATH for argument to install()
type: bug pkg: Kokkos pkg: STK pkg: Teuchos pkg: Intrepid2 pkg: ROL pkg: ShyLU impacting: configure or build pkg: KokkosKernels pkg: Krino
@tasmith4 ## Bug Report @trilinos/kokkos, @trilinos/kokkos-kernels, @trilinos/teuchos ### Description While setting up automated installation testing for Trilinos needed to build and run tests for simpleBuildAgainstTrilinos, I ran into a problem with Kokkos and KokkosKernels install() commands using abs dir for the DESTINATION. For example, Kokkos/CMakeLists.txt has: ``` IF (KOKKOS_HAS_TRILINOS) SET(TRILINOS_INCDIR ${CMAKE_INSTALL_PREFIX}/${${PROJECT_NAME}_INSTALL_INCLUDE_DIR}) ... ``` Fixing this is easy. Just make it: ``` IF (KOKKOS_HAS_TRILINOS) SET(TRILINOS_INCDIR ${${PROJECT_NAME}_INSTALL_INCLUDE_DIR}) ... ``` As explained in the [CMake install() documentation](https://cmake.org/cmake/help/v3.17/command/install.html): > If a relative path is given it is interpreted relative to the value of the [CMAKE_INSTALL_PREFIX](https://cmake.org/cmake/help/v3.17/variable/CMAKE_INSTALL_PREFIX.html#variable:CMAKE_INSTALL_PREFIX) variable. That allows using: ``` $ cmake --install . --prefix <some-other-path> ``` ### Steps to Reproduce 1. SHA1: 042bdd7648a 1. Configure script: Any configure script 1. Configure log: TBD 1. Build log: TBD 1. Input deck: N.A. 1. Configure the project with Kokkos, KokkosKernels, and Teuchos enabled **without setting CMAKE_INSTALL_PREFIX** and and then run `cmake --install . --prefix ${PWD}/install`.
1.0
Many packages erroneously using abs dir CMAKE_PREFIX_PATH for argument to install() - @tasmith4 ## Bug Report @trilinos/kokkos, @trilinos/kokkos-kernels, @trilinos/teuchos ### Description While setting up automated installation testing for Trilinos needed to build and run tests for simpleBuildAgainstTrilinos, I ran into a problem with Kokkos and KokkosKernels install() commands using abs dir for the DESTINATION. For example, Kokkos/CMakeLists.txt has: ``` IF (KOKKOS_HAS_TRILINOS) SET(TRILINOS_INCDIR ${CMAKE_INSTALL_PREFIX}/${${PROJECT_NAME}_INSTALL_INCLUDE_DIR}) ... ``` Fixing this is easy. Just make it: ``` IF (KOKKOS_HAS_TRILINOS) SET(TRILINOS_INCDIR ${${PROJECT_NAME}_INSTALL_INCLUDE_DIR}) ... ``` As explained in the [CMake install() documentation](https://cmake.org/cmake/help/v3.17/command/install.html): > If a relative path is given it is interpreted relative to the value of the [CMAKE_INSTALL_PREFIX](https://cmake.org/cmake/help/v3.17/variable/CMAKE_INSTALL_PREFIX.html#variable:CMAKE_INSTALL_PREFIX) variable. That allows using: ``` $ cmake --install . --prefix <some-other-path> ``` ### Steps to Reproduce 1. SHA1: 042bdd7648a 1. Configure script: Any configure script 1. Configure log: TBD 1. Build log: TBD 1. Input deck: N.A. 1. Configure the project with Kokkos, KokkosKernels, and Teuchos enabled **without setting CMAKE_INSTALL_PREFIX** and and then run `cmake --install . --prefix ${PWD}/install`.
non_perf
many packages erroneously using abs dir cmake prefix path for argument to install bug report trilinos kokkos trilinos kokkos kernels trilinos teuchos description while setting up automated installation testing for trilinos needed to build and run tests for simplebuildagainsttrilinos i ran into a problem with kokkos and kokkoskernels install commands using abs dir for the destination for example kokkos cmakelists txt has if kokkos has trilinos set trilinos incdir cmake install prefix project name install include dir fixing this is easy just make it if kokkos has trilinos set trilinos incdir project name install include dir as explained in the if a relative path is given it is interpreted relative to the value of the variable that allows using cmake install prefix steps to reproduce configure script any configure script configure log tbd build log tbd input deck n a configure the project with kokkos kokkoskernels and teuchos enabled without setting cmake install prefix and and then run cmake install prefix pwd install
0
89,670
18,019,568,097
IssuesEvent
2021-09-16 17:36:22
WordPress/openverse-frontend
https://api.github.com/repos/WordPress/openverse-frontend
closed
[Bug] Managing playback of multiple media files
🟧 priority: high 🛠 goal: fix 💻 aspect: code
## Description <!-- Concisely describe the bug. --> The current setup allows for multiple audio files to be played concurrently, which is a bad user experience. ## Reproduction <!-- Provide detailed steps to reproduce the bug. --> 1. View any page with multiple audio players 2. Press play on multiple audio players 3. Listen to the resulting 'chaos orchestra' ## Expectation <!-- Concisely describe what you expected to happen. --> When pressing 'play' on an audio file, if there is _already_ an active audio file it should be paused. ## Screenshots <!-- Add screenshots to show the problem; or delete the section entirely. --> ## Resolution <!-- Replace the [ ] with [x] to check the box. --> I have proposed a solution in #183. - [ ] 🙋 I would be interested in resolving this bug.
1.0
[Bug] Managing playback of multiple media files - ## Description <!-- Concisely describe the bug. --> The current setup allows for multiple audio files to be played concurrently, which is a bad user experience. ## Reproduction <!-- Provide detailed steps to reproduce the bug. --> 1. View any page with multiple audio players 2. Press play on multiple audio players 3. Listen to the resulting 'chaos orchestra' ## Expectation <!-- Concisely describe what you expected to happen. --> When pressing 'play' on an audio file, if there is _already_ an active audio file it should be paused. ## Screenshots <!-- Add screenshots to show the problem; or delete the section entirely. --> ## Resolution <!-- Replace the [ ] with [x] to check the box. --> I have proposed a solution in #183. - [ ] 🙋 I would be interested in resolving this bug.
non_perf
managing playback of multiple media files description the current setup allows for multiple audio files to be played concurrently which is a bad user experience reproduction view any page with multiple audio players press play on multiple audio players listen to the resulting chaos orchestra expectation when pressing play on an audio file if there is already an active audio file it should be paused screenshots resolution i have proposed a solution in 🙋 i would be interested in resolving this bug
0
28,649
13,771,817,941
IssuesEvent
2020-10-07 22:52:01
metabase/metabase
https://api.github.com/repos/metabase/metabase
closed
Option to limit tables/fields that are synced to specific schemas
Administration/Data Model Administration/Metadata & Sync Type:New Feature Type:Performance
We're working on a relatively small system in redshift that uses a shared redshift cluster. There are lots of schemas most of which are not ours (and we don't have permission to read) but our local postgres is becoming unnecessarily heavy due to the automatic caching of fields in all schemas into metabase_field (it would likely be a few hundred rows, but is getting towards a hundred thousand). A possible enhancement would be to have the option to limit field caching to schemas that have been viewed rather than crawling all schemas up front.
True
Option to limit tables/fields that are synced to specific schemas - We're working on a relatively small system in redshift that uses a shared redshift cluster. There are lots of schemas most of which are not ours (and we don't have permission to read) but our local postgres is becoming unnecessarily heavy due to the automatic caching of fields in all schemas into metabase_field (it would likely be a few hundred rows, but is getting towards a hundred thousand). A possible enhancement would be to have the option to limit field caching to schemas that have been viewed rather than crawling all schemas up front.
perf
option to limit tables fields that are synced to specific schemas we re working on a relatively small system in redshift that uses a shared redshift cluster there are lots of schemas most of which are not ours and we don t have permission to read but our local postgres is becoming unnecessarily heavy due to the automatic caching of fields in all schemas into metabase field it would likely be a few hundred rows but is getting towards a hundred thousand a possible enhancement would be to have the option to limit field caching to schemas that have been viewed rather than crawling all schemas up front
1
117,388
25,106,327,580
IssuesEvent
2022-11-08 16:53:29
serbanghita/Mobile-Detect
https://api.github.com/repos/serbanghita/Mobile-Detect
closed
Feature request: getDevice and getDeviceCategory
area: Code Quality deprecated
Great and usefull work! Is there already a way to get a string which device was detected? If not, it would be great to get 2 methods: function getDevice() -> returns the device which was detected (like "iPhone", "iPad", "Samsung", "SamsungTablet" etc). And (just for convenience): function getDeviceCategory() -> returns "mobile", "tablet" or "desktop" as a string
1.0
Feature request: getDevice and getDeviceCategory - Great and usefull work! Is there already a way to get a string which device was detected? If not, it would be great to get 2 methods: function getDevice() -> returns the device which was detected (like "iPhone", "iPad", "Samsung", "SamsungTablet" etc). And (just for convenience): function getDeviceCategory() -> returns "mobile", "tablet" or "desktop" as a string
non_perf
feature request getdevice and getdevicecategory great and usefull work is there already a way to get a string which device was detected if not it would be great to get methods function getdevice returns the device which was detected like iphone ipad samsung samsungtablet etc and just for convenience function getdevicecategory returns mobile tablet or desktop as a string
0
30,317
14,517,239,049
IssuesEvent
2020-12-13 18:50:44
johnboiles/obs-mac-virtualcam
https://api.github.com/repos/johnboiles/obs-mac-virtualcam
opened
Use `OBSDALCMSampleBufferCreateFromDataNoCopy` to eliminate a framebuffer memory copy
enhancement performance
We could potentially improve performance by using [`OBSDALCMSampleBufferCreateFromDataNoCopy`](https://github.com/johnboiles/obs-mac-virtualcam/blob/d6b5db2f07d92e354e5a1f8f8783360eaca1c8bc/src/dal-plugin/OBSDALCMSampleBufferUtils.mm#L102) instead of `OBSDALCMSampleBufferCreateFromData` which should remove a memory copy of the framebuffer, saving a bit of performance and latency transferring the frames from OBS to the virtual camera. When I tried this, it seemed to work just fine when using the OBS Virtual Camera device in other programs. But when using the plugin as a source in OBS (looping back the output of OBS), it didn't work. Strangely when I'd set the source to a lower resolution than the output it would work for some reason. This made me worried that using `OBSDALCMSampleBufferCreateFromDataNoCopy` could cause problems in some programs. I'm really not sure why this is, but it's possible this is a bug in OBS, and in fact this plugin would be just fine eliminating this memory copy. More investigation is needed.
True
Use `OBSDALCMSampleBufferCreateFromDataNoCopy` to eliminate a framebuffer memory copy - We could potentially improve performance by using [`OBSDALCMSampleBufferCreateFromDataNoCopy`](https://github.com/johnboiles/obs-mac-virtualcam/blob/d6b5db2f07d92e354e5a1f8f8783360eaca1c8bc/src/dal-plugin/OBSDALCMSampleBufferUtils.mm#L102) instead of `OBSDALCMSampleBufferCreateFromData` which should remove a memory copy of the framebuffer, saving a bit of performance and latency transferring the frames from OBS to the virtual camera. When I tried this, it seemed to work just fine when using the OBS Virtual Camera device in other programs. But when using the plugin as a source in OBS (looping back the output of OBS), it didn't work. Strangely when I'd set the source to a lower resolution than the output it would work for some reason. This made me worried that using `OBSDALCMSampleBufferCreateFromDataNoCopy` could cause problems in some programs. I'm really not sure why this is, but it's possible this is a bug in OBS, and in fact this plugin would be just fine eliminating this memory copy. More investigation is needed.
perf
use obsdalcmsamplebuffercreatefromdatanocopy to eliminate a framebuffer memory copy we could potentially improve performance by using instead of obsdalcmsamplebuffercreatefromdata which should remove a memory copy of the framebuffer saving a bit of performance and latency transferring the frames from obs to the virtual camera when i tried this it seemed to work just fine when using the obs virtual camera device in other programs but when using the plugin as a source in obs looping back the output of obs it didn t work strangely when i d set the source to a lower resolution than the output it would work for some reason this made me worried that using obsdalcmsamplebuffercreatefromdatanocopy could cause problems in some programs i m really not sure why this is but it s possible this is a bug in obs and in fact this plugin would be just fine eliminating this memory copy more investigation is needed
1
11,625
7,625,401,191
IssuesEvent
2018-05-03 21:19:07
Microsoft/BotBuilder
https://api.github.com/repos/Microsoft/BotBuilder
closed
Bot memory leak in .Net SDK
.NET SDK bug investigate performance
## Bot Info * SDK Platform: .NET * SDK Version: 3.11.0 ## Issue Description Everytime u send a msg to the bot the memory usage stacks up but never gets down. even when the client ends the session with the bot In the picture u can see the diagnostic session from visual studio the events around 1:10min are msgs the bot received from the client ![botframework](https://user-images.githubusercontent.com/18367963/33077168-56e54376-cecf-11e7-8c9d-0bf54b7c1d6d.png) ## Code Example look at Step 1 from reproduction steps ## Reproduction Steps 1. use the visual studio template https://docs.microsoft.com/en-us/bot-framework/dotnet/bot-builder-dotnet-quickstart 2. update to the newest BotBuilder SDK 3. run the bot and send some msgs from the botframework-emulator ## Expected Behavior No Memory Leak ## Actual Results Memory Leak ### Some additional information im pretty new to the bot framework so i dont know if this a feature and memory gets cleaned after some hours or is this a real problem
True
Bot memory leak in .Net SDK - ## Bot Info * SDK Platform: .NET * SDK Version: 3.11.0 ## Issue Description Everytime u send a msg to the bot the memory usage stacks up but never gets down. even when the client ends the session with the bot In the picture u can see the diagnostic session from visual studio the events around 1:10min are msgs the bot received from the client ![botframework](https://user-images.githubusercontent.com/18367963/33077168-56e54376-cecf-11e7-8c9d-0bf54b7c1d6d.png) ## Code Example look at Step 1 from reproduction steps ## Reproduction Steps 1. use the visual studio template https://docs.microsoft.com/en-us/bot-framework/dotnet/bot-builder-dotnet-quickstart 2. update to the newest BotBuilder SDK 3. run the bot and send some msgs from the botframework-emulator ## Expected Behavior No Memory Leak ## Actual Results Memory Leak ### Some additional information im pretty new to the bot framework so i dont know if this a feature and memory gets cleaned after some hours or is this a real problem
perf
bot memory leak in net sdk bot info sdk platform net sdk version issue description everytime u send a msg to the bot the memory usage stacks up but never gets down even when the client ends the session with the bot in the picture u can see the diagnostic session from visual studio the events around are msgs the bot received from the client code example look at step from reproduction steps reproduction steps use the visual studio template update to the newest botbuilder sdk run the bot and send some msgs from the botframework emulator expected behavior no memory leak actual results memory leak some additional information im pretty new to the bot framework so i dont know if this a feature and memory gets cleaned after some hours or is this a real problem
1
End of preview. Expand in Data Studio

Dataset Card for "binary-10IQR-perf"

More Information needed

Downloads last month
20

Collection including karths/binary-10IQR-perf