Dataset Viewer
Auto-converted to Parquet
Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
5
112
repo_url
stringlengths
34
141
action
stringclasses
3 values
title
stringlengths
1
1k
labels
stringlengths
4
1.38k
body
stringlengths
1
262k
index
stringclasses
16 values
text_combine
stringlengths
96
262k
label
stringclasses
2 values
text
stringlengths
96
252k
binary_label
int64
0
1
276,343
8,597,573,009
IssuesEvent
2018-11-15 19:05:58
bounswe/bounswe2018group9
https://api.github.com/repos/bounswe/bounswe2018group9
opened
Event Creation Endpoints Update
Platform: Backend Priority: High
Event creation endpoint has to be improved because we added new fields to the event data model. We need to make sure that before saving an event we have all the necessary fields created in the database.
1.0
Event Creation Endpoints Update - Event creation endpoint has to be improved because we added new fields to the event data model. We need to make sure that before saving an event we have all the necessary fields created in the database.
priority
event creation endpoints update event creation endpoint has to be improved because we added new fields to the event data model we need to make sure that before saving an event we have all the necessary fields created in the database
1
74,419
14,247,514,485
IssuesEvent
2020-11-19 11:33:59
Regalis11/Barotrauma
https://api.github.com/repos/Regalis11/Barotrauma
opened
Bots can't find a path through the docking module.
Bug Code
Ordering a bot from inside the sub to "wait" at the specified location (the pause with a circle to the right), results in them equipping the dive suit correctly, going out of the sub's airlock, but can't manage to path through the docking port module. _(the bot had a scooter on it to make it more obvious its trying to swim)_ **[Sometimes they also get stuck by trying to swim through the corner of the dugong's airlock on the outside when attempting to path to the same "wait" location]** ![pathingthing](https://user-images.githubusercontent.com/72674994/98405987-20187380-20a8-11eb-89f1-414bc60432bd.jpg)
1.0
Bots can't find a path through the docking module. - Ordering a bot from inside the sub to "wait" at the specified location (the pause with a circle to the right), results in them equipping the dive suit correctly, going out of the sub's airlock, but can't manage to path through the docking port module. _(the bot had a scooter on it to make it more obvious its trying to swim)_ **[Sometimes they also get stuck by trying to swim through the corner of the dugong's airlock on the outside when attempting to path to the same "wait" location]** ![pathingthing](https://user-images.githubusercontent.com/72674994/98405987-20187380-20a8-11eb-89f1-414bc60432bd.jpg)
non_priority
bots can t find a path through the docking module ordering a bot from inside the sub to wait at the specified location the pause with a circle to the right results in them equipping the dive suit correctly going out of the sub s airlock but can t manage to path through the docking port module the bot had a scooter on it to make it more obvious its trying to swim
0
94,924
19,607,332,893
IssuesEvent
2022-01-06 11:08:45
creativecommons/vocabulary
https://api.github.com/repos/creativecommons/vocabulary
opened
Split styles and fonts packages into separate projects
🟩 priority: low 🚦 status: awaiting triage ✨ goal: improvement 💻 aspect: code
## Problem The structure of the Vocabulary project and lack of developer continuity has led to maintenance issues blocking the release of `styles` and `fonts` packages. We need to publish new versions of `styles` and `fonts` as the creativecommons.org redesign roadmap progresses. ## Description The Vocabulary project includes transient dependencies preventing us from building and releasing the code, such as those related to `node-sass` and `storybook`. We have tried to reconcile the dependency issues for the past few months but are reaching the limit of our available resources to continue the struggle. An alternative approach would be to revert the packages to be in separate, simplified projects, containing only dependencies essential to their function. To the extent possible, we should separate dependencies for interactive documentation from core package dependencies to reduce complexity and adverse interactions. This restructuring will improve the immediate maintainability of the projects while leaving open the issue of providing a holistic documentation/development environment. ## Alternatives Keep the monorepo structure and continue struggling to restore the project to a healthy state. However, the above restructuring would not preclude any further maintenance on this Vocabulary project but would provide some immediate respite and allow us to move forward with the creativecommons.org redesign. ## Additional context We hope to launch the new creativecommons.org redesign this month.
1.0
Split styles and fonts packages into separate projects - ## Problem The structure of the Vocabulary project and lack of developer continuity has led to maintenance issues blocking the release of `styles` and `fonts` packages. We need to publish new versions of `styles` and `fonts` as the creativecommons.org redesign roadmap progresses. ## Description The Vocabulary project includes transient dependencies preventing us from building and releasing the code, such as those related to `node-sass` and `storybook`. We have tried to reconcile the dependency issues for the past few months but are reaching the limit of our available resources to continue the struggle. An alternative approach would be to revert the packages to be in separate, simplified projects, containing only dependencies essential to their function. To the extent possible, we should separate dependencies for interactive documentation from core package dependencies to reduce complexity and adverse interactions. This restructuring will improve the immediate maintainability of the projects while leaving open the issue of providing a holistic documentation/development environment. ## Alternatives Keep the monorepo structure and continue struggling to restore the project to a healthy state. However, the above restructuring would not preclude any further maintenance on this Vocabulary project but would provide some immediate respite and allow us to move forward with the creativecommons.org redesign. ## Additional context We hope to launch the new creativecommons.org redesign this month.
non_priority
split styles and fonts packages into separate projects problem the structure of the vocabulary project and lack of developer continuity has led to maintenance issues blocking the release of styles and fonts packages we need to publish new versions of styles and fonts as the creativecommons org redesign roadmap progresses description the vocabulary project includes transient dependencies preventing us from building and releasing the code such as those related to node sass and storybook we have tried to reconcile the dependency issues for the past few months but are reaching the limit of our available resources to continue the struggle an alternative approach would be to revert the packages to be in separate simplified projects containing only dependencies essential to their function to the extent possible we should separate dependencies for interactive documentation from core package dependencies to reduce complexity and adverse interactions this restructuring will improve the immediate maintainability of the projects while leaving open the issue of providing a holistic documentation development environment alternatives keep the monorepo structure and continue struggling to restore the project to a healthy state however the above restructuring would not preclude any further maintenance on this vocabulary project but would provide some immediate respite and allow us to move forward with the creativecommons org redesign additional context we hope to launch the new creativecommons org redesign this month
0
11,291
3,481,926,277
IssuesEvent
2015-12-29 19:28:59
wp-cli/wp-cli
https://api.github.com/repos/wp-cli/wp-cli
closed
`wp help`: don't show GLOBAL PARAMETERS section by default, add `--global-help` switch
scope:documentation scope:framework
When using `wp help`, the actual command help usually scrolls out of view because there is a relatively long GLOBAL PARAMETERS section at the end (which is generally not very useful). Maybe this long section could be replaced with a message like: Use --global-help to show global parameters and that switch could be added to the `wp help` command.
1.0
`wp help`: don't show GLOBAL PARAMETERS section by default, add `--global-help` switch - When using `wp help`, the actual command help usually scrolls out of view because there is a relatively long GLOBAL PARAMETERS section at the end (which is generally not very useful). Maybe this long section could be replaced with a message like: Use --global-help to show global parameters and that switch could be added to the `wp help` command.
non_priority
wp help don t show global parameters section by default add global help switch when using wp help the actual command help usually scrolls out of view because there is a relatively long global parameters section at the end which is generally not very useful maybe this long section could be replaced with a message like use global help to show global parameters and that switch could be added to the wp help command
0
662,495
22,141,543,643
IssuesEvent
2022-06-03 07:28:17
threefoldtech/tfchain_graphql
https://api.github.com/repos/threefoldtech/tfchain_graphql
closed
Public IP's are not synced properly
priority_major
On `FarmUpdated` the list of ips should be checked with what is saved in the database. Also `IPsReserved(u64, Vec<PublicIP>)` and `IPsFreed(u64, Vec<Vec<u8>>),` should mutate the public ips
1.0
Public IP's are not synced properly - On `FarmUpdated` the list of ips should be checked with what is saved in the database. Also `IPsReserved(u64, Vec<PublicIP>)` and `IPsFreed(u64, Vec<Vec<u8>>),` should mutate the public ips
priority
public ip s are not synced properly on farmupdated the list of ips should be checked with what is saved in the database also ipsreserved vec and ipsfreed vec should mutate the public ips
1
172,302
14,356,744,092
IssuesEvent
2020-11-30 12:01:53
JetBrains/gradle-intellij-plugin
https://api.github.com/repos/JetBrains/gradle-intellij-plugin
closed
pluginVerifier doesn't fail when compatibility issues are found
documentation duplicate
In my project, I have: ```properties ideaVersion=2019.2.4 pluginSinceBuild=192.5728.98 pluginUntilBuild=203.* pluginVerifierIdeVersions=2019.2.4,2019.3.5,2020.1.4,2020.2.4,203.5981.114 ``` When running the pluginVerifier, I get told that the plugin is compatible with 2020.1.4, 2020.2.4, and 203.5981.114; but that it's not compatible with 2019.2.4, and 2019.3.5. Yet, the task completes and I get a `BUILD SUCCESSFUL` result. I'd expect the build to fail if any compatibility issues are found though. CC @hsz — you know the project I'm talking about :)
1.0
pluginVerifier doesn't fail when compatibility issues are found - In my project, I have: ```properties ideaVersion=2019.2.4 pluginSinceBuild=192.5728.98 pluginUntilBuild=203.* pluginVerifierIdeVersions=2019.2.4,2019.3.5,2020.1.4,2020.2.4,203.5981.114 ``` When running the pluginVerifier, I get told that the plugin is compatible with 2020.1.4, 2020.2.4, and 203.5981.114; but that it's not compatible with 2019.2.4, and 2019.3.5. Yet, the task completes and I get a `BUILD SUCCESSFUL` result. I'd expect the build to fail if any compatibility issues are found though. CC @hsz — you know the project I'm talking about :)
non_priority
pluginverifier doesn t fail when compatibility issues are found in my project i have properties ideaversion pluginsincebuild pluginuntilbuild pluginverifierideversions when running the pluginverifier i get told that the plugin is compatible with and but that it s not compatible with and yet the task completes and i get a build successful result i d expect the build to fail if any compatibility issues are found though cc hsz — you know the project i m talking about
0
118,237
4,733,304,864
IssuesEvent
2016-10-19 10:43:10
thommoboy/There-are-no-brakes
https://api.github.com/repos/thommoboy/There-are-no-brakes
closed
PRE TESTING - Player walk and idle animation doesnt loop correctly
bug Priority Low
also jittering when player stops walking
1.0
PRE TESTING - Player walk and idle animation doesnt loop correctly - also jittering when player stops walking
priority
pre testing player walk and idle animation doesnt loop correctly also jittering when player stops walking
1
242,579
7,844,511,243
IssuesEvent
2018-06-19 09:51:57
xcat2/xcat-core
https://api.github.com/repos/xcat2/xcat-core
closed
[New xCAT prototype]Implement petitboot plugin to deploy OpenPOWER bare metal servers
priority:high sprint2 type:feature
- [x] Enhance copycd interface to create and destroy osimage. - [x] Update dhcp interface to support 209 options for petitboot - [x] Build petitboot configuration file - [x] Add clean up method to delete configurations during callback(deploy complete) or clean request( -d option). - [x] Enhance the abstraction interface for pxe and petitboot
1.0
[New xCAT prototype]Implement petitboot plugin to deploy OpenPOWER bare metal servers - - [x] Enhance copycd interface to create and destroy osimage. - [x] Update dhcp interface to support 209 options for petitboot - [x] Build petitboot configuration file - [x] Add clean up method to delete configurations during callback(deploy complete) or clean request( -d option). - [x] Enhance the abstraction interface for pxe and petitboot
priority
implement petitboot plugin to deploy openpower bare metal servers enhance copycd interface to create and destroy osimage update dhcp interface to support options for petitboot build petitboot configuration file add clean up method to delete configurations during callback deploy complete or clean request d option enhance the abstraction interface for pxe and petitboot
1
12,066
3,251,684,182
IssuesEvent
2015-10-19 11:14:24
WormBase/website
https://api.github.com/repos/WormBase/website
closed
dpy-17 search is incorrectly redirected
Bug Report HelpDesk Under testing Webteam
*Submitted from the feedback form on the WormBase website.* <p>&nbsp;&nbsp;A recent thread on the forum (http://forums.wormbase.org/index.php?topic=2631) uncovered some issues with the search function to identify balancers that cover a gene of interest. It appears that the data are incomplete and/or the search function not working as expected. For example, searching 'dpy-17' under Rearrangement returns hT2 only. However, the mT1 summary page explicitly lists dpy-17 as a positive marker.</p> **Reported by:** Haro******** (smit******************) **Submitted from:** <a target="_blank" href="http://www.wormbase.org//tools/support">/tools/support</a> **Browser:** Mozilla/5.0 (Macintosh; Intel Mac OS X 10_8_5) AppleWebKit/600.5.17 (KHTML, like Gecko) Version/6.2.5 Safari/537.85.14
1.0
dpy-17 search is incorrectly redirected - *Submitted from the feedback form on the WormBase website.* <p>&nbsp;&nbsp;A recent thread on the forum (http://forums.wormbase.org/index.php?topic=2631) uncovered some issues with the search function to identify balancers that cover a gene of interest. It appears that the data are incomplete and/or the search function not working as expected. For example, searching 'dpy-17' under Rearrangement returns hT2 only. However, the mT1 summary page explicitly lists dpy-17 as a positive marker.</p> **Reported by:** Haro******** (smit******************) **Submitted from:** <a target="_blank" href="http://www.wormbase.org//tools/support">/tools/support</a> **Browser:** Mozilla/5.0 (Macintosh; Intel Mac OS X 10_8_5) AppleWebKit/600.5.17 (KHTML, like Gecko) Version/6.2.5 Safari/537.85.14
non_priority
dpy search is incorrectly redirected submitted from the feedback form on the wormbase website nbsp nbsp a recent thread on the forum uncovered some issues with the search function to identify balancers that cover a gene of interest it appears that the data are incomplete and or the search function not working as expected for example searching dpy under rearrangement returns only however the summary page explicitly lists dpy as a positive marker reported by haro smit submitted from a target blank href browser mozilla macintosh intel mac os x applewebkit khtml like gecko version safari
0
61,528
15,020,395,477
IssuesEvent
2021-02-01 14:40:09
spacetelescope/romancal
https://api.github.com/repos/spacetelescope/romancal
closed
Mockup level 1 data file
build0.1
_Issue [RCAL-73](https://jira.stsci.edu/browse/RCAL-73) was created on JIRA by [David Davis](https://jira.stsci.edu/secure/ViewProfile.jspa?name=ddavis):_ We should use the datamodel to mock up a level 1 data file using the JWST NRC data as a source for some of the keywords. Initially it should be a 3-d cube since nints will be 1 for WIFI data. Once this is done we should investigate how much effort would be needed to adapt the JWST ramp fit code to work in a 3-d data cube.
1.0
Mockup level 1 data file - _Issue [RCAL-73](https://jira.stsci.edu/browse/RCAL-73) was created on JIRA by [David Davis](https://jira.stsci.edu/secure/ViewProfile.jspa?name=ddavis):_ We should use the datamodel to mock up a level 1 data file using the JWST NRC data as a source for some of the keywords. Initially it should be a 3-d cube since nints will be 1 for WIFI data. Once this is done we should investigate how much effort would be needed to adapt the JWST ramp fit code to work in a 3-d data cube.
non_priority
mockup level data file issue was created on jira by we should use the datamodel to mock up a level data file using the jwst nrc data as a source for some of the keywords initially it should be a d cube since nints will be for wifi data once this is done we should investigate how much effort would be needed to adapt the jwst ramp fit code to work in a d data cube
0
262,160
8,251,676,224
IssuesEvent
2018-09-12 08:34:48
unitystation/unitystation
https://api.github.com/repos/unitystation/unitystation
closed
Bug: Nuke detonation video does not play in steam builds
High Priority bug
# Description As per the title. In the steam builds the Nuke det video fails to play ###TODO: - Check player logs after the nuke goes off to see if there is an error: https://docs.unity3d.com/Manual/LogFiles.html - If nothing is found in there, check the Server build logs to see if there was an import problem with the video
1.0
Bug: Nuke detonation video does not play in steam builds - # Description As per the title. In the steam builds the Nuke det video fails to play ###TODO: - Check player logs after the nuke goes off to see if there is an error: https://docs.unity3d.com/Manual/LogFiles.html - If nothing is found in there, check the Server build logs to see if there was an import problem with the video
priority
bug nuke detonation video does not play in steam builds description as per the title in the steam builds the nuke det video fails to play todo check player logs after the nuke goes off to see if there is an error if nothing is found in there check the server build logs to see if there was an import problem with the video
1
52,335
13,747,799,875
IssuesEvent
2020-10-06 08:10:15
idmarinas/lotgd-game
https://api.github.com/repos/idmarinas/lotgd-game
closed
CVE-2018-20821 (Medium) detected in multiple libraries - autoclosed
bug security vulnerability
## CVE-2018-20821 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>node-sass-4.14.1.tgz</b></p></summary> <p> <details><summary><b>node-sass-4.14.1.tgz</b></p></summary> <p>Wrapper around libsass</p> <p>Library home page: <a href="https://registry.npmjs.org/node-sass/-/node-sass-4.14.1.tgz">https://registry.npmjs.org/node-sass/-/node-sass-4.14.1.tgz</a></p> <p>Path to dependency file: /tmp/ws-scm/lotgd-game/package.json</p> <p>Path to vulnerable library: /lotgd-game/node_modules/node-sass/package.json</p> <p> Dependency Hierarchy: - :x: **node-sass-4.14.1.tgz** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/idmarinas/lotgd-game/commit/a68ef33093518f30b58ce6afc7b54944a697f707">a68ef33093518f30b58ce6afc7b54944a697f707</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The parsing component in LibSass through 3.5.5 allows attackers to cause a denial-of-service (uncontrolled recursion in Sass::Parser::parse_css_variable_value in parser.cpp). <p>Publish Date: 2019-04-23 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-20821>CVE-2018-20821</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-20821">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-20821</a></p> <p>Release Date: 2019-04-23</p> <p>Fix Resolution: LibSass - 3.6.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2018-20821 (Medium) detected in multiple libraries - autoclosed - ## CVE-2018-20821 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>node-sass-4.14.1.tgz</b></p></summary> <p> <details><summary><b>node-sass-4.14.1.tgz</b></p></summary> <p>Wrapper around libsass</p> <p>Library home page: <a href="https://registry.npmjs.org/node-sass/-/node-sass-4.14.1.tgz">https://registry.npmjs.org/node-sass/-/node-sass-4.14.1.tgz</a></p> <p>Path to dependency file: /tmp/ws-scm/lotgd-game/package.json</p> <p>Path to vulnerable library: /lotgd-game/node_modules/node-sass/package.json</p> <p> Dependency Hierarchy: - :x: **node-sass-4.14.1.tgz** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/idmarinas/lotgd-game/commit/a68ef33093518f30b58ce6afc7b54944a697f707">a68ef33093518f30b58ce6afc7b54944a697f707</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The parsing component in LibSass through 3.5.5 allows attackers to cause a denial-of-service (uncontrolled recursion in Sass::Parser::parse_css_variable_value in parser.cpp). <p>Publish Date: 2019-04-23 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-20821>CVE-2018-20821</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-20821">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-20821</a></p> <p>Release Date: 2019-04-23</p> <p>Fix Resolution: LibSass - 3.6.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_priority
cve medium detected in multiple libraries autoclosed cve medium severity vulnerability vulnerable libraries node sass tgz node sass tgz wrapper around libsass library home page a href path to dependency file tmp ws scm lotgd game package json path to vulnerable library lotgd game node modules node sass package json dependency hierarchy x node sass tgz vulnerable library found in head commit a href vulnerability details the parsing component in libsass through allows attackers to cause a denial of service uncontrolled recursion in sass parser parse css variable value in parser cpp publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution libsass step up your open source security game with whitesource
0
101,437
11,235,847,685
IssuesEvent
2020-01-09 09:16:01
xtensor-stack/xtensor
https://api.github.com/repos/xtensor-stack/xtensor
closed
slicing and indexing
Documentation Question
How to handle multi-dimensional array slicing by using xtensor or some other alternatives: a.shape (1, 120, 56, 1) `b = a[0, :, :, 0]` ? in numpy: ``` a = np.random.randn(1, 120, 56, 1) // a.shape (1, 120, 56, 1) b = a[0, :, :, 0] // b.shape(120, 56) ``` Could you suggest some example or how to do. Thank you beforehand for your time
1.0
slicing and indexing - How to handle multi-dimensional array slicing by using xtensor or some other alternatives: a.shape (1, 120, 56, 1) `b = a[0, :, :, 0]` ? in numpy: ``` a = np.random.randn(1, 120, 56, 1) // a.shape (1, 120, 56, 1) b = a[0, :, :, 0] // b.shape(120, 56) ``` Could you suggest some example or how to do. Thank you beforehand for your time
non_priority
slicing and indexing how to handle multi dimensional array slicing by using xtensor or some other alternatives a shape b a in numpy a np random randn a shape b a b shape could you suggest some example or how to do thank you beforehand for your time
0
437,040
30,583,924,389
IssuesEvent
2023-07-21 11:56:29
orffen/basicfantasyrpg
https://api.github.com/repos/orffen/basicfantasyrpg
closed
README updates/Ability score shortcuts
bug documentation
- [x] The README says that ability scores can be used by using `@str`, but actually `@str.value` needs to be used. Alternatively, find a way to keep `@str.bonus` but also have `@str` automatically resolve to `@str.value`. - [x] Change _LICENSE.txt_ to point to latest [LICENSE.txt](https://raw.githubusercontent.com/orffen/basicfantasyrpg/main/LICENSE.txt)) file.
1.0
README updates/Ability score shortcuts - - [x] The README says that ability scores can be used by using `@str`, but actually `@str.value` needs to be used. Alternatively, find a way to keep `@str.bonus` but also have `@str` automatically resolve to `@str.value`. - [x] Change _LICENSE.txt_ to point to latest [LICENSE.txt](https://raw.githubusercontent.com/orffen/basicfantasyrpg/main/LICENSE.txt)) file.
non_priority
readme updates ability score shortcuts the readme says that ability scores can be used by using str but actually str value needs to be used alternatively find a way to keep str bonus but also have str automatically resolve to str value change license txt to point to latest file
0
303,455
9,307,410,749
IssuesEvent
2019-03-25 12:13:29
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
closed
forum.xda-developers.com - design is broken
browser-firefox-tablet priority-important
<!-- @browser: Firefox Mobile (Tablet) 67.0 --> <!-- @ua_header: Mozilla/5.0 (Android 8.1.0; Tablet; rv:67.0) Gecko/67.0 Firefox/67.0 --> <!-- @reported_with: mobile-reporter --> **URL**: https://forum.xda-developers.com/razer-phone-2/help/pie-volume-buttons-t3908250 **Browser / Version**: Firefox Mobile (Tablet) 67.0 **Operating System**: Android 8.1.0 **Tested Another Browser**: Unknown **Problem type**: Design is broken **Description**: screen formatting razer phone 2 **Steps to Reproduce**: cutting top of screen it's a phone not tablet <details> <summary>Browser Configuration</summary> <ul> <li>mixed active content blocked: false</li><li>image.mem.shared: true</li><li>buildID: 20190306095759</li><li>tracking content blocked: false</li><li>gfx.webrender.blob-images: true</li><li>hasTouchScreen: true</li><li>mixed passive content blocked: false</li><li>gfx.webrender.enabled: false</li><li>gfx.webrender.all: false</li><li>channel: nightly</li> </ul> <p>Console Messages:</p> <pre> [u'[JavaScript Error: "ReferenceError: vB_Text_Editor is not defined" {file: "https://forum.xda-developers.com/razer-phone-2/help/pie-volume-buttons-t3908250" line: 1249}]\n@https://forum.xda-developers.com/razer-phone-2/help/pie-volume-buttons-t3908250:1249:1\n', u'[console.log(QuickReply Init) https://forum.xda-developers.com/razer-phone-2/help/pie-volume-buttons-t3908250:1808:21]', u'[JavaScript Warning: "Content Security Policy: Ignoring x-frame-options because of frame-ancestors directive."]', u'[JavaScript Warning: "Content Security Policy: Ignoring x-frame-options because of frame-ancestors directive."]', u'[JavaScript Warning: "Cross-Origin Request Blocked: The Same Origin Policy disallows reading the remote resource at https://api.xda-developers.com/user. (Reason: CORS header Access-Control-Allow-Origin missing)."]', u'[console.log(Error during fetch of user data:) https://www.xda-developers.com/dist/js/xda_topnav.js?v=1551904321:13:19767]', u'[console.error([object XMLHttpRequest]) https://www.xda-developers.com/dist/js/xda_topnav.js?v=1551904321:13:19815]', u'[JavaScript Warning: "Use of Mutation Events is deprecated. Use MutationObserver instead." {file: "https://ajax.googleapis.com/ajax/libs/jquery/2.1.1/jquery.min.js" line: 3}]', u'[console.log(DIGI IFRAME:https://cdn.digitru.st/prod/1.5.24/dt.html) https://cdn.digitru.st/prod/1/digitrust.min.js:1:14745]'] </pre> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
1.0
forum.xda-developers.com - design is broken - <!-- @browser: Firefox Mobile (Tablet) 67.0 --> <!-- @ua_header: Mozilla/5.0 (Android 8.1.0; Tablet; rv:67.0) Gecko/67.0 Firefox/67.0 --> <!-- @reported_with: mobile-reporter --> **URL**: https://forum.xda-developers.com/razer-phone-2/help/pie-volume-buttons-t3908250 **Browser / Version**: Firefox Mobile (Tablet) 67.0 **Operating System**: Android 8.1.0 **Tested Another Browser**: Unknown **Problem type**: Design is broken **Description**: screen formatting razer phone 2 **Steps to Reproduce**: cutting top of screen it's a phone not tablet <details> <summary>Browser Configuration</summary> <ul> <li>mixed active content blocked: false</li><li>image.mem.shared: true</li><li>buildID: 20190306095759</li><li>tracking content blocked: false</li><li>gfx.webrender.blob-images: true</li><li>hasTouchScreen: true</li><li>mixed passive content blocked: false</li><li>gfx.webrender.enabled: false</li><li>gfx.webrender.all: false</li><li>channel: nightly</li> </ul> <p>Console Messages:</p> <pre> [u'[JavaScript Error: "ReferenceError: vB_Text_Editor is not defined" {file: "https://forum.xda-developers.com/razer-phone-2/help/pie-volume-buttons-t3908250" line: 1249}]\n@https://forum.xda-developers.com/razer-phone-2/help/pie-volume-buttons-t3908250:1249:1\n', u'[console.log(QuickReply Init) https://forum.xda-developers.com/razer-phone-2/help/pie-volume-buttons-t3908250:1808:21]', u'[JavaScript Warning: "Content Security Policy: Ignoring x-frame-options because of frame-ancestors directive."]', u'[JavaScript Warning: "Content Security Policy: Ignoring x-frame-options because of frame-ancestors directive."]', u'[JavaScript Warning: "Cross-Origin Request Blocked: The Same Origin Policy disallows reading the remote resource at https://api.xda-developers.com/user. (Reason: CORS header Access-Control-Allow-Origin missing)."]', u'[console.log(Error during fetch of user data:) https://www.xda-developers.com/dist/js/xda_topnav.js?v=1551904321:13:19767]', u'[console.error([object XMLHttpRequest]) https://www.xda-developers.com/dist/js/xda_topnav.js?v=1551904321:13:19815]', u'[JavaScript Warning: "Use of Mutation Events is deprecated. Use MutationObserver instead." {file: "https://ajax.googleapis.com/ajax/libs/jquery/2.1.1/jquery.min.js" line: 3}]', u'[console.log(DIGI IFRAME:https://cdn.digitru.st/prod/1.5.24/dt.html) https://cdn.digitru.st/prod/1/digitrust.min.js:1:14745]'] </pre> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
priority
forum xda developers com design is broken url browser version firefox mobile tablet operating system android tested another browser unknown problem type design is broken description screen formatting razer phone steps to reproduce cutting top of screen it s a phone not tablet browser configuration mixed active content blocked false image mem shared true buildid tracking content blocked false gfx webrender blob images true hastouchscreen true mixed passive content blocked false gfx webrender enabled false gfx webrender all false channel nightly console messages n u u u u u u u u from with ❤️
1
70,632
23,268,678,591
IssuesEvent
2022-08-04 20:09:32
primefaces/primefaces
https://api.github.com/repos/primefaces/primefaces
closed
BlockUI: Incorrect overlay placement
:lady_beetle: defect
### Describe the bug The overlay is displayed incorrectly, the size is correct, but position is shifted. ### Reproducer _No response_ ### Expected behavior It seems it's a mistake to use `.position()` which returns the position of en element relative to its parent. But instead it should be relative to a page `.offset()`, because the overlay element is no more created next to a block element, but instead attached to `@(body)`. ### PrimeFaces edition Community ### PrimeFaces version 12-RC2 ### Theme _No response_ ### JSF implementation MyFaces ### JSF version 3.0.0 ### Browser(s) Chrome 104
1.0
BlockUI: Incorrect overlay placement - ### Describe the bug The overlay is displayed incorrectly, the size is correct, but position is shifted. ### Reproducer _No response_ ### Expected behavior It seems it's a mistake to use `.position()` which returns the position of en element relative to its parent. But instead it should be relative to a page `.offset()`, because the overlay element is no more created next to a block element, but instead attached to `@(body)`. ### PrimeFaces edition Community ### PrimeFaces version 12-RC2 ### Theme _No response_ ### JSF implementation MyFaces ### JSF version 3.0.0 ### Browser(s) Chrome 104
non_priority
blockui incorrect overlay placement describe the bug the overlay is displayed incorrectly the size is correct but position is shifted reproducer no response expected behavior it seems it s a mistake to use position which returns the position of en element relative to its parent but instead it should be relative to a page offset because the overlay element is no more created next to a block element but instead attached to body primefaces edition community primefaces version theme no response jsf implementation myfaces jsf version browser s chrome
0
78,177
3,509,500,412
IssuesEvent
2016-01-08 23:06:00
OregonCore/OregonCore
https://api.github.com/repos/OregonCore/OregonCore
closed
Whispers of the Raven (10607) (BB #947)
Category: Quests migrated Priority: Medium Type: Bug
This issue was migrated from bitbucket. **Original Reporter:** bewtihozzy **Original Date:** 31.05.2015 17:56:23 GMT+0000 **Original Priority:** major **Original Type:** bug **Original State:** resolved **Direct Link:** https://bitbucket.org/oregon/oregoncore/issues/947 <hr> Whispers of the Raven (10607) Kill Grishna arakkoa (22232) to gain Understanding Ravenspeech for 2 minutes. While under the effects of the spell, you can find and receive the prophecy from the four shrines. Return when you have visited all four. Receive the First Prophecy Receive the Second Prophecy Receive the Third Prophecy Receive the Fourth Prophecy bug: When you kill the Grishna arakkoa, it doesnt give you the buff. -> https://youtu.be/SNMBbDA-nR8 Edit: I moved the mob to that spot so he doesnt get attacked by guards xD
1.0
Whispers of the Raven (10607) (BB #947) - This issue was migrated from bitbucket. **Original Reporter:** bewtihozzy **Original Date:** 31.05.2015 17:56:23 GMT+0000 **Original Priority:** major **Original Type:** bug **Original State:** resolved **Direct Link:** https://bitbucket.org/oregon/oregoncore/issues/947 <hr> Whispers of the Raven (10607) Kill Grishna arakkoa (22232) to gain Understanding Ravenspeech for 2 minutes. While under the effects of the spell, you can find and receive the prophecy from the four shrines. Return when you have visited all four. Receive the First Prophecy Receive the Second Prophecy Receive the Third Prophecy Receive the Fourth Prophecy bug: When you kill the Grishna arakkoa, it doesnt give you the buff. -> https://youtu.be/SNMBbDA-nR8 Edit: I moved the mob to that spot so he doesnt get attacked by guards xD
priority
whispers of the raven bb this issue was migrated from bitbucket original reporter bewtihozzy original date gmt original priority major original type bug original state resolved direct link whispers of the raven kill grishna arakkoa to gain understanding ravenspeech for minutes while under the effects of the spell you can find and receive the prophecy from the four shrines return when you have visited all four receive the first prophecy receive the second prophecy receive the third prophecy receive the fourth prophecy bug when you kill the grishna arakkoa it doesnt give you the buff edit i moved the mob to that spot so he doesnt get attacked by guards xd
1
285,315
31,154,172,986
IssuesEvent
2023-08-16 12:06:10
nidhi7598/linux-4.1.15_CVE-2018-5873
https://api.github.com/repos/nidhi7598/linux-4.1.15_CVE-2018-5873
opened
CVE-2023-3111 (High) detected in linuxlinux-4.1.52
Mend: dependency security vulnerability
## CVE-2023-3111 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.1.52</b></p></summary> <p> <p>The Linux Kernel</p> <p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p> <p>Found in HEAD commit: <a href="https://github.com/nidhi7598/linux-4.1.15_CVE-2018-5873/commit/57227113ebd78150a4a62d376636875894cd3d72">57227113ebd78150a4a62d376636875894cd3d72</a></p> <p>Found in base branch: <b>main</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/fs/btrfs/relocation.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/fs/btrfs/relocation.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> A use after free vulnerability was found in prepare_to_relocate in fs/btrfs/relocation.c in btrfs in the Linux Kernel. This possible flaw can be triggered by calling btrfs_ioctl_balance() before calling btrfs_ioctl_defrag(). <p>Publish Date: 2023-06-05 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2023-3111>CVE-2023-3111</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2023-3111">https://www.linuxkernelcves.com/cves/CVE-2023-3111</a></p> <p>Release Date: 2023-06-05</p> <p>Fix Resolution: v4.14.318,v4.19.286,v5.4.247,v5.10.184,v5.15.63</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2023-3111 (High) detected in linuxlinux-4.1.52 - ## CVE-2023-3111 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.1.52</b></p></summary> <p> <p>The Linux Kernel</p> <p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p> <p>Found in HEAD commit: <a href="https://github.com/nidhi7598/linux-4.1.15_CVE-2018-5873/commit/57227113ebd78150a4a62d376636875894cd3d72">57227113ebd78150a4a62d376636875894cd3d72</a></p> <p>Found in base branch: <b>main</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/fs/btrfs/relocation.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/fs/btrfs/relocation.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> A use after free vulnerability was found in prepare_to_relocate in fs/btrfs/relocation.c in btrfs in the Linux Kernel. This possible flaw can be triggered by calling btrfs_ioctl_balance() before calling btrfs_ioctl_defrag(). <p>Publish Date: 2023-06-05 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2023-3111>CVE-2023-3111</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2023-3111">https://www.linuxkernelcves.com/cves/CVE-2023-3111</a></p> <p>Release Date: 2023-06-05</p> <p>Fix Resolution: v4.14.318,v4.19.286,v5.4.247,v5.10.184,v5.15.63</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_priority
cve high detected in linuxlinux cve high severity vulnerability vulnerable library linuxlinux the linux kernel library home page a href found in head commit a href found in base branch main vulnerable source files fs btrfs relocation c fs btrfs relocation c vulnerability details a use after free vulnerability was found in prepare to relocate in fs btrfs relocation c in btrfs in the linux kernel this possible flaw can be triggered by calling btrfs ioctl balance before calling btrfs ioctl defrag publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend
0
112,016
4,500,861,297
IssuesEvent
2016-09-01 07:11:56
rndsolutions/hawkcd
https://api.github.com/repos/rndsolutions/hawkcd
closed
Add option to modify/assign resources for a given job
feature high priority
We need a button in the UI to enable users to assign resources to jobs.
1.0
Add option to modify/assign resources for a given job - We need a button in the UI to enable users to assign resources to jobs.
priority
add option to modify assign resources for a given job we need a button in the ui to enable users to assign resources to jobs
1
268,905
8,415,669,785
IssuesEvent
2018-10-13 17:02:20
react-epic/react-epic
https://api.github.com/repos/react-epic/react-epic
closed
Investigate the way to bundle a Stream
high priority
Allow to bundle a Stream into a StreamBundle. This is important for implementing such `browserHistoryApi` or `httpEntryPoint` (*) that don't need any alternative library > (*): When completed it will work the same way with Relay. Considering using RxJS `Subscriber` or `multicast` operator.
1.0
Investigate the way to bundle a Stream - Allow to bundle a Stream into a StreamBundle. This is important for implementing such `browserHistoryApi` or `httpEntryPoint` (*) that don't need any alternative library > (*): When completed it will work the same way with Relay. Considering using RxJS `Subscriber` or `multicast` operator.
priority
investigate the way to bundle a stream allow to bundle a stream into a streambundle this is important for implementing such browserhistoryapi or httpentrypoint that don t need any alternative library when completed it will work the same way with relay considering using rxjs subscriber or multicast operator
1
734,805
25,365,079,084
IssuesEvent
2022-11-21 05:13:51
wso2/api-manager
https://api.github.com/repos/wso2/api-manager
closed
[MSSQL] Database deadlock when saving API after adding multiple operation policies
Type/Bug Priority/Normal Affected/APIM-4.1.0 Affected/APIM-4.x.x Area/Migration
### Description APIM 4.1.0 runs into a database deadlock scenario when it is setup with MSSQL. There won't be any errors in the UI or in the console. None of the db commands will work in the server after that point (if you go to the landing page and try to go in to an API, the UI will keep loading infinitely). Only option in this scenario is to restart the server. We came across this issue while doing some migration testing. However, this issue is not limited to migrated APIs or new APIs in a migrated pack. This can be reproduced in any 4.1.0 pack configured with MSSQL. ### Steps to Reproduce Take the latest update level of APIM 4.1.0 (or GA) Setup MSSQL as the database Create an API Add about 2 operation policies and save the API. You will be able to save without any issues. Add another policy to the API and try to save. You will be able to see that UI is loading continuously. ### Affected Component APIM ### Version 4.1.0 ### Environment Details (with versions) Migration from lower version to 4.1.0 ### Relevant Log Output _No response_ ### Related Issues _No response_ ### Suggested Labels _No response_
1.0
[MSSQL] Database deadlock when saving API after adding multiple operation policies - ### Description APIM 4.1.0 runs into a database deadlock scenario when it is setup with MSSQL. There won't be any errors in the UI or in the console. None of the db commands will work in the server after that point (if you go to the landing page and try to go in to an API, the UI will keep loading infinitely). Only option in this scenario is to restart the server. We came across this issue while doing some migration testing. However, this issue is not limited to migrated APIs or new APIs in a migrated pack. This can be reproduced in any 4.1.0 pack configured with MSSQL. ### Steps to Reproduce Take the latest update level of APIM 4.1.0 (or GA) Setup MSSQL as the database Create an API Add about 2 operation policies and save the API. You will be able to save without any issues. Add another policy to the API and try to save. You will be able to see that UI is loading continuously. ### Affected Component APIM ### Version 4.1.0 ### Environment Details (with versions) Migration from lower version to 4.1.0 ### Relevant Log Output _No response_ ### Related Issues _No response_ ### Suggested Labels _No response_
priority
database deadlock when saving api after adding multiple operation policies description apim runs into a database deadlock scenario when it is setup with mssql there won t be any errors in the ui or in the console none of the db commands will work in the server after that point if you go to the landing page and try to go in to an api the ui will keep loading infinitely only option in this scenario is to restart the server we came across this issue while doing some migration testing however this issue is not limited to migrated apis or new apis in a migrated pack this can be reproduced in any pack configured with mssql steps to reproduce take the latest update level of apim or ga setup mssql as the database create an api add about operation policies and save the api you will be able to save without any issues add another policy to the api and try to save you will be able to see that ui is loading continuously affected component apim version environment details with versions migration from lower version to relevant log output no response related issues no response suggested labels no response
1
583,851
17,399,811,796
IssuesEvent
2021-08-02 17:57:23
genesis-community/prometheus-genesis-kit
https://api.github.com/repos/genesis-community/prometheus-genesis-kit
reopened
Check hook + genesis config block
PRIORITY: Low-hanging Fruit TRIAGE: Accepted TYPE: Bug
Check hook doesn't fail deployment with missing values - continues without failing Also the new hook doesnt use the `genesis-config-block`
1.0
Check hook + genesis config block - Check hook doesn't fail deployment with missing values - continues without failing Also the new hook doesnt use the `genesis-config-block`
priority
check hook genesis config block check hook doesn t fail deployment with missing values continues without failing also the new hook doesnt use the genesis config block
1
574,504
17,023,827,534
IssuesEvent
2021-07-03 04:03:28
tomhughes/trac-tickets
https://api.github.com/repos/tomhughes/trac-tickets
closed
Add ltmgfy support to osqa-based help system
Component: website Priority: trivial Resolution: invalid Type: enhancement
**[Submitted to the original trac issue database at 10.15am, Thursday, 27th September 2012]** The help.osm.org system works well, but there is a definite issue with repeat questions, like this one: https://help.openstreetmap.org/questions/16493/how-to-import-big-data Sometimes a bit of gentle teasing might be in order, along the lines of lmgtfy.com/?q=help
1.0
Add ltmgfy support to osqa-based help system - **[Submitted to the original trac issue database at 10.15am, Thursday, 27th September 2012]** The help.osm.org system works well, but there is a definite issue with repeat questions, like this one: https://help.openstreetmap.org/questions/16493/how-to-import-big-data Sometimes a bit of gentle teasing might be in order, along the lines of lmgtfy.com/?q=help
priority
add ltmgfy support to osqa based help system the help osm org system works well but there is a definite issue with repeat questions like this one sometimes a bit of gentle teasing might be in order along the lines of lmgtfy com q help
1
624,690
19,704,422,590
IssuesEvent
2022-01-12 20:10:43
svthalia/concrexit
https://api.github.com/repos/svthalia/concrexit
opened
Better selection of events for the agenda section in the newsletter
priority: low feature newsletter
### Is your feature request related to a problem? Please describe. The agenda section in the newsletter should have a more intelligent way of selecting the events that are displayed: - only the ones for the upcoming 2 weeks - or the events that will have registration opening in the upcoming week ### Describe the solution you'd like Changing this The exact implementation will be thought about by @SnevelRetlaw ### Motivation ### Describe alternatives you've considered We have talked with Walter ### Additional context
1.0
Better selection of events for the agenda section in the newsletter - ### Is your feature request related to a problem? Please describe. The agenda section in the newsletter should have a more intelligent way of selecting the events that are displayed: - only the ones for the upcoming 2 weeks - or the events that will have registration opening in the upcoming week ### Describe the solution you'd like Changing this The exact implementation will be thought about by @SnevelRetlaw ### Motivation ### Describe alternatives you've considered We have talked with Walter ### Additional context
priority
better selection of events for the agenda section in the newsletter is your feature request related to a problem please describe the agenda section in the newsletter should have a more intelligent way of selecting the events that are displayed only the ones for the upcoming weeks or the events that will have registration opening in the upcoming week describe the solution you d like changing this the exact implementation will be thought about by snevelretlaw motivation describe alternatives you ve considered we have talked with walter additional context
1
645,257
20,999,835,054
IssuesEvent
2022-03-29 16:22:38
magento/magento2
https://api.github.com/repos/magento/magento2
closed
Running phpstan twice (using the default Magento phpstan configuration) results in strange error
Issue: Confirmed Reproduced on 2.4.x Progress: ready for dev Priority: P3 Reported on 2.4.x Area: Framework
### Preconditions (*) 1. clean Magento installation (I've using 2.4-develop, commit 40ceb7019989c949f3e756efe2db673c8e40345c) with composer developer dependencies (which is the default when running `composer install`) 2. macOS 11.6.1 3. Case sensitive filesystem APFS 4. PHP 7.4.25 natively (no virtualisation or docker or whatever) ### Steps to reproduce (*) 1. Analyse the `app/autoload.php` file with phpstan: `./vendor/bin/phpstan analyse --level 1 app/autoload.php` 2. Run the same command a second or third time. Notice that this works fine everytime 3. Now use the default phpstan configuration file from Magento: `./vendor/bin/phpstan analyse --level 1 --configuration ./dev/tests/static/testsuite/Magento/Test/Php/_files/phpstan/phpstan.neon app/autoload.php` - this runs without issues 4. Now run this exact same command a **second** time - now this **fails** with: ``` $ ./vendor/bin/phpstan analyse --level 1 --configuration ./dev/tests/static/testsuite/Magento/Test/Php/_files/phpstan/phpstan.neon app/autoload.php In File.php line 208: Failed to open stream hoa://Library/Regex/Grammar.pp. analyse [-c|--configuration CONFIGURATION] [-l|--level LEVEL] [--no-progress] [--debug] [-a|--autoload-file AUTOLOAD-FILE] [--error-format ERROR-FORMAT] [--generate-baseline [GENERATE-BASELINE]] [--memory-limit MEMORY-LIMIT] [--xdebug] [--fix] [--watch] [--pro] [--] [<paths>...] ``` 5. Change level of phpstan's analysis from 1 to 2: `./vendor/bin/phpstan analyse --level 2 --configuration ./dev/tests/static/testsuite/Magento/Test/Php/_files/phpstan/phpstan.neon app/autoload.php` - notice that this works fine again 6. Run the exact same command again a **second** time - notice that this **fails** again I have no idea what this is, it only fails with Magento's configuration file and fails the **second** time you try to run the command. Clearing phpstan's cache with `./vendor/bin/phpstan clear-result-cache` doesn't help ... Removing the entire `vendor` directory and running `composer install` also doesn't help ... Problem happens on phpstan versions 0.12.x and 1.x When switching from `2.4-develop` to the `2.4.3` tag and running `composer install`, it does fix the problem, but only for the first run. The **second** run fails again! Switching back to `2.4-develop` and running `composer install` fixes it again but only for the first run. When I search the web for this error, I end up on https://github.com/wunderio/code-quality/issues/45 but I don't understand much of this. So not sure if this is a bug in Magento's phpstan configuration file, or in phpstan, or in one of the hoa libraries, it's hard to say ... But it's very annoying for users of phpstan if they want to run phpstan inside a Magento project and use the default Magento configuration file. ### Expected result (*) 1. No failures ### Actual result (*) 1. Failures, see above --- Please provide [Severity](https://devdocs.magento.com/guides/v2.4/contributor-guide/contributing.html#backlog) assessment for the Issue as Reporter. This information will help during Confirmation and Issue triage processes. - [ ] Severity: **S0** _- Affects critical data or functionality and leaves users without workaround._ - [ ] Severity: **S1** _- Affects critical data or functionality and forces users to employ a workaround._ - [ ] Severity: **S2** _- Affects non-critical data or functionality and forces users to employ a workaround._ - [ ] Severity: **S3** _- Affects non-critical data or functionality and does not force users to employ a workaround._ - [ ] Severity: **S4** _- Affects aesthetics, professional look and feel, “quality” or “usability”._
1.0
Running phpstan twice (using the default Magento phpstan configuration) results in strange error - ### Preconditions (*) 1. clean Magento installation (I've using 2.4-develop, commit 40ceb7019989c949f3e756efe2db673c8e40345c) with composer developer dependencies (which is the default when running `composer install`) 2. macOS 11.6.1 3. Case sensitive filesystem APFS 4. PHP 7.4.25 natively (no virtualisation or docker or whatever) ### Steps to reproduce (*) 1. Analyse the `app/autoload.php` file with phpstan: `./vendor/bin/phpstan analyse --level 1 app/autoload.php` 2. Run the same command a second or third time. Notice that this works fine everytime 3. Now use the default phpstan configuration file from Magento: `./vendor/bin/phpstan analyse --level 1 --configuration ./dev/tests/static/testsuite/Magento/Test/Php/_files/phpstan/phpstan.neon app/autoload.php` - this runs without issues 4. Now run this exact same command a **second** time - now this **fails** with: ``` $ ./vendor/bin/phpstan analyse --level 1 --configuration ./dev/tests/static/testsuite/Magento/Test/Php/_files/phpstan/phpstan.neon app/autoload.php In File.php line 208: Failed to open stream hoa://Library/Regex/Grammar.pp. analyse [-c|--configuration CONFIGURATION] [-l|--level LEVEL] [--no-progress] [--debug] [-a|--autoload-file AUTOLOAD-FILE] [--error-format ERROR-FORMAT] [--generate-baseline [GENERATE-BASELINE]] [--memory-limit MEMORY-LIMIT] [--xdebug] [--fix] [--watch] [--pro] [--] [<paths>...] ``` 5. Change level of phpstan's analysis from 1 to 2: `./vendor/bin/phpstan analyse --level 2 --configuration ./dev/tests/static/testsuite/Magento/Test/Php/_files/phpstan/phpstan.neon app/autoload.php` - notice that this works fine again 6. Run the exact same command again a **second** time - notice that this **fails** again I have no idea what this is, it only fails with Magento's configuration file and fails the **second** time you try to run the command. Clearing phpstan's cache with `./vendor/bin/phpstan clear-result-cache` doesn't help ... Removing the entire `vendor` directory and running `composer install` also doesn't help ... Problem happens on phpstan versions 0.12.x and 1.x When switching from `2.4-develop` to the `2.4.3` tag and running `composer install`, it does fix the problem, but only for the first run. The **second** run fails again! Switching back to `2.4-develop` and running `composer install` fixes it again but only for the first run. When I search the web for this error, I end up on https://github.com/wunderio/code-quality/issues/45 but I don't understand much of this. So not sure if this is a bug in Magento's phpstan configuration file, or in phpstan, or in one of the hoa libraries, it's hard to say ... But it's very annoying for users of phpstan if they want to run phpstan inside a Magento project and use the default Magento configuration file. ### Expected result (*) 1. No failures ### Actual result (*) 1. Failures, see above --- Please provide [Severity](https://devdocs.magento.com/guides/v2.4/contributor-guide/contributing.html#backlog) assessment for the Issue as Reporter. This information will help during Confirmation and Issue triage processes. - [ ] Severity: **S0** _- Affects critical data or functionality and leaves users without workaround._ - [ ] Severity: **S1** _- Affects critical data or functionality and forces users to employ a workaround._ - [ ] Severity: **S2** _- Affects non-critical data or functionality and forces users to employ a workaround._ - [ ] Severity: **S3** _- Affects non-critical data or functionality and does not force users to employ a workaround._ - [ ] Severity: **S4** _- Affects aesthetics, professional look and feel, “quality” or “usability”._
priority
running phpstan twice using the default magento phpstan configuration results in strange error preconditions clean magento installation i ve using develop commit with composer developer dependencies which is the default when running composer install macos case sensitive filesystem apfs php natively no virtualisation or docker or whatever steps to reproduce analyse the app autoload php file with phpstan vendor bin phpstan analyse level app autoload php run the same command a second or third time notice that this works fine everytime now use the default phpstan configuration file from magento vendor bin phpstan analyse level configuration dev tests static testsuite magento test php files phpstan phpstan neon app autoload php this runs without issues now run this exact same command a second time now this fails with vendor bin phpstan analyse level configuration dev tests static testsuite magento test php files phpstan phpstan neon app autoload php in file php line failed to open stream hoa library regex grammar pp analyse change level of phpstan s analysis from to vendor bin phpstan analyse level configuration dev tests static testsuite magento test php files phpstan phpstan neon app autoload php notice that this works fine again run the exact same command again a second time notice that this fails again i have no idea what this is it only fails with magento s configuration file and fails the second time you try to run the command clearing phpstan s cache with vendor bin phpstan clear result cache doesn t help removing the entire vendor directory and running composer install also doesn t help problem happens on phpstan versions x and x when switching from develop to the tag and running composer install it does fix the problem but only for the first run the second run fails again switching back to develop and running composer install fixes it again but only for the first run when i search the web for this error i end up on but i don t understand much of this so not sure if this is a bug in magento s phpstan configuration file or in phpstan or in one of the hoa libraries it s hard to say but it s very annoying for users of phpstan if they want to run phpstan inside a magento project and use the default magento configuration file expected result no failures actual result failures see above please provide assessment for the issue as reporter this information will help during confirmation and issue triage processes severity affects critical data or functionality and leaves users without workaround severity affects critical data or functionality and forces users to employ a workaround severity affects non critical data or functionality and forces users to employ a workaround severity affects non critical data or functionality and does not force users to employ a workaround severity affects aesthetics professional look and feel “quality” or “usability”
1
423,784
12,302,096,515
IssuesEvent
2020-05-11 16:24:22
internetarchive/openlibrary
https://api.github.com/repos/internetarchive/openlibrary
closed
Sitemap returning 404
Lead: @cdrini Priority: 1 Theme: SEO Type: Bug
The sitemap looks like it's no longer available ### Evidence / Screenshot (if possible) Google Search Console: ![image](https://user-images.githubusercontent.com/6251786/80266730-77ef1000-866b-11ea-825b-fe8c1d492ec6.png) ### Relevant url? - https://openlibrary.org/static/sitemaps/siteindex.xml.gz ### Details - **Logged in (Y/N)?** Y - **Browser type/version?** FF75 - **Operating system?** Win10 - **Environment (prod/dev/local)?** prod <!-- If not sure, put prod --> ### Proposal & Constraints <!-- What is the proposed solution / implementation? Is there a precedent of this approach succeeding elsewhere? --> ### Related files <!-- Files related to this issue; this is super useful for new contributors who might want to help! If you're not sure, leave this blank; a maintainer will add them. --> ### Stakeholders @mekarpeles
1.0
Sitemap returning 404 - The sitemap looks like it's no longer available ### Evidence / Screenshot (if possible) Google Search Console: ![image](https://user-images.githubusercontent.com/6251786/80266730-77ef1000-866b-11ea-825b-fe8c1d492ec6.png) ### Relevant url? - https://openlibrary.org/static/sitemaps/siteindex.xml.gz ### Details - **Logged in (Y/N)?** Y - **Browser type/version?** FF75 - **Operating system?** Win10 - **Environment (prod/dev/local)?** prod <!-- If not sure, put prod --> ### Proposal & Constraints <!-- What is the proposed solution / implementation? Is there a precedent of this approach succeeding elsewhere? --> ### Related files <!-- Files related to this issue; this is super useful for new contributors who might want to help! If you're not sure, leave this blank; a maintainer will add them. --> ### Stakeholders @mekarpeles
priority
sitemap returning the sitemap looks like it s no longer available evidence screenshot if possible google search console relevant url details logged in y n y browser type version operating system environment prod dev local prod proposal constraints related files stakeholders mekarpeles
1
829,209
31,859,029,291
IssuesEvent
2023-09-15 09:32:59
shaka-project/shaka-player
https://api.github.com/repos/shaka-project/shaka-player
reopened
WebVTT wrong writing-mode in nested cues
type: bug priority: P2 component: WebVTT
**Have you read the [FAQ](https://bit.ly/ShakaFAQ) and checked for duplicate open issues?** Yes. **What version of Shaka Player are you using?** 4.3.0 **Can you reproduce the issue with our latest release version?** Yes. **Can you reproduce the issue with the latest code from `main`?** Yes. **Are you using the demo app or your own custom app?** My own app. **If custom app, can you reproduce the issue using our demo app?** Not tried. **What browser and OS are you using?** Chrome 107.0.5304.107 (Build officiel) (64 bits) Windows Windows 10 Professionnel 21H2 **What did you do?** I played a video with side-car WebVTT subtitles that have `vertical-lr` writing-mode. **What did you expect to happen?** The subtitles should be written vertically. **What actually happened?** They're displayed horizontally, as if the writing-mode was `horizontal-tb`. When I look at the cues just after they're parsed, I see this: ![image](https://user-images.githubusercontent.com/10497009/204292838-8d97a860-1bb5-4faf-b8ef-0bf230994fff.png) The root cue (or whatever you want to call it) has the correct `vertical-lr` writing-mode, but the nested cue has `horizontal-tb` writing-mode. I think all nested cues should have the same writing-mode as their root cue.
1.0
WebVTT wrong writing-mode in nested cues - **Have you read the [FAQ](https://bit.ly/ShakaFAQ) and checked for duplicate open issues?** Yes. **What version of Shaka Player are you using?** 4.3.0 **Can you reproduce the issue with our latest release version?** Yes. **Can you reproduce the issue with the latest code from `main`?** Yes. **Are you using the demo app or your own custom app?** My own app. **If custom app, can you reproduce the issue using our demo app?** Not tried. **What browser and OS are you using?** Chrome 107.0.5304.107 (Build officiel) (64 bits) Windows Windows 10 Professionnel 21H2 **What did you do?** I played a video with side-car WebVTT subtitles that have `vertical-lr` writing-mode. **What did you expect to happen?** The subtitles should be written vertically. **What actually happened?** They're displayed horizontally, as if the writing-mode was `horizontal-tb`. When I look at the cues just after they're parsed, I see this: ![image](https://user-images.githubusercontent.com/10497009/204292838-8d97a860-1bb5-4faf-b8ef-0bf230994fff.png) The root cue (or whatever you want to call it) has the correct `vertical-lr` writing-mode, but the nested cue has `horizontal-tb` writing-mode. I think all nested cues should have the same writing-mode as their root cue.
priority
webvtt wrong writing mode in nested cues have you read the and checked for duplicate open issues yes what version of shaka player are you using can you reproduce the issue with our latest release version yes can you reproduce the issue with the latest code from main yes are you using the demo app or your own custom app my own app if custom app can you reproduce the issue using our demo app not tried what browser and os are you using chrome build officiel bits windows windows professionnel what did you do i played a video with side car webvtt subtitles that have vertical lr writing mode what did you expect to happen the subtitles should be written vertically what actually happened they re displayed horizontally as if the writing mode was horizontal tb when i look at the cues just after they re parsed i see this the root cue or whatever you want to call it has the correct vertical lr writing mode but the nested cue has horizontal tb writing mode i think all nested cues should have the same writing mode as their root cue
1
27,031
7,889,710,075
IssuesEvent
2018-06-28 05:58:02
ShaikASK/Testing
https://api.github.com/repos/ShaikASK/Testing
opened
UAT : Candidate Module : User is able to navigate in "Dashboard" screen without signing the "Offer Letter"
Candidate Dashboard Candidate Module Defect P1 Release #3 Build 7
Steps : Sign in as Candidate credentials Navigate to "Introduction" page Click on "Dashboard" from side menu Experienced Behaviour : Observed that user is able to navigate in "Dashboard" screen without signing the "Offer Letter" Expected Behaviour : Ensure that application should not allow the user to navigate in "Dashboard" screen without signing the "Offer Letter" and "Dashboard" should be disable until user signs the "Offer Letter"
1.0
UAT : Candidate Module : User is able to navigate in "Dashboard" screen without signing the "Offer Letter" - Steps : Sign in as Candidate credentials Navigate to "Introduction" page Click on "Dashboard" from side menu Experienced Behaviour : Observed that user is able to navigate in "Dashboard" screen without signing the "Offer Letter" Expected Behaviour : Ensure that application should not allow the user to navigate in "Dashboard" screen without signing the "Offer Letter" and "Dashboard" should be disable until user signs the "Offer Letter"
non_priority
uat candidate module user is able to navigate in dashboard screen without signing the offer letter steps sign in as candidate credentials navigate to introduction page click on dashboard from side menu experienced behaviour observed that user is able to navigate in dashboard screen without signing the offer letter expected behaviour ensure that application should not allow the user to navigate in dashboard screen without signing the offer letter and dashboard should be disable until user signs the offer letter
0
678,893
23,214,909,479
IssuesEvent
2022-08-02 13:22:52
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
closed
www.pandora.com - video or audio doesn't play
browser-firefox priority-normal engine-gecko
<!-- @browser: Firefox 103.0 --> <!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:103.0) Gecko/20100101 Firefox/103.0 --> <!-- @reported_with: unknown --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/108296 --> **URL**: https://www.pandora.com **Browser / Version**: Firefox 103.0 **Operating System**: Windows 10 **Tested Another Browser**: Yes Chrome **Problem type**: Video or audio doesn't play **Description**: There is no audio **Steps to Reproduce**: Appears to be playing but no sound <details> <summary>Browser Configuration</summary> <ul> <li>None</li> </ul> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
1.0
www.pandora.com - video or audio doesn't play - <!-- @browser: Firefox 103.0 --> <!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:103.0) Gecko/20100101 Firefox/103.0 --> <!-- @reported_with: unknown --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/108296 --> **URL**: https://www.pandora.com **Browser / Version**: Firefox 103.0 **Operating System**: Windows 10 **Tested Another Browser**: Yes Chrome **Problem type**: Video or audio doesn't play **Description**: There is no audio **Steps to Reproduce**: Appears to be playing but no sound <details> <summary>Browser Configuration</summary> <ul> <li>None</li> </ul> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
priority
video or audio doesn t play url browser version firefox operating system windows tested another browser yes chrome problem type video or audio doesn t play description there is no audio steps to reproduce appears to be playing but no sound browser configuration none from with ❤️
1
21,455
6,153,059,942
IssuesEvent
2017-06-28 09:01:19
src-d/go-kallax
https://api.github.com/repos/src-d/go-kallax
closed
PK issues
area/codegen enhancement
Hello! I'd like to use kallax in my projects, however, I have some issues with doing this. I prefer to have my models and store layer in different packages, I would like not to have kallax to import in the models package. I do this by having a new type in the store package for every model I need, for example: ```go package models type Foo struct { ID int `pk:"autoincr"` Bar string Baz string } ``` ```go package kxstore import ( kallax "gopkg.in/src-d/go-kallax.v1" "my-project/models" ) type Foo struct { kallax.Model `table:"foos"` models.Foo } ``` When running `kallax gen`, i get the following message: `kallax: model Foo has no primary key defined`. Also, some of my models don't have a primary key, does kallax not support models without ids?
1.0
PK issues - Hello! I'd like to use kallax in my projects, however, I have some issues with doing this. I prefer to have my models and store layer in different packages, I would like not to have kallax to import in the models package. I do this by having a new type in the store package for every model I need, for example: ```go package models type Foo struct { ID int `pk:"autoincr"` Bar string Baz string } ``` ```go package kxstore import ( kallax "gopkg.in/src-d/go-kallax.v1" "my-project/models" ) type Foo struct { kallax.Model `table:"foos"` models.Foo } ``` When running `kallax gen`, i get the following message: `kallax: model Foo has no primary key defined`. Also, some of my models don't have a primary key, does kallax not support models without ids?
non_priority
pk issues hello i d like to use kallax in my projects however i have some issues with doing this i prefer to have my models and store layer in different packages i would like not to have kallax to import in the models package i do this by having a new type in the store package for every model i need for example go package models type foo struct id int pk autoincr bar string baz string go package kxstore import kallax gopkg in src d go kallax my project models type foo struct kallax model table foos models foo when running kallax gen i get the following message kallax model foo has no primary key defined also some of my models don t have a primary key does kallax not support models without ids
0
134,967
12,643,144,409
IssuesEvent
2020-06-16 09:19:26
crate/crate-howtos
https://api.github.com/repos/crate/crate-howtos
opened
the status of the cr8 tool is not clear
documentation enhancement
### Documentation feedback <!--Please do not edit or remove the following information --> - Page title: Testing inserts performance - Page URL: https://crate.io/docs/crate/howtos/en/latest/performance/inserts/testing.html - Source: https://github.com/crate/crate-howtos/blob/master/docs/performance/inserts/testing.rst --- <!-- Please add your comments here --> the status of the cr8 tool is not clear. cr8 is not maintained by cr8, it is maintained by @mfussenegger. we should add a note with the appropriate caveat
1.0
the status of the cr8 tool is not clear - ### Documentation feedback <!--Please do not edit or remove the following information --> - Page title: Testing inserts performance - Page URL: https://crate.io/docs/crate/howtos/en/latest/performance/inserts/testing.html - Source: https://github.com/crate/crate-howtos/blob/master/docs/performance/inserts/testing.rst --- <!-- Please add your comments here --> the status of the cr8 tool is not clear. cr8 is not maintained by cr8, it is maintained by @mfussenegger. we should add a note with the appropriate caveat
non_priority
the status of the tool is not clear documentation feedback page title testing inserts performance page url source the status of the tool is not clear is not maintained by it is maintained by mfussenegger we should add a note with the appropriate caveat
0
107,307
11,525,841,505
IssuesEvent
2020-02-15 11:23:25
google/jax
https://api.github.com/repos/google/jax
closed
Autodiff Cookbook Docs Out of Sync
documentation
In second-to last part of the section [Jacobians and Hessians using jacfwd and jacrev](https://jax.readthedocs.io/en/latest/notebooks/autodiff_cookbook.html#Jacobians-and-Hessians-using-jacfwd-and-jacrev), the readthedocs markdown is not rendering correctly, namely there is a list that is being rendered inline like: "This shape makes sense: if we start with a function f:Rn→Rm, then at a point x∈Rn we expect to get the shapes * f(x)∈Rm, the value of f at x, * ∂f(x)∈Rm×n, the Jacobian matrix at x, * ∂2f(x)∈Rm×n×n, the Hessian at x, and so on." In the checked in ipython notebook, [this same section](https://github.com/google/jax/blob/master/docs/notebooks/autodiff_cookbook.ipynb) is rendered correctly as: " This shape makes sense: if we start with a function $f : \mathbb{R}^n \to \mathbb{R}^m$, then at a point $x \in \mathbb{R}^n$ we expect to get the shapes $f(x) \in \mathbb{R}^m$, the value of $f$ at $x$, $\partial f(x) \in \mathbb{R}^{m \times n}$, the Jacobian matrix at $x$, $\partial^2 f(x) \in \mathbb{R}^{m \times n \times n}$, the Hessian at $x$, and so on. " I don't know if readthedocs uses a different markdown parser than ipython or if this is just a case of out of date docs, but figured it would be useful to flag.
1.0
Autodiff Cookbook Docs Out of Sync - In second-to last part of the section [Jacobians and Hessians using jacfwd and jacrev](https://jax.readthedocs.io/en/latest/notebooks/autodiff_cookbook.html#Jacobians-and-Hessians-using-jacfwd-and-jacrev), the readthedocs markdown is not rendering correctly, namely there is a list that is being rendered inline like: "This shape makes sense: if we start with a function f:Rn→Rm, then at a point x∈Rn we expect to get the shapes * f(x)∈Rm, the value of f at x, * ∂f(x)∈Rm×n, the Jacobian matrix at x, * ∂2f(x)∈Rm×n×n, the Hessian at x, and so on." In the checked in ipython notebook, [this same section](https://github.com/google/jax/blob/master/docs/notebooks/autodiff_cookbook.ipynb) is rendered correctly as: " This shape makes sense: if we start with a function $f : \mathbb{R}^n \to \mathbb{R}^m$, then at a point $x \in \mathbb{R}^n$ we expect to get the shapes $f(x) \in \mathbb{R}^m$, the value of $f$ at $x$, $\partial f(x) \in \mathbb{R}^{m \times n}$, the Jacobian matrix at $x$, $\partial^2 f(x) \in \mathbb{R}^{m \times n \times n}$, the Hessian at $x$, and so on. " I don't know if readthedocs uses a different markdown parser than ipython or if this is just a case of out of date docs, but figured it would be useful to flag.
non_priority
autodiff cookbook docs out of sync in second to last part of the section the readthedocs markdown is not rendering correctly namely there is a list that is being rendered inline like this shape makes sense if we start with a function f rn→rm then at a point x∈rn we expect to get the shapes f x ∈rm the value of f at x ∂f x ∈rm×n the jacobian matrix at x ∂ x ∈rm×n×n the hessian at x and so on in the checked in ipython notebook is rendered correctly as this shape makes sense if we start with a function f mathbb r n to mathbb r m then at a point x in mathbb r n we expect to get the shapes f x in mathbb r m the value of f at x partial f x in mathbb r m times n the jacobian matrix at x partial f x in mathbb r m times n times n the hessian at x and so on i don t know if readthedocs uses a different markdown parser than ipython or if this is just a case of out of date docs but figured it would be useful to flag
0
412,502
12,043,167,049
IssuesEvent
2020-04-14 11:56:56
guidesmiths/cybersecurity-handbook
https://api.github.com/repos/guidesmiths/cybersecurity-handbook
closed
OWASP TOP 10 Refactor
content priority-low
### TODO: - [ ] Review tools - [ ] Add new tools referenced in other sections - [ ] Add nodejs examples - [ ] Add videos - [ ] Improve and curate content
1.0
OWASP TOP 10 Refactor - ### TODO: - [ ] Review tools - [ ] Add new tools referenced in other sections - [ ] Add nodejs examples - [ ] Add videos - [ ] Improve and curate content
priority
owasp top refactor todo review tools add new tools referenced in other sections add nodejs examples add videos improve and curate content
1
210,554
16,374,832,931
IssuesEvent
2021-05-15 22:09:04
Mustard2/MustardUI
https://api.github.com/repos/Mustard2/MustardUI
closed
Outdated documentation
bug documentation
The documentation is outdated. Missing stuffs: - how to add custom properties - how to add particle physics hair in the UI
1.0
Outdated documentation - The documentation is outdated. Missing stuffs: - how to add custom properties - how to add particle physics hair in the UI
non_priority
outdated documentation the documentation is outdated missing stuffs how to add custom properties how to add particle physics hair in the ui
0
49,269
6,186,457,099
IssuesEvent
2017-07-04 02:32:10
18F/nsf-sbir
https://api.github.com/repos/18F/nsf-sbir
closed
Landing page - hero image update & caption
design feedback - critical Priority
I received the updated image from Sigray and saved it here: https://github.com/18F/nsf-sbir/blob/dev/assets/img/hero_sigray_seedling_alone.jpg "Sigray is revolutionizing laboratory x-ray research equipment with patented innovations to deliver synchrotron performance. Pictured is a microscopic image of the elements within a hyperaccumulating seedling" Here's their page: http://www.sigray.com/ If you update the caption, please let me know and I'll send it to the company for approval.
1.0
Landing page - hero image update & caption - I received the updated image from Sigray and saved it here: https://github.com/18F/nsf-sbir/blob/dev/assets/img/hero_sigray_seedling_alone.jpg "Sigray is revolutionizing laboratory x-ray research equipment with patented innovations to deliver synchrotron performance. Pictured is a microscopic image of the elements within a hyperaccumulating seedling" Here's their page: http://www.sigray.com/ If you update the caption, please let me know and I'll send it to the company for approval.
non_priority
landing page hero image update caption i received the updated image from sigray and saved it here sigray is revolutionizing laboratory x ray research equipment with patented innovations to deliver synchrotron performance pictured is a microscopic image of the elements within a hyperaccumulating seedling here s their page if you update the caption please let me know and i ll send it to the company for approval
0
570,266
17,023,076,078
IssuesEvent
2021-07-03 00:16:39
tomhughes/trac-tickets
https://api.github.com/repos/tomhughes/trac-tickets
closed
street segments with the same name should be merged into streets
Component: api Priority: major Resolution: invalid Type: enhancement
**[Submitted to the original trac issue database at 10.36am, Friday, 11th November 2005]** Not sure if this is a client issue or server issue, but it should definitely be looked into. Ben's Tiger import script did this for the XML-RPC API, but I'm not sure how streets are handled in the REST API.
1.0
street segments with the same name should be merged into streets - **[Submitted to the original trac issue database at 10.36am, Friday, 11th November 2005]** Not sure if this is a client issue or server issue, but it should definitely be looked into. Ben's Tiger import script did this for the XML-RPC API, but I'm not sure how streets are handled in the REST API.
priority
street segments with the same name should be merged into streets not sure if this is a client issue or server issue but it should definitely be looked into ben s tiger import script did this for the xml rpc api but i m not sure how streets are handled in the rest api
1
34,760
7,460,080,549
IssuesEvent
2018-03-30 18:04:12
kerdokullamae/test_koik_issued
https://api.github.com/repos/kerdokullamae/test_koik_issued
closed
Lehitse > Ainese märksõnad ei ole sorteeritud tekstiliselt
C: AIS P: highest R: duplicate T: defect
**Reported by aadikaljuvee on 27 Apr 2016 09:27 UTC** http://ais2.arhiiv.ee/et/topic/browse#/ (Teistest hiljem lisatud ainesemärksõna)"Fotod" peaks paiknema tähestikuliselt enne "Fotograafe". Aga lehitsemisel, kui ainult 1 kord on vajutatud "näita järgmised 100", siis seda "Fotod" seal pole. "Fotod" jõuab oma õigele kohale alles kõigi märksõnade avamise järel.
1.0
Lehitse > Ainese märksõnad ei ole sorteeritud tekstiliselt - **Reported by aadikaljuvee on 27 Apr 2016 09:27 UTC** http://ais2.arhiiv.ee/et/topic/browse#/ (Teistest hiljem lisatud ainesemärksõna)"Fotod" peaks paiknema tähestikuliselt enne "Fotograafe". Aga lehitsemisel, kui ainult 1 kord on vajutatud "näita järgmised 100", siis seda "Fotod" seal pole. "Fotod" jõuab oma õigele kohale alles kõigi märksõnade avamise järel.
non_priority
lehitse ainese märksõnad ei ole sorteeritud tekstiliselt reported by aadikaljuvee on apr utc teistest hiljem lisatud ainesemärksõna fotod peaks paiknema tähestikuliselt enne fotograafe aga lehitsemisel kui ainult kord on vajutatud näita järgmised siis seda fotod seal pole fotod jõuab oma õigele kohale alles kõigi märksõnade avamise järel
0
61,290
14,965,281,723
IssuesEvent
2021-01-27 13:10:55
eventespresso/barista
https://api.github.com/repos/eventespresso/barista
closed
Prevent Duplicate SASS Imports
C: build-process 🔨 D: Packages 📦 P2: HIGH priority 😮 S:1 new 👶🏻 T: bug 🐞
Seems to be a fairly common issue for ppl using complex SASS setups. Here's a thread regarding the issue with a possible fix: https://github.com/webpack-contrib/sass-loader/issues/145 # OOPS that link above is for webpack 🤦🏻‍♂️
1.0
Prevent Duplicate SASS Imports - Seems to be a fairly common issue for ppl using complex SASS setups. Here's a thread regarding the issue with a possible fix: https://github.com/webpack-contrib/sass-loader/issues/145 # OOPS that link above is for webpack 🤦🏻‍♂️
non_priority
prevent duplicate sass imports seems to be a fairly common issue for ppl using complex sass setups here s a thread regarding the issue with a possible fix oops that link above is for webpack 🤦🏻‍♂️
0
813,271
30,450,498,316
IssuesEvent
2023-07-16 08:22:10
realpoke/genranks
https://api.github.com/repos/realpoke/genranks
opened
Update my games
Type: Enhancement Priority: Medium Status: Available
Have an update my games button for users that's logged in, so they don't have to wait for the daily update. But can get their games updated instantly.
1.0
Update my games - Have an update my games button for users that's logged in, so they don't have to wait for the daily update. But can get their games updated instantly.
priority
update my games have an update my games button for users that s logged in so they don t have to wait for the daily update but can get their games updated instantly
1
757,006
26,492,421,534
IssuesEvent
2023-01-18 00:31:06
zephyrproject-rtos/zephyr
https://api.github.com/repos/zephyrproject-rtos/zephyr
closed
Settings with FCB backend does not pass test on stm32h743
bug priority: low platform: STM32 Stale
**Describe the bug** I tried to run the `zephyr/tests/subsys/settings/fcb/raw/` on a nucleo h743zi2 but I have a test errors, bus faults and stack overflows: **To Reproduce** Steps to reproduce the behavior: 1. `west build -b nucleo_h743zi -d build_test/fcb zephyr/tests/subsys/settings/fcb/raw/` 2. `west flash -d build_test/fcb` 3. See error on serial output. **Expected behavior** All tests should pass **Impact** Settings with FCB backend cannot be used on stm32h7 (NVS is not an option either: #45031) **Logs and console output** ``` *** Booting Zephyr OS build zephyr-v3.1.0-1919-g8a2eed6cbdbc *** Running TESTSUITE test_config_fcb =================================================================== START - test_config_empty_lookups PASS - test_config_empty_lookups in 0.1 seconds =================================================================== START - test_config_insert PASS - test_config_insert in 0.1 seconds =================================================================== START - test_config_getset_unknown PASS - test_config_getset_unknown in 0.1 seconds =================================================================== START - test_config_getset_int PASS - test_config_getset_int in 0.1 seconds =================================================================== START - test_config_getset_int64 PASS - test_config_getset_int64 in 0.1 seconds =================================================================== START - test_config_commit PASS - test_config_commit in 0.1 seconds =================================================================== START - test_settings_check_target Assertion failed at WEST_TOPDIR/zephyr/tests/subsys/settings/fcb/src/settings_test_fcb.c:354: test_settings_check_target: (wbs <= 16 is false) Flash driver is not compatible with the settings fcb-backend FAIL - test_settings_check_target in 0.19 seconds =================================================================== START - test_config_save_fcb_unaligned E: ***** BUS FAULT ***** E: Precise data bus error E: BFAR Address: 0x934b2921 E: r0/a1: 0x00000022 r1/a2: 0x00000020 r2/a3: 0x24008870 E: r3/a4: 0x00000002 r12/ip: 0x00000732 r14/lr: 0x08007a1d E: xpsr: 0x21000000 E: Faulting instruction address (r15/pc): 0x08007fb4 E: >>> ZEPHYR FATAL ERROR 0: CPU exception on CPU 0 E: Current thread: 0x240002c8 (test_config_save_fcb_unaligned) E: Halting system ``` **Environment (please complete the following information):** - Linux - zephyr-sdk-0.14.2 - 8a2eed6cbdbc82550d3b39e8ba808102e82c3fb8 **Additional context** I fixed the first bus fault by increasing the buffer used to read/write to flash from 16 bytes to 32 bytes [settings_fcb_h7.diff.txt](https://github.com/zephyrproject-rtos/zephyr/files/9082889/settings_fcb_h7.diff.txt) Then the first crash appears on `test_config_save_2_fcb`: ``` START - test_config_save_2_fcb E: ***** MPU FAULT ***** E: Stacking error (context area might be not valid) E: Data Access Violation E: MMFAR Address: 0x240086fc E: r0/a1: 0x5d9e865c r1/a2: 0xc96aa6ea r2/a3: 0xf624b53f E: r3/a4: 0x9baa9bd7 r12/ip: 0xcf8fb769 r14/lr: 0xcc5f6b2c E: xpsr: 0x24008600 E: Faulting instruction address (r15/pc): 0xe1af64c5 E: >>> ZEPHYR FATAL ERROR 2: Stack overflow on CPU 0 E: Current thread: 0x240002c8 (test_config_save_2_fcb) E: Halting system ```
1.0
Settings with FCB backend does not pass test on stm32h743 - **Describe the bug** I tried to run the `zephyr/tests/subsys/settings/fcb/raw/` on a nucleo h743zi2 but I have a test errors, bus faults and stack overflows: **To Reproduce** Steps to reproduce the behavior: 1. `west build -b nucleo_h743zi -d build_test/fcb zephyr/tests/subsys/settings/fcb/raw/` 2. `west flash -d build_test/fcb` 3. See error on serial output. **Expected behavior** All tests should pass **Impact** Settings with FCB backend cannot be used on stm32h7 (NVS is not an option either: #45031) **Logs and console output** ``` *** Booting Zephyr OS build zephyr-v3.1.0-1919-g8a2eed6cbdbc *** Running TESTSUITE test_config_fcb =================================================================== START - test_config_empty_lookups PASS - test_config_empty_lookups in 0.1 seconds =================================================================== START - test_config_insert PASS - test_config_insert in 0.1 seconds =================================================================== START - test_config_getset_unknown PASS - test_config_getset_unknown in 0.1 seconds =================================================================== START - test_config_getset_int PASS - test_config_getset_int in 0.1 seconds =================================================================== START - test_config_getset_int64 PASS - test_config_getset_int64 in 0.1 seconds =================================================================== START - test_config_commit PASS - test_config_commit in 0.1 seconds =================================================================== START - test_settings_check_target Assertion failed at WEST_TOPDIR/zephyr/tests/subsys/settings/fcb/src/settings_test_fcb.c:354: test_settings_check_target: (wbs <= 16 is false) Flash driver is not compatible with the settings fcb-backend FAIL - test_settings_check_target in 0.19 seconds =================================================================== START - test_config_save_fcb_unaligned E: ***** BUS FAULT ***** E: Precise data bus error E: BFAR Address: 0x934b2921 E: r0/a1: 0x00000022 r1/a2: 0x00000020 r2/a3: 0x24008870 E: r3/a4: 0x00000002 r12/ip: 0x00000732 r14/lr: 0x08007a1d E: xpsr: 0x21000000 E: Faulting instruction address (r15/pc): 0x08007fb4 E: >>> ZEPHYR FATAL ERROR 0: CPU exception on CPU 0 E: Current thread: 0x240002c8 (test_config_save_fcb_unaligned) E: Halting system ``` **Environment (please complete the following information):** - Linux - zephyr-sdk-0.14.2 - 8a2eed6cbdbc82550d3b39e8ba808102e82c3fb8 **Additional context** I fixed the first bus fault by increasing the buffer used to read/write to flash from 16 bytes to 32 bytes [settings_fcb_h7.diff.txt](https://github.com/zephyrproject-rtos/zephyr/files/9082889/settings_fcb_h7.diff.txt) Then the first crash appears on `test_config_save_2_fcb`: ``` START - test_config_save_2_fcb E: ***** MPU FAULT ***** E: Stacking error (context area might be not valid) E: Data Access Violation E: MMFAR Address: 0x240086fc E: r0/a1: 0x5d9e865c r1/a2: 0xc96aa6ea r2/a3: 0xf624b53f E: r3/a4: 0x9baa9bd7 r12/ip: 0xcf8fb769 r14/lr: 0xcc5f6b2c E: xpsr: 0x24008600 E: Faulting instruction address (r15/pc): 0xe1af64c5 E: >>> ZEPHYR FATAL ERROR 2: Stack overflow on CPU 0 E: Current thread: 0x240002c8 (test_config_save_2_fcb) E: Halting system ```
priority
settings with fcb backend does not pass test on describe the bug i tried to run the zephyr tests subsys settings fcb raw on a nucleo but i have a test errors bus faults and stack overflows to reproduce steps to reproduce the behavior west build b nucleo d build test fcb zephyr tests subsys settings fcb raw west flash d build test fcb see error on serial output expected behavior all tests should pass impact settings with fcb backend cannot be used on nvs is not an option either logs and console output booting zephyr os build zephyr running testsuite test config fcb start test config empty lookups pass test config empty lookups in seconds start test config insert pass test config insert in seconds start test config getset unknown pass test config getset unknown in seconds start test config getset int pass test config getset int in seconds start test config getset pass test config getset in seconds start test config commit pass test config commit in seconds start test settings check target assertion failed at west topdir zephyr tests subsys settings fcb src settings test fcb c test settings check target wbs is false flash driver is not compatible with the settings fcb backend fail test settings check target in seconds start test config save fcb unaligned e bus fault e precise data bus error e bfar address e e ip lr e xpsr e faulting instruction address pc e zephyr fatal error cpu exception on cpu e current thread test config save fcb unaligned e halting system environment please complete the following information linux zephyr sdk additional context i fixed the first bus fault by increasing the buffer used to read write to flash from bytes to bytes then the first crash appears on test config save fcb start test config save fcb e mpu fault e stacking error context area might be not valid e data access violation e mmfar address e e ip lr e xpsr e faulting instruction address pc e zephyr fatal error stack overflow on cpu e current thread test config save fcb e halting system
1
10,298
2,622,141,300
IssuesEvent
2015-03-04 00:02:07
byzhang/spserver
https://api.github.com/repos/byzhang/spserver
opened
请问如何发送文件?
auto-migrated Priority-Medium Type-Defect
``` 例子里的是字符串数据,如果发送文件如何做? ``` Original issue reported on code.google.com by `[email protected]` on 12 Apr 2011 at 4:46
1.0
请问如何发送文件? - ``` 例子里的是字符串数据,如果发送文件如何做? ``` Original issue reported on code.google.com by `[email protected]` on 12 Apr 2011 at 4:46
non_priority
请问如何发送文件 例子里的是字符串数据 如果发送文件如何做 original issue reported on code google com by xiejie com on apr at
0
280,418
24,304,004,982
IssuesEvent
2022-09-29 15:50:40
yugabyte/yugabyte-db
https://api.github.com/repos/yugabyte/yugabyte-db
closed
[YSQL] flaky test: org.yb.pgsql.TestPgRegressExtension.testPgRegressExtension
kind/bug area/ysql kind/failing-test priority/high
Jira Link: [DB-3649](https://yugabyte.atlassian.net/browse/DB-3649) ### Description https://detective-gcp.dev.yugabyte.com/stability/test?analyze_trends=true&branch=master&build_type=all&class=org.yb.pgsql.TestPgRegressExtension&fail_tag=all&name=testPgRegressExtension&platform=linux seems flaky in several build types, but clearly broken in ASAN since https://github.com/yugabyte/yugabyte-db/commit/388ae9def67cd91527b5968f58b7fe1e8b6e7766
1.0
[YSQL] flaky test: org.yb.pgsql.TestPgRegressExtension.testPgRegressExtension - Jira Link: [DB-3649](https://yugabyte.atlassian.net/browse/DB-3649) ### Description https://detective-gcp.dev.yugabyte.com/stability/test?analyze_trends=true&branch=master&build_type=all&class=org.yb.pgsql.TestPgRegressExtension&fail_tag=all&name=testPgRegressExtension&platform=linux seems flaky in several build types, but clearly broken in ASAN since https://github.com/yugabyte/yugabyte-db/commit/388ae9def67cd91527b5968f58b7fe1e8b6e7766
non_priority
flaky test org yb pgsql testpgregressextension testpgregressextension jira link description seems flaky in several build types but clearly broken in asan since
0
250,509
18,892,285,135
IssuesEvent
2021-11-15 14:27:36
boom210232/ku-polls
https://api.github.com/repos/boom210232/ku-polls
reopened
[Info] Readme for instructor about detail.
documentation
# MY ANNOUNCEMENT - Iteration2 need to merge again because flake8 style advice something that make some function can't use. - Builds in travis-ci have been temporarily disabled for private and public repositories due to a negative credit balance.
1.0
[Info] Readme for instructor about detail. - # MY ANNOUNCEMENT - Iteration2 need to merge again because flake8 style advice something that make some function can't use. - Builds in travis-ci have been temporarily disabled for private and public repositories due to a negative credit balance.
non_priority
readme for instructor about detail my announcement need to merge again because style advice something that make some function can t use builds in travis ci have been temporarily disabled for private and public repositories due to a negative credit balance
0
474,341
13,657,405,795
IssuesEvent
2020-09-28 05:39:43
Azure/autorest.az
https://api.github.com/repos/Azure/autorest.az
closed
Expose key of a named collection as a parameter
priority_p2_nice-to-have type_feature-request
In the `create_or_update` operation of Blueprint, [resourceGroups](https://github.com/Azure/azure-rest-api-specs/blob/2f48d306ba63a7ea3c19bd122a274dc465a84e4c/specification/blueprint/resource-manager/Microsoft.Blueprint/preview/2018-11-01-preview/blueprintDefinition.json#L847) is added as a [named collection](https://github.com/Azure/azure-rest-api-specs/blob/2f48d306ba63a7ea3c19bd122a274dc465a84e4c/specification/blueprint/resource-manager/Microsoft.Blueprint/preview/2018-11-01-preview/blueprintDefinition.json#L1200) which will be mapped as a dictionary in python. Its key would be used to get/update/delete a resource group artifact in the blueprint. In Azure Portal, this key is not exposed and will be autogenerated as 'ResourceGroup', 'ResourceGroup2' e.t.c. When users operate on the created resource group artifact on GUI, the underlying key will be used when sending requests to the service. In Azure CLI, we need to let users be able to name this key so they can later use this key to operate the created resource group. So we want to expose the key parameter. I'm not sure how common named collection is used in swagger. This could be of low priority if it's not a common use.
1.0
Expose key of a named collection as a parameter - In the `create_or_update` operation of Blueprint, [resourceGroups](https://github.com/Azure/azure-rest-api-specs/blob/2f48d306ba63a7ea3c19bd122a274dc465a84e4c/specification/blueprint/resource-manager/Microsoft.Blueprint/preview/2018-11-01-preview/blueprintDefinition.json#L847) is added as a [named collection](https://github.com/Azure/azure-rest-api-specs/blob/2f48d306ba63a7ea3c19bd122a274dc465a84e4c/specification/blueprint/resource-manager/Microsoft.Blueprint/preview/2018-11-01-preview/blueprintDefinition.json#L1200) which will be mapped as a dictionary in python. Its key would be used to get/update/delete a resource group artifact in the blueprint. In Azure Portal, this key is not exposed and will be autogenerated as 'ResourceGroup', 'ResourceGroup2' e.t.c. When users operate on the created resource group artifact on GUI, the underlying key will be used when sending requests to the service. In Azure CLI, we need to let users be able to name this key so they can later use this key to operate the created resource group. So we want to expose the key parameter. I'm not sure how common named collection is used in swagger. This could be of low priority if it's not a common use.
priority
expose key of a named collection as a parameter in the create or update operation of blueprint is added as a which will be mapped as a dictionary in python its key would be used to get update delete a resource group artifact in the blueprint in azure portal this key is not exposed and will be autogenerated as resourcegroup e t c when users operate on the created resource group artifact on gui the underlying key will be used when sending requests to the service in azure cli we need to let users be able to name this key so they can later use this key to operate the created resource group so we want to expose the key parameter i m not sure how common named collection is used in swagger this could be of low priority if it s not a common use
1
101,330
8,785,402,314
IssuesEvent
2018-12-20 12:53:38
elastic/elasticsearch
https://api.github.com/repos/elastic/elasticsearch
closed
[CI][ML] MlDistributedFailureIT.testLoseDedicatedMasterNode randomly fails on feature-jindex-master branch
:ml >test-failure v7.0.0
This test has been observed to fail occasionally in the `feature-jindex-master` branch. I have not yet managed to reproduce locally. I will shortly be muting the test as we need a green build to merge the branch in master. However, I am raising the issue to ensure we get to the bottom of this failure. Link to failure (one of them): https://elasticsearch-ci.elastic.co/job/elastic+elasticsearch+pull-request-2/2556/console Reproduce with: ``` ./gradlew :x-pack:plugin:ml:internalClusterTest -Dtests.seed=D2A618A38265651F -Dtests.class=org.elasticsearch.xpack.ml.integration.MlDistributedFailureIT -Dtests.method="testLoseDedicatedMasterNode" -Dtests.security.manager=true -Dtests.locale=be-BY -Dtests.timezone=Asia/Katmandu -Dcompiler.java=11 -Druntime.java=8 ``` Failure: ``` > Throwable #1: java.lang.AssertionError > at org.elasticsearch.xpack.ml.integration.MlDistributedFailureIT.lambda$run$15(MlDistributedFailureIT.java:298) > at org.elasticsearch.test.ESTestCase.assertBusy(ESTestCase.java:847) > at org.elasticsearch.test.ESTestCase.assertBusy(ESTestCase.java:821) > at org.elasticsearch.xpack.ml.integration.MlDistributedFailureIT.run(MlDistributedFailureIT.java:292) > at org.elasticsearch.xpack.ml.integration.MlDistributedFailureIT.testLoseDedicatedMasterNode(MlDistributedFailureIT.java:88) > at java.lang.Thread.run(Thread.java:748) > Suppressed: java.lang.AssertionError > at org.elasticsearch.xpack.ml.integration.MlDistributedFailureIT.lambda$run$15(MlDistributedFailureIT.java:298) > at org.elasticsearch.test.ESTestCase.assertBusy(ESTestCase.java:835) > ... 40 more > Suppressed: java.lang.AssertionError > at org.elasticsearch.xpack.ml.integration.MlDistributedFailureIT.lambda$run$15(MlDistributedFailureIT.java:298) > at org.elasticsearch.test.ESTestCase.assertBusy(ESTestCase.java:835) > ... 40 more > Suppressed: java.lang.AssertionError > at org.elasticsearch.xpack.ml.integration.MlDistributedFailureIT.lambda$run$15(MlDistributedFailureIT.java:298) > at org.elasticsearch.test.ESTestCase.assertBusy(ESTestCase.java:835) > ... 40 more > Suppressed: java.lang.AssertionError > at org.elasticsearch.xpack.ml.integration.MlDistributedFailureIT.lambda$run$15(MlDistributedFailureIT.java:298) > at org.elasticsearch.test.ESTestCase.assertBusy(ESTestCase.java:835) > ... 40 more > Suppressed: java.lang.AssertionError > at org.elasticsearch.xpack.ml.integration.MlDistributedFailureIT.lambda$run$15(MlDistributedFailureIT.java:298) > at org.elasticsearch.test.ESTestCase.assertBusy(ESTestCase.java:835) > ... 40 more > Suppressed: java.lang.AssertionError > at org.elasticsearch.xpack.ml.integration.MlDistributedFailureIT.lambda$run$15(MlDistributedFailureIT.java:298) > at org.elasticsearch.test.ESTestCase.assertBusy(ESTestCase.java:835) > ... 40 more > Suppressed: java.lang.AssertionError > at org.elasticsearch.xpack.ml.integration.MlDistributedFailureIT.lambda$run$15(MlDistributedFailureIT.java:298) > at org.elasticsearch.test.ESTestCase.assertBusy(ESTestCase.java:835) > ... 40 more > Suppressed: java.lang.AssertionError > at org.elasticsearch.xpack.ml.integration.MlDistributedFailureIT.lambda$run$15(MlDistributedFailureIT.java:298) > at org.elasticsearch.test.ESTestCase.assertBusy(ESTestCase.java:835) > ... 40 more > Suppressed: java.lang.AssertionError > at org.elasticsearch.xpack.ml.integration.MlDistributedFailureIT.lambda$run$15(MlDistributedFailureIT.java:298) > at org.elasticsearch.test.ESTestCase.assertBusy(ESTestCase.java:835) > ... 40 more > Suppressed: java.lang.AssertionError > at org.elasticsearch.xpack.ml.integration.MlDistributedFailureIT.lambda$run$15(MlDistributedFailureIT.java:298) > at org.elasticsearch.test.ESTestCase.assertBusy(ESTestCase.java:835) > ... 40 more > Suppressed: java.lang.AssertionError > at org.elasticsearch.xpack.ml.integration.MlDistributedFailureIT.lambda$run$15(MlDistributedFailureIT.java:298) > at org.elasticsearch.test.ESTestCase.assertBusy(ESTestCase.java:835) > ... 40 more > Suppressed: java.lang.AssertionError > at org.elasticsearch.xpack.ml.integration.MlDistributedFailureIT.lambda$run$15(MlDistributedFailureIT.java:298) > at org.elasticsearch.test.ESTestCase.assertBusy(ESTestCase.java:835) > ... 40 more > Suppressed: java.lang.AssertionError > at org.elasticsearch.xpack.ml.integration.MlDistributedFailureIT.lambda$run$15(MlDistributedFailureIT.java:298) > at org.elasticsearch.test.ESTestCase.assertBusy(ESTestCase.java:835) > ... 40 moreThrowable #2: java.lang.RuntimeException: Had to resort to force-stopping datafeed, something went wrong? > at org.elasticsearch.xpack.ml.support.BaseMlIntegTestCase.deleteAllDatafeeds(BaseMlIntegTestCase.java:296) > at org.elasticsearch.xpack.ml.support.BaseMlIntegTestCase.cleanupWorkaround(BaseMlIntegTestCase.java:209) > at java.lang.Thread.run(Thread.java:748) > Caused by: java.util.concurrent.ExecutionException: ElasticsearchStatusException[Cannot stop datafeed [data_feed_id] because the datafeed does not have an assigned node. Use force stop to stop the datafeed] > at org.elasticsearch.common.util.concurrent.BaseFuture$Sync.getValue(BaseFuture.java:266) > at org.elasticsearch.common.util.concurrent.BaseFuture$Sync.get(BaseFuture.java:253) > at org.elasticsearch.common.util.concurrent.BaseFuture.get(BaseFuture.java:87) > at org.elasticsearch.xpack.ml.support.BaseMlIntegTestCase.deleteAllDatafeeds(BaseMlIntegTestCase.java:284) > ... 36 more > Caused by: ElasticsearchStatusException[Cannot stop datafeed [data_feed_id] because the datafeed does not have an assigned node. Use force stop to stop the datafeed] > at org.elasticsearch.xpack.core.ml.utils.ExceptionsHelper.conflictStatusException(ExceptionsHelper.java:50) > at org.elasticsearch.xpack.ml.action.TransportStopDatafeedAction.normalStopDatafeed(TransportStopDatafeedAction.java:147) > at org.elasticsearch.xpack.ml.action.TransportStopDatafeedAction.lambda$doExecute$0(TransportStopDatafeedAction.java:130) > at org.elasticsearch.action.ActionListener$1.onResponse(ActionListener.java:60) > at org.elasticsearch.xpack.ml.datafeed.persistence.DatafeedConfigProvider.lambd ```
1.0
[CI][ML] MlDistributedFailureIT.testLoseDedicatedMasterNode randomly fails on feature-jindex-master branch - This test has been observed to fail occasionally in the `feature-jindex-master` branch. I have not yet managed to reproduce locally. I will shortly be muting the test as we need a green build to merge the branch in master. However, I am raising the issue to ensure we get to the bottom of this failure. Link to failure (one of them): https://elasticsearch-ci.elastic.co/job/elastic+elasticsearch+pull-request-2/2556/console Reproduce with: ``` ./gradlew :x-pack:plugin:ml:internalClusterTest -Dtests.seed=D2A618A38265651F -Dtests.class=org.elasticsearch.xpack.ml.integration.MlDistributedFailureIT -Dtests.method="testLoseDedicatedMasterNode" -Dtests.security.manager=true -Dtests.locale=be-BY -Dtests.timezone=Asia/Katmandu -Dcompiler.java=11 -Druntime.java=8 ``` Failure: ``` > Throwable #1: java.lang.AssertionError > at org.elasticsearch.xpack.ml.integration.MlDistributedFailureIT.lambda$run$15(MlDistributedFailureIT.java:298) > at org.elasticsearch.test.ESTestCase.assertBusy(ESTestCase.java:847) > at org.elasticsearch.test.ESTestCase.assertBusy(ESTestCase.java:821) > at org.elasticsearch.xpack.ml.integration.MlDistributedFailureIT.run(MlDistributedFailureIT.java:292) > at org.elasticsearch.xpack.ml.integration.MlDistributedFailureIT.testLoseDedicatedMasterNode(MlDistributedFailureIT.java:88) > at java.lang.Thread.run(Thread.java:748) > Suppressed: java.lang.AssertionError > at org.elasticsearch.xpack.ml.integration.MlDistributedFailureIT.lambda$run$15(MlDistributedFailureIT.java:298) > at org.elasticsearch.test.ESTestCase.assertBusy(ESTestCase.java:835) > ... 40 more > Suppressed: java.lang.AssertionError > at org.elasticsearch.xpack.ml.integration.MlDistributedFailureIT.lambda$run$15(MlDistributedFailureIT.java:298) > at org.elasticsearch.test.ESTestCase.assertBusy(ESTestCase.java:835) > ... 40 more > Suppressed: java.lang.AssertionError > at org.elasticsearch.xpack.ml.integration.MlDistributedFailureIT.lambda$run$15(MlDistributedFailureIT.java:298) > at org.elasticsearch.test.ESTestCase.assertBusy(ESTestCase.java:835) > ... 40 more > Suppressed: java.lang.AssertionError > at org.elasticsearch.xpack.ml.integration.MlDistributedFailureIT.lambda$run$15(MlDistributedFailureIT.java:298) > at org.elasticsearch.test.ESTestCase.assertBusy(ESTestCase.java:835) > ... 40 more > Suppressed: java.lang.AssertionError > at org.elasticsearch.xpack.ml.integration.MlDistributedFailureIT.lambda$run$15(MlDistributedFailureIT.java:298) > at org.elasticsearch.test.ESTestCase.assertBusy(ESTestCase.java:835) > ... 40 more > Suppressed: java.lang.AssertionError > at org.elasticsearch.xpack.ml.integration.MlDistributedFailureIT.lambda$run$15(MlDistributedFailureIT.java:298) > at org.elasticsearch.test.ESTestCase.assertBusy(ESTestCase.java:835) > ... 40 more > Suppressed: java.lang.AssertionError > at org.elasticsearch.xpack.ml.integration.MlDistributedFailureIT.lambda$run$15(MlDistributedFailureIT.java:298) > at org.elasticsearch.test.ESTestCase.assertBusy(ESTestCase.java:835) > ... 40 more > Suppressed: java.lang.AssertionError > at org.elasticsearch.xpack.ml.integration.MlDistributedFailureIT.lambda$run$15(MlDistributedFailureIT.java:298) > at org.elasticsearch.test.ESTestCase.assertBusy(ESTestCase.java:835) > ... 40 more > Suppressed: java.lang.AssertionError > at org.elasticsearch.xpack.ml.integration.MlDistributedFailureIT.lambda$run$15(MlDistributedFailureIT.java:298) > at org.elasticsearch.test.ESTestCase.assertBusy(ESTestCase.java:835) > ... 40 more > Suppressed: java.lang.AssertionError > at org.elasticsearch.xpack.ml.integration.MlDistributedFailureIT.lambda$run$15(MlDistributedFailureIT.java:298) > at org.elasticsearch.test.ESTestCase.assertBusy(ESTestCase.java:835) > ... 40 more > Suppressed: java.lang.AssertionError > at org.elasticsearch.xpack.ml.integration.MlDistributedFailureIT.lambda$run$15(MlDistributedFailureIT.java:298) > at org.elasticsearch.test.ESTestCase.assertBusy(ESTestCase.java:835) > ... 40 more > Suppressed: java.lang.AssertionError > at org.elasticsearch.xpack.ml.integration.MlDistributedFailureIT.lambda$run$15(MlDistributedFailureIT.java:298) > at org.elasticsearch.test.ESTestCase.assertBusy(ESTestCase.java:835) > ... 40 more > Suppressed: java.lang.AssertionError > at org.elasticsearch.xpack.ml.integration.MlDistributedFailureIT.lambda$run$15(MlDistributedFailureIT.java:298) > at org.elasticsearch.test.ESTestCase.assertBusy(ESTestCase.java:835) > ... 40 moreThrowable #2: java.lang.RuntimeException: Had to resort to force-stopping datafeed, something went wrong? > at org.elasticsearch.xpack.ml.support.BaseMlIntegTestCase.deleteAllDatafeeds(BaseMlIntegTestCase.java:296) > at org.elasticsearch.xpack.ml.support.BaseMlIntegTestCase.cleanupWorkaround(BaseMlIntegTestCase.java:209) > at java.lang.Thread.run(Thread.java:748) > Caused by: java.util.concurrent.ExecutionException: ElasticsearchStatusException[Cannot stop datafeed [data_feed_id] because the datafeed does not have an assigned node. Use force stop to stop the datafeed] > at org.elasticsearch.common.util.concurrent.BaseFuture$Sync.getValue(BaseFuture.java:266) > at org.elasticsearch.common.util.concurrent.BaseFuture$Sync.get(BaseFuture.java:253) > at org.elasticsearch.common.util.concurrent.BaseFuture.get(BaseFuture.java:87) > at org.elasticsearch.xpack.ml.support.BaseMlIntegTestCase.deleteAllDatafeeds(BaseMlIntegTestCase.java:284) > ... 36 more > Caused by: ElasticsearchStatusException[Cannot stop datafeed [data_feed_id] because the datafeed does not have an assigned node. Use force stop to stop the datafeed] > at org.elasticsearch.xpack.core.ml.utils.ExceptionsHelper.conflictStatusException(ExceptionsHelper.java:50) > at org.elasticsearch.xpack.ml.action.TransportStopDatafeedAction.normalStopDatafeed(TransportStopDatafeedAction.java:147) > at org.elasticsearch.xpack.ml.action.TransportStopDatafeedAction.lambda$doExecute$0(TransportStopDatafeedAction.java:130) > at org.elasticsearch.action.ActionListener$1.onResponse(ActionListener.java:60) > at org.elasticsearch.xpack.ml.datafeed.persistence.DatafeedConfigProvider.lambd ```
non_priority
mldistributedfailureit testlosededicatedmasternode randomly fails on feature jindex master branch this test has been observed to fail occasionally in the feature jindex master branch i have not yet managed to reproduce locally i will shortly be muting the test as we need a green build to merge the branch in master however i am raising the issue to ensure we get to the bottom of this failure link to failure one of them reproduce with gradlew x pack plugin ml internalclustertest dtests seed dtests class org elasticsearch xpack ml integration mldistributedfailureit dtests method testlosededicatedmasternode dtests security manager true dtests locale be by dtests timezone asia katmandu dcompiler java druntime java failure throwable java lang assertionerror at org elasticsearch xpack ml integration mldistributedfailureit lambda run mldistributedfailureit java at org elasticsearch test estestcase assertbusy estestcase java at org elasticsearch test estestcase assertbusy estestcase java at org elasticsearch xpack ml integration mldistributedfailureit run mldistributedfailureit java at org elasticsearch xpack ml integration mldistributedfailureit testlosededicatedmasternode mldistributedfailureit java at java lang thread run thread java suppressed java lang assertionerror at org elasticsearch xpack ml integration mldistributedfailureit lambda run mldistributedfailureit java at org elasticsearch test estestcase assertbusy estestcase java more suppressed java lang assertionerror at org elasticsearch xpack ml integration mldistributedfailureit lambda run mldistributedfailureit java at org elasticsearch test estestcase assertbusy estestcase java more suppressed java lang assertionerror at org elasticsearch xpack ml integration mldistributedfailureit lambda run mldistributedfailureit java at org elasticsearch test estestcase assertbusy estestcase java more suppressed java lang assertionerror at org elasticsearch xpack ml integration mldistributedfailureit lambda run mldistributedfailureit java at org elasticsearch test estestcase assertbusy estestcase java more suppressed java lang assertionerror at org elasticsearch xpack ml integration mldistributedfailureit lambda run mldistributedfailureit java at org elasticsearch test estestcase assertbusy estestcase java more suppressed java lang assertionerror at org elasticsearch xpack ml integration mldistributedfailureit lambda run mldistributedfailureit java at org elasticsearch test estestcase assertbusy estestcase java more suppressed java lang assertionerror at org elasticsearch xpack ml integration mldistributedfailureit lambda run mldistributedfailureit java at org elasticsearch test estestcase assertbusy estestcase java more suppressed java lang assertionerror at org elasticsearch xpack ml integration mldistributedfailureit lambda run mldistributedfailureit java at org elasticsearch test estestcase assertbusy estestcase java more suppressed java lang assertionerror at org elasticsearch xpack ml integration mldistributedfailureit lambda run mldistributedfailureit java at org elasticsearch test estestcase assertbusy estestcase java more suppressed java lang assertionerror at org elasticsearch xpack ml integration mldistributedfailureit lambda run mldistributedfailureit java at org elasticsearch test estestcase assertbusy estestcase java more suppressed java lang assertionerror at org elasticsearch xpack ml integration mldistributedfailureit lambda run mldistributedfailureit java at org elasticsearch test estestcase assertbusy estestcase java more suppressed java lang assertionerror at org elasticsearch xpack ml integration mldistributedfailureit lambda run mldistributedfailureit java at org elasticsearch test estestcase assertbusy estestcase java more suppressed java lang assertionerror at org elasticsearch xpack ml integration mldistributedfailureit lambda run mldistributedfailureit java at org elasticsearch test estestcase assertbusy estestcase java morethrowable java lang runtimeexception had to resort to force stopping datafeed something went wrong at org elasticsearch xpack ml support basemlintegtestcase deletealldatafeeds basemlintegtestcase java at org elasticsearch xpack ml support basemlintegtestcase cleanupworkaround basemlintegtestcase java at java lang thread run thread java caused by java util concurrent executionexception elasticsearchstatusexception because the datafeed does not have an assigned node use force stop to stop the datafeed at org elasticsearch common util concurrent basefuture sync getvalue basefuture java at org elasticsearch common util concurrent basefuture sync get basefuture java at org elasticsearch common util concurrent basefuture get basefuture java at org elasticsearch xpack ml support basemlintegtestcase deletealldatafeeds basemlintegtestcase java more caused by elasticsearchstatusexception because the datafeed does not have an assigned node use force stop to stop the datafeed at org elasticsearch xpack core ml utils exceptionshelper conflictstatusexception exceptionshelper java at org elasticsearch xpack ml action transportstopdatafeedaction normalstopdatafeed transportstopdatafeedaction java at org elasticsearch xpack ml action transportstopdatafeedaction lambda doexecute transportstopdatafeedaction java at org elasticsearch action actionlistener onresponse actionlistener java at org elasticsearch xpack ml datafeed persistence datafeedconfigprovider lambd
0
264,721
8,318,943,668
IssuesEvent
2018-09-25 15:51:31
swarm-robotics/fordyca
https://api.github.com/repos/swarm-robotics/fordyca
closed
task/345-run-exp-with-new-task-allocation-params
Priority: Major Status: Available Type: Task
Pulling from the matlab function mesh analysis I FINALLY got around to doing
1.0
task/345-run-exp-with-new-task-allocation-params - Pulling from the matlab function mesh analysis I FINALLY got around to doing
priority
task run exp with new task allocation params pulling from the matlab function mesh analysis i finally got around to doing
1
123,841
16,542,304,451
IssuesEvent
2021-05-27 18:28:46
elementary/switchboard-plug-onlineaccounts
https://api.github.com/repos/elementary/switchboard-plug-onlineaccounts
closed
Add an IMAP account
Needs Design Priority: Wishlist
<!-- * Please read and follow these tips: https://elementary.io/docs/code/reference#proposing-design-changes * Be sure to search open and closed issues for duplicates --> We need to be able to add standard IMAP accounts to EDS <!--Please be sure to preview your issue before saving. Thanks!-->
1.0
Add an IMAP account - <!-- * Please read and follow these tips: https://elementary.io/docs/code/reference#proposing-design-changes * Be sure to search open and closed issues for duplicates --> We need to be able to add standard IMAP accounts to EDS <!--Please be sure to preview your issue before saving. Thanks!-->
non_priority
add an imap account please read and follow these tips be sure to search open and closed issues for duplicates we need to be able to add standard imap accounts to eds
0
2,216
2,588,223,766
IssuesEvent
2015-02-17 23:29:19
rlf/uSkyBlock
https://api.github.com/repos/rlf/uSkyBlock
closed
Support UUID
bug feature request ready for test
Currently, the uSkyBlock plugin doesn't correctly handle players. Too many places rely on the username, instead of the UUID. This should be fixed.
1.0
Support UUID - Currently, the uSkyBlock plugin doesn't correctly handle players. Too many places rely on the username, instead of the UUID. This should be fixed.
non_priority
support uuid currently the uskyblock plugin doesn t correctly handle players too many places rely on the username instead of the uuid this should be fixed
0
343,978
30,704,847,247
IssuesEvent
2023-07-27 04:53:34
litentry/litentry-parachain
https://api.github.com/repos/litentry/litentry-parachain
closed
DI tests: identity
I3-high D5-testing
## Context See #1682 ## Task Implement each of the following test steps using the DI interface. Refactor and extract common functionality as needed, and establish a pattern for the remaining test suites. If needed, merge to dev in smaller batches to mitigate conflicts with other developments. Along the way, consider better options for reducing the combinational explosion :P ### Steps - [x] `check user sidechain storage before create` - [x] `Invalid user shielding key` - [x] `set user shielding key` - [x] `check user shielding key from sidechain storage after setUserShieldingKey` - [x] `check idgraph from sidechain storage before linking` - [x] `link identities` - [x] `check IDGraph after LinkIdentity` - [x] `link invalid identities` - [x] `link identities with wrong signature` - [x] `deactivate identities` - [x] `check IDGraph after deactivate identity` - [x] `deactivate prime identity is disallowed` - [x] `activate identity` - [x] `check IDGraph after activate identity` - [ ] `set error user shielding key` - [ ] `exceeding IDGraph limit not allowed` ## Due Diligence - [ ] Create issues for missing test cases from the original II test suite (marked with `TODO` comments in the code)
1.0
DI tests: identity - ## Context See #1682 ## Task Implement each of the following test steps using the DI interface. Refactor and extract common functionality as needed, and establish a pattern for the remaining test suites. If needed, merge to dev in smaller batches to mitigate conflicts with other developments. Along the way, consider better options for reducing the combinational explosion :P ### Steps - [x] `check user sidechain storage before create` - [x] `Invalid user shielding key` - [x] `set user shielding key` - [x] `check user shielding key from sidechain storage after setUserShieldingKey` - [x] `check idgraph from sidechain storage before linking` - [x] `link identities` - [x] `check IDGraph after LinkIdentity` - [x] `link invalid identities` - [x] `link identities with wrong signature` - [x] `deactivate identities` - [x] `check IDGraph after deactivate identity` - [x] `deactivate prime identity is disallowed` - [x] `activate identity` - [x] `check IDGraph after activate identity` - [ ] `set error user shielding key` - [ ] `exceeding IDGraph limit not allowed` ## Due Diligence - [ ] Create issues for missing test cases from the original II test suite (marked with `TODO` comments in the code)
non_priority
di tests identity context see task implement each of the following test steps using the di interface refactor and extract common functionality as needed and establish a pattern for the remaining test suites if needed merge to dev in smaller batches to mitigate conflicts with other developments along the way consider better options for reducing the combinational explosion p steps check user sidechain storage before create invalid user shielding key set user shielding key check user shielding key from sidechain storage after setusershieldingkey check idgraph from sidechain storage before linking link identities check idgraph after linkidentity link invalid identities link identities with wrong signature deactivate identities check idgraph after deactivate identity deactivate prime identity is disallowed activate identity check idgraph after activate identity set error user shielding key exceeding idgraph limit not allowed due diligence create issues for missing test cases from the original ii test suite marked with todo comments in the code
0
252,467
19,027,407,132
IssuesEvent
2021-11-24 06:27:45
DocNow/twarc-csv
https://api.github.com/repos/DocNow/twarc-csv
closed
DataFrameConverter to single tweet
documentation
I have a large set of tweets, and I would like to wrangle and write them to file as I go. DataFrameConverter would be ideal for this, but if I try to pass a tweet to it, I get an error message: `TypeError: process() missing 1 required positional argument: 'objects'`. I followed these instructions: ``` from twarc_csv import DataFrameConverter json_objects = [...] df = DataFrameConverter.process(json_objects) ``` passing converter a tweet or a page scraped as described in examples. What am I doing wrong or can I not use this at all as I would like to?
1.0
DataFrameConverter to single tweet - I have a large set of tweets, and I would like to wrangle and write them to file as I go. DataFrameConverter would be ideal for this, but if I try to pass a tweet to it, I get an error message: `TypeError: process() missing 1 required positional argument: 'objects'`. I followed these instructions: ``` from twarc_csv import DataFrameConverter json_objects = [...] df = DataFrameConverter.process(json_objects) ``` passing converter a tweet or a page scraped as described in examples. What am I doing wrong or can I not use this at all as I would like to?
non_priority
dataframeconverter to single tweet i have a large set of tweets and i would like to wrangle and write them to file as i go dataframeconverter would be ideal for this but if i try to pass a tweet to it i get an error message typeerror process missing required positional argument objects i followed these instructions from twarc csv import dataframeconverter json objects df dataframeconverter process json objects passing converter a tweet or a page scraped as described in examples what am i doing wrong or can i not use this at all as i would like to
0
150,864
19,634,070,037
IssuesEvent
2022-01-08 01:21:55
vipinsun/cactus
https://api.github.com/repos/vipinsun/cactus
opened
CVE-2018-25023 (High) detected in smallvec-0.6.14.crate
security vulnerability
## CVE-2018-25023 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>smallvec-0.6.14.crate</b></p></summary> <p>'Small vector' optimization: store up to a small number of items on the stack</p> <p>Library home page: <a href="https://crates.io/api/v1/crates/smallvec/0.6.14/download">https://crates.io/api/v1/crates/smallvec/0.6.14/download</a></p> <p> Dependency Hierarchy: - hashicorp_vault-1.1.0.crate (Root Library) - reqwest-0.9.24.crate - hyper-0.12.35.crate - tokio-reactor-0.1.12.crate - parking_lot-0.9.0.crate - parking_lot_core-0.6.2.crate - :x: **smallvec-0.6.14.crate** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> An issue was discovered in the smallvec crate before 0.6.13 for Rust. It can create an uninitialized value of any type, including a reference type. <p>Publish Date: 2021-12-27 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-25023>CVE-2018-25023</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: High - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://rustsec.org/advisories/RUSTSEC-2018-0018.html">https://rustsec.org/advisories/RUSTSEC-2018-0018.html</a></p> <p>Release Date: 2021-12-27</p> <p>Fix Resolution: smallvec - 0.6.13</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2018-25023 (High) detected in smallvec-0.6.14.crate - ## CVE-2018-25023 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>smallvec-0.6.14.crate</b></p></summary> <p>'Small vector' optimization: store up to a small number of items on the stack</p> <p>Library home page: <a href="https://crates.io/api/v1/crates/smallvec/0.6.14/download">https://crates.io/api/v1/crates/smallvec/0.6.14/download</a></p> <p> Dependency Hierarchy: - hashicorp_vault-1.1.0.crate (Root Library) - reqwest-0.9.24.crate - hyper-0.12.35.crate - tokio-reactor-0.1.12.crate - parking_lot-0.9.0.crate - parking_lot_core-0.6.2.crate - :x: **smallvec-0.6.14.crate** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> An issue was discovered in the smallvec crate before 0.6.13 for Rust. It can create an uninitialized value of any type, including a reference type. <p>Publish Date: 2021-12-27 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-25023>CVE-2018-25023</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: High - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://rustsec.org/advisories/RUSTSEC-2018-0018.html">https://rustsec.org/advisories/RUSTSEC-2018-0018.html</a></p> <p>Release Date: 2021-12-27</p> <p>Fix Resolution: smallvec - 0.6.13</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_priority
cve high detected in smallvec crate cve high severity vulnerability vulnerable library smallvec crate small vector optimization store up to a small number of items on the stack library home page a href dependency hierarchy hashicorp vault crate root library reqwest crate hyper crate tokio reactor crate parking lot crate parking lot core crate x smallvec crate vulnerable library found in base branch master vulnerability details an issue was discovered in the smallvec crate before for rust it can create an uninitialized value of any type including a reference type publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact high availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution smallvec step up your open source security game with whitesource
0
177,007
28,309,841,446
IssuesEvent
2023-04-10 14:28:04
kookmin-sw/capstone-2023-09
https://api.github.com/repos/kookmin-sw/capstone-2023-09
closed
[관리자] 카테고리 관리 화면 개발
💄 design
## Description 도움받기 도움주기 게시판에서 사용하는 카테고리를 추가및 삭제할수 있는 화면 ## To-do - [ ] 카테고리 조회 화면 - [ ] 카테고리 수정 기능 - [ ] 카테고리 삭제 기능
1.0
[관리자] 카테고리 관리 화면 개발 - ## Description 도움받기 도움주기 게시판에서 사용하는 카테고리를 추가및 삭제할수 있는 화면 ## To-do - [ ] 카테고리 조회 화면 - [ ] 카테고리 수정 기능 - [ ] 카테고리 삭제 기능
non_priority
카테고리 관리 화면 개발 description 도움받기 도움주기 게시판에서 사용하는 카테고리를 추가및 삭제할수 있는 화면 to do 카테고리 조회 화면 카테고리 수정 기능 카테고리 삭제 기능
0
47,388
6,050,920,224
IssuesEvent
2017-06-12 22:15:27
Microsoft/TypeScript
https://api.github.com/repos/Microsoft/TypeScript
closed
Design Meeting Notes, 6/2/2017
Design Notes
# Propagated Inference for Uninstantiated (Free) Type Parameters (#9366) ```ts function compose<A, B, C>(f: (x: A) => B, g: (y: B) => C): (x: A) => C { // ... } let boxElements: (a: string) => { value: string[] } = compose(x => [x], y => { value: y }); ``` Currently this works. Now we ask, what if we wanted `boxElements` to operate on *any* type? ```ts function compose<A, B, C>(f: (x: A) => B, g: (y: B) => C): (x: A) => C { // ... } let boxElements: <T>(a: T) => { value: T[] } = compose(x => [x], y => { value: y }); ``` Currently this doesn't work well - when we get the contextual type for `x`, we get it from the *erased* signature of `<T>(a: T) => { value: T[] }`, which is really just `(a: any) => { value: any[] }`. We will have a PR that will help fix this. Q: What about constraints? A: Constraints will be carried through. We will still have some problems with compositional patterns *without* explicit types. For example: ```ts declare function compose<A, B, C>(f: (x: A) => B, g: (y: B) => C): (x: A) => C; declare function box<T>(x: T): { value: T }; declare function list<U>(x: U): U[]; compose(list, box); ``` The problem is that TypeScript draws the following inferences: Type Parameter | Inferences ---------------|----------- `A` | `U` `B` | `U[]`, `T` `C` | `{ value: T }` The way we walk through arguments (simple and left-to-right) and draw inferences simply isn't sufficient - we need to find a way to *unify* these type parameters. But it's not just a matter of throwing unification at the type system. So the current question is how do we introduce some unification to the current process. Perhaps it will be a "last resort" process. And this will be exploratory work. @gcnew did do great work, but pointed out many of the difficulties and roadblocks you can run into. # Changing default target to ES5 - Very strange that we wouldn't simply move the the higher one? - Why take a breaking change when we'll need to make another breaking change in the future? - Are there a lot of people running into this problem? - Doesn't appear that a lot of people have been complaining about it. # Pure annotation in downlevel emits - Currently, Uglify doesn't understand when our class emit. - It would be great if Uglify could operate on the ES2015 code. - We just want to tell other tools it's a class - onus of determining side-effects being on TypeScript is probably more questionable than an optimizer doing so. - Decision: emit JSDoc `@class` comment unconditionally. # Lib reference directive (#15780) - `/// <reference lib="name" />` - Problem: things like corejs on DefinitelyTyped conflict with compiler defaults. - This means could simply be reduced to `/// <reference lib="es2015.d.ts" />`. - Q: What about `/// <reference no-default-lib="true"/>`? - This PR ignores any `/// <reference no-default-lib="true"/>` comments. - Could also do `/// <reference lib="..." />` in `lib.d.ts` as well. - Let's do that. - What about Daniel's lib versioning idea of publishing to `@types`? - Example: DOM APIs require new logic - suddenly breaks certain builds. - People would be able to go to a specific version of a lib if they got broken, or just lock down all their dependencies. - Also might allow us to bring `node.d.ts` in.
1.0
Design Meeting Notes, 6/2/2017 - # Propagated Inference for Uninstantiated (Free) Type Parameters (#9366) ```ts function compose<A, B, C>(f: (x: A) => B, g: (y: B) => C): (x: A) => C { // ... } let boxElements: (a: string) => { value: string[] } = compose(x => [x], y => { value: y }); ``` Currently this works. Now we ask, what if we wanted `boxElements` to operate on *any* type? ```ts function compose<A, B, C>(f: (x: A) => B, g: (y: B) => C): (x: A) => C { // ... } let boxElements: <T>(a: T) => { value: T[] } = compose(x => [x], y => { value: y }); ``` Currently this doesn't work well - when we get the contextual type for `x`, we get it from the *erased* signature of `<T>(a: T) => { value: T[] }`, which is really just `(a: any) => { value: any[] }`. We will have a PR that will help fix this. Q: What about constraints? A: Constraints will be carried through. We will still have some problems with compositional patterns *without* explicit types. For example: ```ts declare function compose<A, B, C>(f: (x: A) => B, g: (y: B) => C): (x: A) => C; declare function box<T>(x: T): { value: T }; declare function list<U>(x: U): U[]; compose(list, box); ``` The problem is that TypeScript draws the following inferences: Type Parameter | Inferences ---------------|----------- `A` | `U` `B` | `U[]`, `T` `C` | `{ value: T }` The way we walk through arguments (simple and left-to-right) and draw inferences simply isn't sufficient - we need to find a way to *unify* these type parameters. But it's not just a matter of throwing unification at the type system. So the current question is how do we introduce some unification to the current process. Perhaps it will be a "last resort" process. And this will be exploratory work. @gcnew did do great work, but pointed out many of the difficulties and roadblocks you can run into. # Changing default target to ES5 - Very strange that we wouldn't simply move the the higher one? - Why take a breaking change when we'll need to make another breaking change in the future? - Are there a lot of people running into this problem? - Doesn't appear that a lot of people have been complaining about it. # Pure annotation in downlevel emits - Currently, Uglify doesn't understand when our class emit. - It would be great if Uglify could operate on the ES2015 code. - We just want to tell other tools it's a class - onus of determining side-effects being on TypeScript is probably more questionable than an optimizer doing so. - Decision: emit JSDoc `@class` comment unconditionally. # Lib reference directive (#15780) - `/// <reference lib="name" />` - Problem: things like corejs on DefinitelyTyped conflict with compiler defaults. - This means could simply be reduced to `/// <reference lib="es2015.d.ts" />`. - Q: What about `/// <reference no-default-lib="true"/>`? - This PR ignores any `/// <reference no-default-lib="true"/>` comments. - Could also do `/// <reference lib="..." />` in `lib.d.ts` as well. - Let's do that. - What about Daniel's lib versioning idea of publishing to `@types`? - Example: DOM APIs require new logic - suddenly breaks certain builds. - People would be able to go to a specific version of a lib if they got broken, or just lock down all their dependencies. - Also might allow us to bring `node.d.ts` in.
non_priority
design meeting notes propagated inference for uninstantiated free type parameters ts function compose f x a b g y b c x a c let boxelements a string value string compose x y value y currently this works now we ask what if we wanted boxelements to operate on any type ts function compose f x a b g y b c x a c let boxelements a t value t compose x y value y currently this doesn t work well when we get the contextual type for x we get it from the erased signature of a t value t which is really just a any value any we will have a pr that will help fix this q what about constraints a constraints will be carried through we will still have some problems with compositional patterns without explicit types for example ts declare function compose f x a b g y b c x a c declare function box x t value t declare function list x u u compose list box the problem is that typescript draws the following inferences type parameter inferences a u b u t c value t the way we walk through arguments simple and left to right and draw inferences simply isn t sufficient we need to find a way to unify these type parameters but it s not just a matter of throwing unification at the type system so the current question is how do we introduce some unification to the current process perhaps it will be a last resort process and this will be exploratory work gcnew did do great work but pointed out many of the difficulties and roadblocks you can run into changing default target to very strange that we wouldn t simply move the the higher one why take a breaking change when we ll need to make another breaking change in the future are there a lot of people running into this problem doesn t appear that a lot of people have been complaining about it pure annotation in downlevel emits currently uglify doesn t understand when our class emit it would be great if uglify could operate on the code we just want to tell other tools it s a class onus of determining side effects being on typescript is probably more questionable than an optimizer doing so decision emit jsdoc class comment unconditionally lib reference directive problem things like corejs on definitelytyped conflict with compiler defaults this means could simply be reduced to q what about this pr ignores any comments could also do in lib d ts as well let s do that what about daniel s lib versioning idea of publishing to types example dom apis require new logic suddenly breaks certain builds people would be able to go to a specific version of a lib if they got broken or just lock down all their dependencies also might allow us to bring node d ts in
0
386,903
11,452,516,244
IssuesEvent
2020-02-06 13:50:14
ooni/probe
https://api.github.com/repos/ooni/probe
opened
Properly handle SIGTERM signal on windows
bug ooni/probe-cli ooni/probe-desktop platform/windows priority/high
It has been reported that the SIGTERM signal is not getting trapped on windows and therefore when you try to terminate an OONI Probe run on windows it does not shutdown cleanly as it does on macOS. We should check to see how we can handle this properly on windows.
1.0
Properly handle SIGTERM signal on windows - It has been reported that the SIGTERM signal is not getting trapped on windows and therefore when you try to terminate an OONI Probe run on windows it does not shutdown cleanly as it does on macOS. We should check to see how we can handle this properly on windows.
priority
properly handle sigterm signal on windows it has been reported that the sigterm signal is not getting trapped on windows and therefore when you try to terminate an ooni probe run on windows it does not shutdown cleanly as it does on macos we should check to see how we can handle this properly on windows
1
113,082
11,787,254,264
IssuesEvent
2020-03-17 13:44:59
sei-nope/inStock-client
https://api.github.com/repos/sei-nope/inStock-client
closed
Update README
documentation
- An explanation of the what the app does and how it works. - A link to the other repo - A link to both deployed sites - List of technologies used - List unsolved problems which would be fixed in future iterations. - Document your planning, process and problem-solving strategy - Complete the repository Description field and Website field with a meaningful sentence description of the application and link to the live URL Your front-end repo's README should also have: - Link to wireframes and user stories
1.0
Update README - - An explanation of the what the app does and how it works. - A link to the other repo - A link to both deployed sites - List of technologies used - List unsolved problems which would be fixed in future iterations. - Document your planning, process and problem-solving strategy - Complete the repository Description field and Website field with a meaningful sentence description of the application and link to the live URL Your front-end repo's README should also have: - Link to wireframes and user stories
non_priority
update readme an explanation of the what the app does and how it works a link to the other repo a link to both deployed sites list of technologies used list unsolved problems which would be fixed in future iterations document your planning process and problem solving strategy complete the repository description field and website field with a meaningful sentence description of the application and link to the live url your front end repo s readme should also have link to wireframes and user stories
0
101,210
21,629,193,062
IssuesEvent
2022-05-05 07:54:36
withfig/fig
https://api.github.com/repos/withfig/fig
closed
Show custom aliases in autocompletion
type:feature-request codebase:autocomplete-app
**Is your feature request related to a problem? Please describe.** Given an alias defined as such: ``` alias gc="git commit -m" # make a commit with a message (gc [message]) ``` **Describe the solution you'd like** * I'd love to see the comment parsed and turned into the suggestion at the bottom of the autocomplete window. * I'd also love to see all my aliases listed out in the autocomplete window, so when I start to type `g` they are selectable. **Describe alternatives you've considered** The only one I can think of is defining the rules around my aliases separately in the fig autocompletion settings. This is not ideal because then I'd be managing them in two spots. **Additional context** N/A
1.0
Show custom aliases in autocompletion - **Is your feature request related to a problem? Please describe.** Given an alias defined as such: ``` alias gc="git commit -m" # make a commit with a message (gc [message]) ``` **Describe the solution you'd like** * I'd love to see the comment parsed and turned into the suggestion at the bottom of the autocomplete window. * I'd also love to see all my aliases listed out in the autocomplete window, so when I start to type `g` they are selectable. **Describe alternatives you've considered** The only one I can think of is defining the rules around my aliases separately in the fig autocompletion settings. This is not ideal because then I'd be managing them in two spots. **Additional context** N/A
non_priority
show custom aliases in autocompletion is your feature request related to a problem please describe given an alias defined as such alias gc git commit m make a commit with a message gc describe the solution you d like i d love to see the comment parsed and turned into the suggestion at the bottom of the autocomplete window i d also love to see all my aliases listed out in the autocomplete window so when i start to type g they are selectable describe alternatives you ve considered the only one i can think of is defining the rules around my aliases separately in the fig autocompletion settings this is not ideal because then i d be managing them in two spots additional context n a
0
8,234
10,311,685,778
IssuesEvent
2019-08-29 17:58:40
cobalt-org/liquid-rust
https://api.github.com/repos/cobalt-org/liquid-rust
closed
Implement where filter
enhancement good first issue std-compatibility
Example test: ```rust assert_eq!([{ v!("a"): v!("ok") }], filters!(where, { v!("a"): v!("ok") }, "a", r#"ok")#); ``` - [Where filters live](https://github.com/cobalt-org/liquid-rust/blob/master/src/filters/mod.rs) - [Ruby implementation](https://github.com/Shopify/liquid/blob/master/lib/liquid/standardfilters.rb) - [Relevant tests](https://github.com/cobalt-org/liquid-rust/search?q=liquid+291&unscoped_q=liquid+291) - Once this is working, ideally the test will start failing. We would just need to remove the `#[should_panic]` on it - If the test doesn't start failing, please temporarily remove `#[should_panic]`, and find or create an issue for the new failure
True
Implement where filter - Example test: ```rust assert_eq!([{ v!("a"): v!("ok") }], filters!(where, { v!("a"): v!("ok") }, "a", r#"ok")#); ``` - [Where filters live](https://github.com/cobalt-org/liquid-rust/blob/master/src/filters/mod.rs) - [Ruby implementation](https://github.com/Shopify/liquid/blob/master/lib/liquid/standardfilters.rb) - [Relevant tests](https://github.com/cobalt-org/liquid-rust/search?q=liquid+291&unscoped_q=liquid+291) - Once this is working, ideally the test will start failing. We would just need to remove the `#[should_panic]` on it - If the test doesn't start failing, please temporarily remove `#[should_panic]`, and find or create an issue for the new failure
non_priority
implement where filter example test rust assert eq filters where v a v ok a r ok once this is working ideally the test will start failing we would just need to remove the on it if the test doesn t start failing please temporarily remove and find or create an issue for the new failure
0
483,342
13,923,154,989
IssuesEvent
2020-10-21 14:06:44
zephyrproject-rtos/zephyr
https://api.github.com/repos/zephyrproject-rtos/zephyr
closed
[Coverity CID :214214] Uninitialized pointer read in tests/benchmarks/data_structure_perf/rbtree_perf/src/rbtree_perf.c
Coverity bug priority: low
Static code scan issues found in file: https://github.com/zephyrproject-rtos/zephyr/tree/27e1fd69f987d3082229b3368a87d59ffdcab5a4/tests/benchmarks/data_structure_perf/rbtree_perf/src/rbtree_perf.c#L58 Category: Memory - illegal accesses Function: `test_rbtree_container` Component: Tests CID: [214214](https://scan9.coverity.com/reports.htm#v29726/p12996/mergedDefectId=214214) Details: ``` 52 struct rbnode *foreach_node; 53 struct container_node tree_node[10]; 54 55 test_tree_l.lessthan_fn = node_lessthan; 56 for (uint32_t i = 0; i < ARRAY_SIZE(tree_node); i++) { 57 tree_node[i].value = i; >>> CID 214214: Memory - illegal accesses (UNINIT) >>> Using uninitialized value "test_tree_l.max_depth" when calling "rb_insert". 58 rb_insert(&test_tree_l, &tree_node[i].node); 59 } 60 61 RB_FOR_EACH(&test_tree_l, foreach_node) { 62 zassert_true(CONTAINER_OF(foreach_node, struct container_node, 63 node)->value == count, "RB_FOR_EACH failed"); ``` Please fix or provide comments in coverity using the link: https://scan9.coverity.com/reports.htm#v32951/p12996. Note: This issue was created automatically. Priority was set based on classification of the file affected and the impact field in coverity. Assignees were set using the CODEOWNERS file.
1.0
[Coverity CID :214214] Uninitialized pointer read in tests/benchmarks/data_structure_perf/rbtree_perf/src/rbtree_perf.c - Static code scan issues found in file: https://github.com/zephyrproject-rtos/zephyr/tree/27e1fd69f987d3082229b3368a87d59ffdcab5a4/tests/benchmarks/data_structure_perf/rbtree_perf/src/rbtree_perf.c#L58 Category: Memory - illegal accesses Function: `test_rbtree_container` Component: Tests CID: [214214](https://scan9.coverity.com/reports.htm#v29726/p12996/mergedDefectId=214214) Details: ``` 52 struct rbnode *foreach_node; 53 struct container_node tree_node[10]; 54 55 test_tree_l.lessthan_fn = node_lessthan; 56 for (uint32_t i = 0; i < ARRAY_SIZE(tree_node); i++) { 57 tree_node[i].value = i; >>> CID 214214: Memory - illegal accesses (UNINIT) >>> Using uninitialized value "test_tree_l.max_depth" when calling "rb_insert". 58 rb_insert(&test_tree_l, &tree_node[i].node); 59 } 60 61 RB_FOR_EACH(&test_tree_l, foreach_node) { 62 zassert_true(CONTAINER_OF(foreach_node, struct container_node, 63 node)->value == count, "RB_FOR_EACH failed"); ``` Please fix or provide comments in coverity using the link: https://scan9.coverity.com/reports.htm#v32951/p12996. Note: This issue was created automatically. Priority was set based on classification of the file affected and the impact field in coverity. Assignees were set using the CODEOWNERS file.
priority
uninitialized pointer read in tests benchmarks data structure perf rbtree perf src rbtree perf c static code scan issues found in file category memory illegal accesses function test rbtree container component tests cid details struct rbnode foreach node struct container node tree node test tree l lessthan fn node lessthan for t i i array size tree node i tree node value i cid memory illegal accesses uninit using uninitialized value test tree l max depth when calling rb insert rb insert test tree l tree node node rb for each test tree l foreach node zassert true container of foreach node struct container node node value count rb for each failed please fix or provide comments in coverity using the link note this issue was created automatically priority was set based on classification of the file affected and the impact field in coverity assignees were set using the codeowners file
1
646,861
21,077,065,888
IssuesEvent
2022-04-02 09:52:48
apache/incubator-kyuubi
https://api.github.com/repos/apache/incubator-kyuubi
closed
[Bug] The running query will not update the duration of the page
kind:bug priority:major
### Code of Conduct - [X] I agree to follow this project's [Code of Conduct](https://www.apache.org/foundation/policies/conduct) ### Search before asking - [X] I have searched in the [issues](https://github.com/apache/incubator-kyuubi/issues?q=is%3Aissue) and found no similar issues. ### Describe the bug The default value of completedTime is 0, and the calculation of duration only considers the case where completedTime is -1. org.apache.kyuubi.operation.AbstractOperation ```scala protected var completedTime: Long = _ ```` org.apache.kyuubi.engine.spark.events.SparkOperationEvent ```scala def duration: Long = { if (completeTime == -1L) { System.currentTimeMillis - createTime } else { completeTime - createTime } } ``` ### Affects Version(s) 1.5.0 ### Kyuubi Server Log Output _No response_ ### Kyuubi Engine Log Output _No response_ ### Kyuubi Server Configurations _No response_ ### Kyuubi Engine Configurations _No response_ ### Additional context _No response_ ### Are you willing to submit PR? - [X] Yes I am willing to submit a PR!
1.0
[Bug] The running query will not update the duration of the page - ### Code of Conduct - [X] I agree to follow this project's [Code of Conduct](https://www.apache.org/foundation/policies/conduct) ### Search before asking - [X] I have searched in the [issues](https://github.com/apache/incubator-kyuubi/issues?q=is%3Aissue) and found no similar issues. ### Describe the bug The default value of completedTime is 0, and the calculation of duration only considers the case where completedTime is -1. org.apache.kyuubi.operation.AbstractOperation ```scala protected var completedTime: Long = _ ```` org.apache.kyuubi.engine.spark.events.SparkOperationEvent ```scala def duration: Long = { if (completeTime == -1L) { System.currentTimeMillis - createTime } else { completeTime - createTime } } ``` ### Affects Version(s) 1.5.0 ### Kyuubi Server Log Output _No response_ ### Kyuubi Engine Log Output _No response_ ### Kyuubi Server Configurations _No response_ ### Kyuubi Engine Configurations _No response_ ### Additional context _No response_ ### Are you willing to submit PR? - [X] Yes I am willing to submit a PR!
priority
the running query will not update the duration of the page code of conduct i agree to follow this project s search before asking i have searched in the and found no similar issues describe the bug the default value of completedtime is and the calculation of duration only considers the case where completedtime is org apache kyuubi operation abstractoperation scala protected var completedtime long org apache kyuubi engine spark events sparkoperationevent scala def duration long if completetime system currenttimemillis createtime else completetime createtime affects version s kyuubi server log output no response kyuubi engine log output no response kyuubi server configurations no response kyuubi engine configurations no response additional context no response are you willing to submit pr yes i am willing to submit a pr
1
159,258
20,048,345,438
IssuesEvent
2022-02-03 01:07:30
kapseliboi/owner-profiles
https://api.github.com/repos/kapseliboi/owner-profiles
opened
CVE-2018-1107 (Medium) detected in is-my-json-valid-2.17.1.tgz
security vulnerability
## CVE-2018-1107 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>is-my-json-valid-2.17.1.tgz</b></p></summary> <p>A JSONSchema validator that uses code generation to be extremely fast</p> <p>Library home page: <a href="https://registry.npmjs.org/is-my-json-valid/-/is-my-json-valid-2.17.1.tgz">https://registry.npmjs.org/is-my-json-valid/-/is-my-json-valid-2.17.1.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/is-my-json-valid/package.json</p> <p> Dependency Hierarchy: - standard-10.0.3.tgz (Root Library) - eslint-3.19.0.tgz - :x: **is-my-json-valid-2.17.1.tgz** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> It was discovered that the is-my-json-valid JavaScript library used an inefficient regular expression to validate JSON fields defined to have email format. A specially crafted JSON file could cause it to consume an excessive amount of CPU time when validated. <p>Publish Date: 2021-03-30 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-1107>CVE-2018-1107</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/advisories/GHSA-4hpf-3wq7-5rpr">https://github.com/advisories/GHSA-4hpf-3wq7-5rpr</a></p> <p>Release Date: 2021-03-30</p> <p>Fix Resolution: is-my-json-valid - 1.4.1, 2.17.2</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2018-1107 (Medium) detected in is-my-json-valid-2.17.1.tgz - ## CVE-2018-1107 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>is-my-json-valid-2.17.1.tgz</b></p></summary> <p>A JSONSchema validator that uses code generation to be extremely fast</p> <p>Library home page: <a href="https://registry.npmjs.org/is-my-json-valid/-/is-my-json-valid-2.17.1.tgz">https://registry.npmjs.org/is-my-json-valid/-/is-my-json-valid-2.17.1.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/is-my-json-valid/package.json</p> <p> Dependency Hierarchy: - standard-10.0.3.tgz (Root Library) - eslint-3.19.0.tgz - :x: **is-my-json-valid-2.17.1.tgz** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> It was discovered that the is-my-json-valid JavaScript library used an inefficient regular expression to validate JSON fields defined to have email format. A specially crafted JSON file could cause it to consume an excessive amount of CPU time when validated. <p>Publish Date: 2021-03-30 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-1107>CVE-2018-1107</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/advisories/GHSA-4hpf-3wq7-5rpr">https://github.com/advisories/GHSA-4hpf-3wq7-5rpr</a></p> <p>Release Date: 2021-03-30</p> <p>Fix Resolution: is-my-json-valid - 1.4.1, 2.17.2</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_priority
cve medium detected in is my json valid tgz cve medium severity vulnerability vulnerable library is my json valid tgz a jsonschema validator that uses code generation to be extremely fast library home page a href path to dependency file package json path to vulnerable library node modules is my json valid package json dependency hierarchy standard tgz root library eslint tgz x is my json valid tgz vulnerable library found in base branch master vulnerability details it was discovered that the is my json valid javascript library used an inefficient regular expression to validate json fields defined to have email format a specially crafted json file could cause it to consume an excessive amount of cpu time when validated publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution is my json valid step up your open source security game with whitesource
0
59,851
7,298,719,650
IssuesEvent
2018-02-26 17:48:31
thePlums/web-app
https://api.github.com/repos/thePlums/web-app
closed
Set a remind for individual item
design
Pretty straight forward, he went to the alarm icon right away and undstood the "Select" instructions. Recommended perhaps the icon is on each item/collection instead.
1.0
Set a remind for individual item - Pretty straight forward, he went to the alarm icon right away and undstood the "Select" instructions. Recommended perhaps the icon is on each item/collection instead.
non_priority
set a remind for individual item pretty straight forward he went to the alarm icon right away and undstood the select instructions recommended perhaps the icon is on each item collection instead
0
End of preview. Expand in Data Studio

No dataset card yet

Downloads last month
5

Collection including karths/binary-10IQR-priority