Dataset Viewer
Auto-converted to Parquet
Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
7
112
repo_url
stringlengths
36
141
action
stringclasses
3 values
title
stringlengths
1
853
labels
stringlengths
4
898
body
stringlengths
2
262k
index
stringclasses
13 values
text_combine
stringlengths
96
262k
label
stringclasses
2 values
text
stringlengths
96
250k
binary_label
int64
0
1
769,701
27,016,818,788
IssuesEvent
2023-02-10 20:14:16
tallyhowallet/extension
https://api.github.com/repos/tallyhowallet/extension
closed
Unable to connect to yearn
Type: Bug Status: Pending Priority: Medium
### Discord Discussion Link _No response_ ### What browsers are you seeing the problem on? Chrome ### What were you trying to do? Connect Tally to yearn! ### What did not work? When Tally is enabled as default wallet... I'm not able to connect with Tally; rather I get a notice re: mm <img width="1331" alt="Screen Shot 2022-02-23 at 4 31 02 PM" src="https://user-images.githubusercontent.com/7005061/155411994-e1717b7f-07c5-4ca7-95e4-4ac3f4fd1650.png"> https://user-images.githubusercontent.com/7005061/155412004-12fdd36f-b116-4665-95a3-5d6496ec9214.mov ### Version _No response_ ### Relevant log output _No response_
1.0
Unable to connect to yearn - ### Discord Discussion Link _No response_ ### What browsers are you seeing the problem on? Chrome ### What were you trying to do? Connect Tally to yearn! ### What did not work? When Tally is enabled as default wallet... I'm not able to connect with Tally; rather I get a notice re: mm <img width="1331" alt="Screen Shot 2022-02-23 at 4 31 02 PM" src="https://user-images.githubusercontent.com/7005061/155411994-e1717b7f-07c5-4ca7-95e4-4ac3f4fd1650.png"> https://user-images.githubusercontent.com/7005061/155412004-12fdd36f-b116-4665-95a3-5d6496ec9214.mov ### Version _No response_ ### Relevant log output _No response_
non_build
unable to connect to yearn discord discussion link no response what browsers are you seeing the problem on chrome what were you trying to do connect tally to yearn what did not work when tally is enabled as default wallet i m not able to connect with tally rather i get a notice re mm img width alt screen shot at pm src version no response relevant log output no response
0
15,859
6,047,695,660
IssuesEvent
2017-06-12 14:57:38
meteor/meteor
https://api.github.com/repos/meteor/meteor
closed
Define concatenated css files with html import?
feature Project:Isobuild:Minifiers
http://stackoverflow.com/questions/31077273/how-can-an-external-stylesheet-be-used-to-style-polymer-1-0-elements As asked in this stackoverflow, currently when using polymer you are able to style paper-elements (or custom elements in general cross browser in the future) with custom properties or whatever they're called sorry. However, you need to add is="custom-style" to your style tag. Meteor takes all the css files and concatenates them automatically and uses the link tag to import them, so you don't even end up with a style tag. A solution for the problem is doing an html import and having the styles included in that file inline. But as mentioned in my comment on stackoverflow this means you need to add at least one additional http request. I know I'm super nitpicky here but developing for the Philippines (slowest internet speed in Asia, 3.6MBPS average) and Asia just means you want to optimize even the slighest detail for speed. Would it theoretically be possible to concatenate everything into a html import instead of linking to a css file so you don't need to generate two seperate css files?
1.0
Define concatenated css files with html import? - http://stackoverflow.com/questions/31077273/how-can-an-external-stylesheet-be-used-to-style-polymer-1-0-elements As asked in this stackoverflow, currently when using polymer you are able to style paper-elements (or custom elements in general cross browser in the future) with custom properties or whatever they're called sorry. However, you need to add is="custom-style" to your style tag. Meteor takes all the css files and concatenates them automatically and uses the link tag to import them, so you don't even end up with a style tag. A solution for the problem is doing an html import and having the styles included in that file inline. But as mentioned in my comment on stackoverflow this means you need to add at least one additional http request. I know I'm super nitpicky here but developing for the Philippines (slowest internet speed in Asia, 3.6MBPS average) and Asia just means you want to optimize even the slighest detail for speed. Would it theoretically be possible to concatenate everything into a html import instead of linking to a css file so you don't need to generate two seperate css files?
build
define concatenated css files with html import as asked in this stackoverflow currently when using polymer you are able to style paper elements or custom elements in general cross browser in the future with custom properties or whatever they re called sorry however you need to add is custom style to your style tag meteor takes all the css files and concatenates them automatically and uses the link tag to import them so you don t even end up with a style tag a solution for the problem is doing an html import and having the styles included in that file inline but as mentioned in my comment on stackoverflow this means you need to add at least one additional http request i know i m super nitpicky here but developing for the philippines slowest internet speed in asia average and asia just means you want to optimize even the slighest detail for speed would it theoretically be possible to concatenate everything into a html import instead of linking to a css file so you don t need to generate two seperate css files
1
46,165
11,795,098,320
IssuesEvent
2020-03-18 08:15:07
xamarin/xamarin-android
https://api.github.com/repos/xamarin/xamarin-android
opened
R8 throws an error when code shrinker is set to none
Area: App+Library Build
### Steps to Reproduce 0. Create empty project 1. Install package com.onesginal, 2. Set DexCompiler to d8 3. Set Code Shrinker to none 4. Enable multidex ### Expected Behavior There are no errors from R8 (because code shrinker is set to none) ### Actual Behavior R8 Fires error ``` 2>R8 : warning : Missing class: com.google.android.gms.location.LocationListener 2>R8 : error : Compilation can't be completed because some library classes are missing. ``` ### Version Information All details, sample repo, and more information are here: https://github.com/OneSignal/OneSignal-Xamarin-SDK/issues/185 Why R8 is fired when mulidex is enabled? If that is default behavior, then visual studio csproj editor should chcck "codeshrinker" to d8 automatically when selecting multidex, otherwise developer don't know whats happening.
1.0
R8 throws an error when code shrinker is set to none - ### Steps to Reproduce 0. Create empty project 1. Install package com.onesginal, 2. Set DexCompiler to d8 3. Set Code Shrinker to none 4. Enable multidex ### Expected Behavior There are no errors from R8 (because code shrinker is set to none) ### Actual Behavior R8 Fires error ``` 2>R8 : warning : Missing class: com.google.android.gms.location.LocationListener 2>R8 : error : Compilation can't be completed because some library classes are missing. ``` ### Version Information All details, sample repo, and more information are here: https://github.com/OneSignal/OneSignal-Xamarin-SDK/issues/185 Why R8 is fired when mulidex is enabled? If that is default behavior, then visual studio csproj editor should chcck "codeshrinker" to d8 automatically when selecting multidex, otherwise developer don't know whats happening.
build
throws an error when code shrinker is set to none steps to reproduce create empty project install package com onesginal set dexcompiler to set code shrinker to none enable multidex expected behavior there are no errors from because code shrinker is set to none actual behavior fires error warning missing class com google android gms location locationlistener error compilation can t be completed because some library classes are missing version information all details sample repo and more information are here why is fired when mulidex is enabled if that is default behavior then visual studio csproj editor should chcck codeshrinker to automatically when selecting multidex otherwise developer don t know whats happening
1
22,608
11,685,931,389
IssuesEvent
2020-03-05 09:58:42
Automattic/jetpack
https://api.github.com/repos/Automattic/jetpack
closed
Tiled Gallery block: images are served with Photon, but not resized
Gutenberg Performance Tiled Galleries [Block] Tiled Gallery [Pri] High [Type] Enhancement
#### Steps to reproduce the issue 1. Add a Tiled Gallery with Gutenberg using large, hi-res images. 2. The images on the back end and front end aren't resized with Photon. They are loading in the large, full-sized images instead. This causes the browser to freeze up and leads to slow loading on the front end. #### What I expected Photon should have added the "w" and "h" attributes to load in smaller versions of the images. This works correctly when using Tiled Gallery and the Classic Editor. #### What happened instead Tiled Gallery (Gutenberg) loads in the full-sized images. This causes the browser to freeze on the back end and the images to load in slowly on the front end if the images are very large.
True
Tiled Gallery block: images are served with Photon, but not resized - #### Steps to reproduce the issue 1. Add a Tiled Gallery with Gutenberg using large, hi-res images. 2. The images on the back end and front end aren't resized with Photon. They are loading in the large, full-sized images instead. This causes the browser to freeze up and leads to slow loading on the front end. #### What I expected Photon should have added the "w" and "h" attributes to load in smaller versions of the images. This works correctly when using Tiled Gallery and the Classic Editor. #### What happened instead Tiled Gallery (Gutenberg) loads in the full-sized images. This causes the browser to freeze on the back end and the images to load in slowly on the front end if the images are very large.
non_build
tiled gallery block images are served with photon but not resized steps to reproduce the issue add a tiled gallery with gutenberg using large hi res images the images on the back end and front end aren t resized with photon they are loading in the large full sized images instead this causes the browser to freeze up and leads to slow loading on the front end what i expected photon should have added the w and h attributes to load in smaller versions of the images this works correctly when using tiled gallery and the classic editor what happened instead tiled gallery gutenberg loads in the full sized images this causes the browser to freeze on the back end and the images to load in slowly on the front end if the images are very large
0
53,634
13,189,973,272
IssuesEvent
2020-08-13 09:22:30
googleapis/java-bigtable-hbase
https://api.github.com/repos/googleapis/java-bigtable-hbase
opened
bigtable.grpc.scanner.ReadRowsAcceptanceTest: many tests failed
buildcop: issue priority: p1 type: bug
Many tests failed at the same time in this package. * I will close this issue when there are no more failures in this package _and_ there is at least one pass. * No new issues will be filed for this package until this issue is closed. * If there are already issues for individual test cases, I will close them when the corresponding test passes. You can close them earlier, if you prefer, and I won't reopen them while this issue is still open. Here are the tests that failed: * test[invalidNoCommit] * test[invalidNoCellKeyBeforeCommit] * test[invalidNoCellKeyBeforeValue] * test[invalidNewColFamilyMustSpecifyQualifier] * test[bareCommitImpliesTs=0] * test[simpleRowWithTimestamp] * test[missingTimestamp,ImpliedTs=0] * test[emptyCellValue] * test[twoUnsplitCells] * test[twoQualifiers] * test[twoFamilies] * test[withLabels] * test[splitCell,BareCommit] * test[splitCell] * test[splitFourWays] * test[twoSplitCells] * test[multiQualifierSplits] * test[multiQualifierMultiSplit] * test[multiFamilySplit] * test[invalidNoCommitBetweenRows] * test[invalidNoCommitAfterFirstRow] * test[invalidLastRowMissingCommit] * test[invalidDuplicateRowKey] * test[invalidNewRowMissingRowKey] * test[twoRows] * test[twoRowsImplicitTimestamp] * test[twoRowsEmptyValue] * test[twoRows,OneWithMultipleCells] * test[twoRows,MultipleCells] * test[twoRows,MultipleCells,MultipleFamilies] * test[twoRows,FourCells,2Labels] * test[twoRowsWithSplits,SameTimestamp] * test[invalidBareReset] * test[invalidBadReset,NoCommit] * test[invalidMissingKeyAfterReset] * test[noDataAfterReset] * test[simpleReset] * test[resetToNewVal] * test[resetToNewQual] * test[resetWithSplits] * test[resetTwoCells] * test[twoResets] * test[resetThenTwoCells] * test[resetToNewRow] * test[resetInBetweenChunks] * test[invalidResetWithChunk] * test[invalidCommitWithChunk] * test[emptyCellChunk] ----- commit: 81641f77b42506cdd7f0ddcb2192b5f0151aa413 buildURL: [Build Status](https://source.cloud.google.com/results/invocations/1c4ccf74-76a3-4494-a56d-e98266b137a0), [Sponge](http://sponge2/1c4ccf74-76a3-4494-a56d-e98266b137a0) status: failed
1.0
bigtable.grpc.scanner.ReadRowsAcceptanceTest: many tests failed - Many tests failed at the same time in this package. * I will close this issue when there are no more failures in this package _and_ there is at least one pass. * No new issues will be filed for this package until this issue is closed. * If there are already issues for individual test cases, I will close them when the corresponding test passes. You can close them earlier, if you prefer, and I won't reopen them while this issue is still open. Here are the tests that failed: * test[invalidNoCommit] * test[invalidNoCellKeyBeforeCommit] * test[invalidNoCellKeyBeforeValue] * test[invalidNewColFamilyMustSpecifyQualifier] * test[bareCommitImpliesTs=0] * test[simpleRowWithTimestamp] * test[missingTimestamp,ImpliedTs=0] * test[emptyCellValue] * test[twoUnsplitCells] * test[twoQualifiers] * test[twoFamilies] * test[withLabels] * test[splitCell,BareCommit] * test[splitCell] * test[splitFourWays] * test[twoSplitCells] * test[multiQualifierSplits] * test[multiQualifierMultiSplit] * test[multiFamilySplit] * test[invalidNoCommitBetweenRows] * test[invalidNoCommitAfterFirstRow] * test[invalidLastRowMissingCommit] * test[invalidDuplicateRowKey] * test[invalidNewRowMissingRowKey] * test[twoRows] * test[twoRowsImplicitTimestamp] * test[twoRowsEmptyValue] * test[twoRows,OneWithMultipleCells] * test[twoRows,MultipleCells] * test[twoRows,MultipleCells,MultipleFamilies] * test[twoRows,FourCells,2Labels] * test[twoRowsWithSplits,SameTimestamp] * test[invalidBareReset] * test[invalidBadReset,NoCommit] * test[invalidMissingKeyAfterReset] * test[noDataAfterReset] * test[simpleReset] * test[resetToNewVal] * test[resetToNewQual] * test[resetWithSplits] * test[resetTwoCells] * test[twoResets] * test[resetThenTwoCells] * test[resetToNewRow] * test[resetInBetweenChunks] * test[invalidResetWithChunk] * test[invalidCommitWithChunk] * test[emptyCellChunk] ----- commit: 81641f77b42506cdd7f0ddcb2192b5f0151aa413 buildURL: [Build Status](https://source.cloud.google.com/results/invocations/1c4ccf74-76a3-4494-a56d-e98266b137a0), [Sponge](http://sponge2/1c4ccf74-76a3-4494-a56d-e98266b137a0) status: failed
build
bigtable grpc scanner readrowsacceptancetest many tests failed many tests failed at the same time in this package i will close this issue when there are no more failures in this package and there is at least one pass no new issues will be filed for this package until this issue is closed if there are already issues for individual test cases i will close them when the corresponding test passes you can close them earlier if you prefer and i won t reopen them while this issue is still open here are the tests that failed test test test test test test test test test test test test test test test test test test test test test test test test test test test test test test test test test test test test test test test test test test test test test test test test commit buildurl status failed
1
95,785
27,616,694,522
IssuesEvent
2023-03-09 19:58:17
flutter/flutter
https://api.github.com/repos/flutter/flutter
closed
Gradlew build fails on multiple first party packages
platform-android plugin t: gradle a: build passed first triage P4
## Steps to Reproduce 1. In your project folder, do `cd android ` and then `gradlew build` **Expected results:** Successful build **Actual results:** Build fails P.S. - I don't want the mentioned solution. I need to abort on lint error. <details> <summary>Cmd Output</summary> ``` > Task :local_auth:lint FAILED FAILURE: Build failed with an exception. * What went wrong: Execution failed for task ':local_auth:lint'. > Lint found errors in the project; aborting build. Fix the issues identified by lint, or add the following to your build script to proceed with errors: ... android { lintOptions { abortOnError false } } ... Errors found: {MyPath}\.pub-cache\hosted\pub.dartlang.org\local_auth-1.1.6\android\src\main\res\values\styles.xml:3: Error: @android:style/Theme.Material.Dialog.Alert requires API level 21 (current min is 16) [NewApi] <style name="AlertDialogCustom" parent="@android:style/Theme.Material.Dialog.Alert"> ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ {MyPath}\.pub-cache\hosted\pub.dartlang.org\local_auth-1.1.6\android\src\main\res\values\styles.xml:7: Error: android:colorAccent requires API level 21 (current min is 16) [NewApi] <item name="android:colorAccent">#FF009688</item> ~~~~~~~~~~~~~~~~~~~~~~~~~~ * Try: Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights. ``` </details> <details> <summary>Flutter Doctor</summary> ``` [√] Flutter (Channel stable, 2.2.2, on Microsoft Windows [Version 10.0.19042.928], locale en-IN) • Flutter version 2.2.2 at {MyPath} • Framework revision d79295af24 (6 days ago), 2021-06-11 08:56:01 -0700 • Engine revision 91c9fc8fe0 • Dart version 2.13.3 [√] Android toolchain - develop for Android devices (Android SDK version 30.0.3) • Android SDK at {MyPath} • Platform android-30, build-tools 30.0.3 • ANDROID_HOME = {MyPath} • Java binary at: {MyPath} • Java version OpenJDK Runtime Environment (build 11.0.8+10-b944.6842174) • All Android licenses accepted. [X] Chrome - develop for the web (Cannot find Chrome executable at .\Google\Chrome\Application\chrome.exe) ! Cannot find Chrome. Try setting CHROME_EXECUTABLE to a Chrome executable. [√] Android Studio • Android Studio at {MyPath} • Flutter plugin can be installed from: https://plugins.jetbrains.com/plugin/9212-flutter • Dart plugin can be installed from: https://plugins.jetbrains.com/plugin/6351-dart • android-studio-dir = {MyPath} • Java version OpenJDK Runtime Environment (build 11.0.8+10-b944.6842174) [√] Connected device (1 available) • Edge (web) • edge • web-javascript • Microsoft Edge 92.0.902.9 (unsupported) ! Doctor found issues in 1 category. ``` </details>
1.0
Gradlew build fails on multiple first party packages - ## Steps to Reproduce 1. In your project folder, do `cd android ` and then `gradlew build` **Expected results:** Successful build **Actual results:** Build fails P.S. - I don't want the mentioned solution. I need to abort on lint error. <details> <summary>Cmd Output</summary> ``` > Task :local_auth:lint FAILED FAILURE: Build failed with an exception. * What went wrong: Execution failed for task ':local_auth:lint'. > Lint found errors in the project; aborting build. Fix the issues identified by lint, or add the following to your build script to proceed with errors: ... android { lintOptions { abortOnError false } } ... Errors found: {MyPath}\.pub-cache\hosted\pub.dartlang.org\local_auth-1.1.6\android\src\main\res\values\styles.xml:3: Error: @android:style/Theme.Material.Dialog.Alert requires API level 21 (current min is 16) [NewApi] <style name="AlertDialogCustom" parent="@android:style/Theme.Material.Dialog.Alert"> ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ {MyPath}\.pub-cache\hosted\pub.dartlang.org\local_auth-1.1.6\android\src\main\res\values\styles.xml:7: Error: android:colorAccent requires API level 21 (current min is 16) [NewApi] <item name="android:colorAccent">#FF009688</item> ~~~~~~~~~~~~~~~~~~~~~~~~~~ * Try: Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights. ``` </details> <details> <summary>Flutter Doctor</summary> ``` [√] Flutter (Channel stable, 2.2.2, on Microsoft Windows [Version 10.0.19042.928], locale en-IN) • Flutter version 2.2.2 at {MyPath} • Framework revision d79295af24 (6 days ago), 2021-06-11 08:56:01 -0700 • Engine revision 91c9fc8fe0 • Dart version 2.13.3 [√] Android toolchain - develop for Android devices (Android SDK version 30.0.3) • Android SDK at {MyPath} • Platform android-30, build-tools 30.0.3 • ANDROID_HOME = {MyPath} • Java binary at: {MyPath} • Java version OpenJDK Runtime Environment (build 11.0.8+10-b944.6842174) • All Android licenses accepted. [X] Chrome - develop for the web (Cannot find Chrome executable at .\Google\Chrome\Application\chrome.exe) ! Cannot find Chrome. Try setting CHROME_EXECUTABLE to a Chrome executable. [√] Android Studio • Android Studio at {MyPath} • Flutter plugin can be installed from: https://plugins.jetbrains.com/plugin/9212-flutter • Dart plugin can be installed from: https://plugins.jetbrains.com/plugin/6351-dart • android-studio-dir = {MyPath} • Java version OpenJDK Runtime Environment (build 11.0.8+10-b944.6842174) [√] Connected device (1 available) • Edge (web) • edge • web-javascript • Microsoft Edge 92.0.902.9 (unsupported) ! Doctor found issues in 1 category. ``` </details>
build
gradlew build fails on multiple first party packages steps to reproduce in your project folder do cd android and then gradlew build expected results successful build actual results build fails p s i don t want the mentioned solution i need to abort on lint error cmd output task local auth lint failed failure build failed with an exception what went wrong execution failed for task local auth lint lint found errors in the project aborting build fix the issues identified by lint or add the following to your build script to proceed with errors android lintoptions abortonerror false errors found mypath pub cache hosted pub dartlang org local auth android src main res values styles xml error android style theme material dialog alert requires api level current min is mypath pub cache hosted pub dartlang org local auth android src main res values styles xml error android coloraccent requires api level current min is try run with stacktrace option to get the stack trace run with info or debug option to get more log output run with scan to get full insights flutter doctor flutter channel stable on microsoft windows locale en in • flutter version at mypath • framework revision days ago • engine revision • dart version android toolchain develop for android devices android sdk version • android sdk at mypath • platform android build tools • android home mypath • java binary at mypath • java version openjdk runtime environment build • all android licenses accepted chrome develop for the web cannot find chrome executable at google chrome application chrome exe cannot find chrome try setting chrome executable to a chrome executable android studio • android studio at mypath • flutter plugin can be installed from • dart plugin can be installed from • android studio dir mypath • java version openjdk runtime environment build connected device available • edge web • edge • web javascript • microsoft edge unsupported doctor found issues in category
1
199,259
6,987,368,868
IssuesEvent
2017-12-14 08:58:51
sunpy/sunpy
https://api.github.com/repos/sunpy/sunpy
opened
Decorator to be able to prepend and/or append to doc strings
Effort Low Feature Request Package Novice Priority Low Refactoring
Once #2369 is merged in we have under `sunpy/sun/sun.py`, a decorator (called `add_parameter_docs`) that we use to append to the functions under `sun.py` as they all take the same parameter input. We would like this to moved under to `sunpy/util/decorators.py` where we can import this function more generally as well as let it take arguments allowing us to pass in a prepend and/or append argument to the decorator that would add it to a decorated function's docstring. Maybe `.format` on the docstring if a dictionary is provided as an argument to the decorator.
1.0
Decorator to be able to prepend and/or append to doc strings - Once #2369 is merged in we have under `sunpy/sun/sun.py`, a decorator (called `add_parameter_docs`) that we use to append to the functions under `sun.py` as they all take the same parameter input. We would like this to moved under to `sunpy/util/decorators.py` where we can import this function more generally as well as let it take arguments allowing us to pass in a prepend and/or append argument to the decorator that would add it to a decorated function's docstring. Maybe `.format` on the docstring if a dictionary is provided as an argument to the decorator.
non_build
decorator to be able to prepend and or append to doc strings once is merged in we have under sunpy sun sun py a decorator called add parameter docs that we use to append to the functions under sun py as they all take the same parameter input we would like this to moved under to sunpy util decorators py where we can import this function more generally as well as let it take arguments allowing us to pass in a prepend and or append argument to the decorator that would add it to a decorated function s docstring maybe format on the docstring if a dictionary is provided as an argument to the decorator
0
92,094
26,578,723,822
IssuesEvent
2023-01-22 06:15:24
runatlantis/atlantis
https://api.github.com/repos/runatlantis/atlantis
closed
Deprecate `set-output` in favor of `$GITHUB_OUTPUT`
bug build github-actions
<!--- Please keep this note for the community ---> ### Community Note * Please vote on this issue by adding a 👍 [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request. Searching for pre-existing feature requests helps us consolidate datapoints for identical requirements into a single place, thank you! * Please do not leave "+1" or other comments that do not add relevant new information or questions, they generate extra noise for issue followers and do not help prioritize the request. * If you are interested in working on this issue or have submitted a pull request, please leave a comment. <!--- Thank you for keeping this note for the community ---> --- <!--- When filing a bug, please include the following headings if possible. Any example text in this template can be deleted. ---> ### Overview of the Issue <!--- Please describe the issue you are having and how you encountered the problem. ---> > To avoid untrusted logged data to use save-stateand set-output workflow commands without the intention of the workflow author we have introduced a new set of environment files to manage state and output. https://github.blog/changelog/2022-10-11-github-actions-deprecating-save-state-and-set-output-commands/#examples ### Reproduction Steps <!--- In order to effectively and quickly resolve the issue, please provide exact steps that allow us the reproduce the problem. If no steps are provided, then it will likely take longer to get the issue resolved. ---> n/a ### Logs <!--- Provide log files from Atlantis server <details> <summary>Logs</summary> ``` log output ``` </details> ---> n/a ### Environment details <!--- If not already included, please provide the following: - Atlantis version: - If not running the latest Atlantis version have you tried to reproduce this issue on the latest version: - Atlantis flags: Atlantis server-side config file: ``` # config file ``` Repo `atlantis.yaml` file: ``` # config file ``` Any other information you can provide about the environment/deployment. ---> n/a ### Additional Context <!--- Additional context on the problem. Docs, links to blogs, or other material that lead you to discover this issue or were helpful in troubleshooting the issue. ---> See link above I only see it `set-output` in one location https://github.com/runatlantis/atlantis/blob/d620d5de98b0178a16429fa89095507109b4a815/.github/workflows/release.yml#L35 I do not see `save-state` anywhere.
1.0
Deprecate `set-output` in favor of `$GITHUB_OUTPUT` - <!--- Please keep this note for the community ---> ### Community Note * Please vote on this issue by adding a 👍 [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request. Searching for pre-existing feature requests helps us consolidate datapoints for identical requirements into a single place, thank you! * Please do not leave "+1" or other comments that do not add relevant new information or questions, they generate extra noise for issue followers and do not help prioritize the request. * If you are interested in working on this issue or have submitted a pull request, please leave a comment. <!--- Thank you for keeping this note for the community ---> --- <!--- When filing a bug, please include the following headings if possible. Any example text in this template can be deleted. ---> ### Overview of the Issue <!--- Please describe the issue you are having and how you encountered the problem. ---> > To avoid untrusted logged data to use save-stateand set-output workflow commands without the intention of the workflow author we have introduced a new set of environment files to manage state and output. https://github.blog/changelog/2022-10-11-github-actions-deprecating-save-state-and-set-output-commands/#examples ### Reproduction Steps <!--- In order to effectively and quickly resolve the issue, please provide exact steps that allow us the reproduce the problem. If no steps are provided, then it will likely take longer to get the issue resolved. ---> n/a ### Logs <!--- Provide log files from Atlantis server <details> <summary>Logs</summary> ``` log output ``` </details> ---> n/a ### Environment details <!--- If not already included, please provide the following: - Atlantis version: - If not running the latest Atlantis version have you tried to reproduce this issue on the latest version: - Atlantis flags: Atlantis server-side config file: ``` # config file ``` Repo `atlantis.yaml` file: ``` # config file ``` Any other information you can provide about the environment/deployment. ---> n/a ### Additional Context <!--- Additional context on the problem. Docs, links to blogs, or other material that lead you to discover this issue or were helpful in troubleshooting the issue. ---> See link above I only see it `set-output` in one location https://github.com/runatlantis/atlantis/blob/d620d5de98b0178a16429fa89095507109b4a815/.github/workflows/release.yml#L35 I do not see `save-state` anywhere.
build
deprecate set output in favor of github output community note please vote on this issue by adding a 👍 to the original issue to help the community and maintainers prioritize this request searching for pre existing feature requests helps us consolidate datapoints for identical requirements into a single place thank you please do not leave or other comments that do not add relevant new information or questions they generate extra noise for issue followers and do not help prioritize the request if you are interested in working on this issue or have submitted a pull request please leave a comment overview of the issue to avoid untrusted logged data to use save stateand set output workflow commands without the intention of the workflow author we have introduced a new set of environment files to manage state and output reproduction steps in order to effectively and quickly resolve the issue please provide exact steps that allow us the reproduce the problem if no steps are provided then it will likely take longer to get the issue resolved n a logs provide log files from atlantis server logs log output n a environment details if not already included please provide the following atlantis version if not running the latest atlantis version have you tried to reproduce this issue on the latest version atlantis flags atlantis server side config file config file repo atlantis yaml file config file any other information you can provide about the environment deployment n a additional context additional context on the problem docs links to blogs or other material that lead you to discover this issue or were helpful in troubleshooting the issue see link above i only see it set output in one location i do not see save state anywhere
1
437,013
12,558,321,970
IssuesEvent
2020-06-07 15:34:02
tensorfork/tensorfork
https://api.github.com/repos/tensorfork/tensorfork
opened
BigGAN: D data augmentations per Zhao et al 2020
enhancement good first issue priority: high
The new paper ["Image Augmentations for GAN Training"](https://arxiv.org/abs/2006.02595#google), Zhao et al 2020b, reports: > Data augmentations have been widely studied to improve the accuracy and robustness of classifiers. However, the potential of image augmentation in improving GAN models for image synthesis has not been thoroughly investigated in previous studies. In this work, we systematically study the effectiveness of various existing augmentation techniques for GAN training in a variety of settings. We provide insights and guidelines on how to augment images for both vanilla GANs and GANs with regularizations, improving the fidelity of the generated images substantially. Surprisingly, we find that vanilla GANs attain generation quality on par with recent state-of-the-art results if we use augmentations on both real and generated images. When this GAN training is combined with other augmentation-based regularization techniques, such as contrastive loss and consistency regularization, the augmentations further improve the quality of generated images. We provide new state-of-the-art results for conditional generation on CIFAR-10 with both consistency loss and contrastive loss as additional regularizations. This follows up Zhao et al 2020a, ["Improved Consistency Regularization for GANs"](https://arxiv.org/abs/2002.04724), which we've discussed at length before. The major finding of 2020b here is two-part: 1. *no* data augmentation helps when applied *only* to real samples, and they often hurt, a lot! ![[higher=worse](https://arxiv.org/pdf/2006.02595.pdf#page=3)](https://user-images.githubusercontent.com/352559/83972469-4abe9000-a8ae-11ea-8f64-1bd23dabfb97.png) Note that ProGAN/StyleGAN, Mooch BigGAN, compare_gan, and MSG-GAN either do no data augmentations whatsoever, or they apply very limited data augmentation (flip/mirror and/or random cropping) to real images as part of the data loading pipeline of reals - and never inside the training loop where it could be applied to fakes too. We too currently do only random cropping as a hack to deal with non-square (real) images, but we seem to get better results when doing rescale+black-background (on reals). 2. *several* of the same data augmentations help when applied to *both* real & fake samples, and they often help, a lot! ![[lower=better](https://arxiv.org/pdf/2006.02595.pdf#page=4)](https://user-images.githubusercontent.com/352559/83972481-6aee4f00-a8ae-11ea-8e3a-8782cbc51535.png) Even for a vanilla GAN, using the right data augmentation in the right way can improve FID by as much as a quarter to a half, with no other change whatsoever. The only difference is that the augmentation is done inside the training loop to both real & fakes, instead of just fakes. (2020b uses a single data augmentation at a time, applied to the entire minibatch. There seems to be no reason you couldn't use all of the good data augmentations, and pick one at random per iteration to apply to an entire minibatch, to get even larger benefits.) Specifically, the 'spatial' data augmentations like zooming in/out or shifting the image up/down/left/right (wrapping around) help the most, while data augmentations like adding pixel noise or distorting colors do not do anything on net. (They do not consider flipping or some other possible ones like Sobel distortion, which the SimCLR paper considered.) More details on the helpful augmentation implementations: ![[augmentation definitions](https://arxiv.org/pdf/2006.02595.pdf#appendix.B)](https://user-images.githubusercontent.com/352559/83972533-aa1ca000-a8ae-11ea-981c-c9b91db00c5f.png) The intuition from 2020a is that if you do data augmentation only on reals, D will learn that reals are distorted in such a way, while the G samples are undistorted, and so the data augmentation, instead of teaching D more meaningful representations of images, instead misleads D, resulting in 'leakage' where D makes G start to generate distorted samples too. This is visible in generated samples, wastes training, and apparently is bad enough that it eliminates the benefits that data augmentation should have given you. (This may account for why so many GANs omit data augmentation entirely, or use an extremely weaksauce one like flipping - flipping can't result in 'leakage' because usually, a flipped image is just another valid image and is not 'distorted'. If you don't use data augmentation, it can't help you, but it also can't hurt you when you use it wrong.) But if you do it on both reals & fakes, D no longer learns the wrong thing. It merely gets a much more challenging task that it can learn from better, as it must distinguish real from fake even when the images are distorted in various ways, forcing it to generalize more. On the downside: while the ablations are fairly thorough and they do many runs so it's unlikely to be the usual GAN bugaboo of very high variance results and they test on meaningful architectures like BigGAN, so the results look promising, the data augmentation is tested only on CIFAR-10, which is a very small and simple data compared to Danbooru, so there's a chance it won't work nearly as well for us as for their vanilla GAN, and the biggest benefits come from adding in BCR and SimCLR (#11 ), which we don't have yet. --- This is a remarkably simple improvement we should add to BigGAN immediately to test out. Since it doesn't modify the dataset or arch, we can even add it to current runs. We probably want (in descending order of how much it helps in Figure 3): ZoomIn, ZoomOut, TranslationX, TranslationY, TranslationX+Y, mirroring/flipping, and CutOut. (CutMix is a lot harder to implement since it requires combing multiple images, and it appears to potentially hurt, so we'll skip that one.) 1. ZoomIn is just a random crop and then a resize up to fit 2. ZoomOut resizes the whole image, and then fills in the padding not with 0s, but with the edge of the image. (This makes it different from [`tf.contrib.image.translate`](https://www.tensorflow.org/versions/r1.15/api_docs/python/tf/contrib/image/translate), which pads with 0s; there is a [`tf.pad`](https://www.tensorflow.org/versions/r1.15/api_docs/python/tf/pad) which has a 'mirror' option which I think may be what is done here, but we need an image version.) 3. TranslationX: shifts the image by a few pixels left/right; the excess pixels get wrapped around to the other side 4. TranslationY: same thing, up/down, wrapping around 5. TranslationX+Y: shift both left/right and up/down (so can be implemented by calling `TranslationX(TranslationY(image))`) 6. CutOut: superimposes a small black square at a random point in the image, erasing the existing pixels; apparently implemented as [`tfa.image.cutout`](https://www.tensorflow.org/addons/api_docs/python/tfa/image/cutout) 7. mirroring: the usual, something like `tf.image.flip_left_right`
1.0
BigGAN: D data augmentations per Zhao et al 2020 - The new paper ["Image Augmentations for GAN Training"](https://arxiv.org/abs/2006.02595#google), Zhao et al 2020b, reports: > Data augmentations have been widely studied to improve the accuracy and robustness of classifiers. However, the potential of image augmentation in improving GAN models for image synthesis has not been thoroughly investigated in previous studies. In this work, we systematically study the effectiveness of various existing augmentation techniques for GAN training in a variety of settings. We provide insights and guidelines on how to augment images for both vanilla GANs and GANs with regularizations, improving the fidelity of the generated images substantially. Surprisingly, we find that vanilla GANs attain generation quality on par with recent state-of-the-art results if we use augmentations on both real and generated images. When this GAN training is combined with other augmentation-based regularization techniques, such as contrastive loss and consistency regularization, the augmentations further improve the quality of generated images. We provide new state-of-the-art results for conditional generation on CIFAR-10 with both consistency loss and contrastive loss as additional regularizations. This follows up Zhao et al 2020a, ["Improved Consistency Regularization for GANs"](https://arxiv.org/abs/2002.04724), which we've discussed at length before. The major finding of 2020b here is two-part: 1. *no* data augmentation helps when applied *only* to real samples, and they often hurt, a lot! ![[higher=worse](https://arxiv.org/pdf/2006.02595.pdf#page=3)](https://user-images.githubusercontent.com/352559/83972469-4abe9000-a8ae-11ea-8f64-1bd23dabfb97.png) Note that ProGAN/StyleGAN, Mooch BigGAN, compare_gan, and MSG-GAN either do no data augmentations whatsoever, or they apply very limited data augmentation (flip/mirror and/or random cropping) to real images as part of the data loading pipeline of reals - and never inside the training loop where it could be applied to fakes too. We too currently do only random cropping as a hack to deal with non-square (real) images, but we seem to get better results when doing rescale+black-background (on reals). 2. *several* of the same data augmentations help when applied to *both* real & fake samples, and they often help, a lot! ![[lower=better](https://arxiv.org/pdf/2006.02595.pdf#page=4)](https://user-images.githubusercontent.com/352559/83972481-6aee4f00-a8ae-11ea-8e3a-8782cbc51535.png) Even for a vanilla GAN, using the right data augmentation in the right way can improve FID by as much as a quarter to a half, with no other change whatsoever. The only difference is that the augmentation is done inside the training loop to both real & fakes, instead of just fakes. (2020b uses a single data augmentation at a time, applied to the entire minibatch. There seems to be no reason you couldn't use all of the good data augmentations, and pick one at random per iteration to apply to an entire minibatch, to get even larger benefits.) Specifically, the 'spatial' data augmentations like zooming in/out or shifting the image up/down/left/right (wrapping around) help the most, while data augmentations like adding pixel noise or distorting colors do not do anything on net. (They do not consider flipping or some other possible ones like Sobel distortion, which the SimCLR paper considered.) More details on the helpful augmentation implementations: ![[augmentation definitions](https://arxiv.org/pdf/2006.02595.pdf#appendix.B)](https://user-images.githubusercontent.com/352559/83972533-aa1ca000-a8ae-11ea-981c-c9b91db00c5f.png) The intuition from 2020a is that if you do data augmentation only on reals, D will learn that reals are distorted in such a way, while the G samples are undistorted, and so the data augmentation, instead of teaching D more meaningful representations of images, instead misleads D, resulting in 'leakage' where D makes G start to generate distorted samples too. This is visible in generated samples, wastes training, and apparently is bad enough that it eliminates the benefits that data augmentation should have given you. (This may account for why so many GANs omit data augmentation entirely, or use an extremely weaksauce one like flipping - flipping can't result in 'leakage' because usually, a flipped image is just another valid image and is not 'distorted'. If you don't use data augmentation, it can't help you, but it also can't hurt you when you use it wrong.) But if you do it on both reals & fakes, D no longer learns the wrong thing. It merely gets a much more challenging task that it can learn from better, as it must distinguish real from fake even when the images are distorted in various ways, forcing it to generalize more. On the downside: while the ablations are fairly thorough and they do many runs so it's unlikely to be the usual GAN bugaboo of very high variance results and they test on meaningful architectures like BigGAN, so the results look promising, the data augmentation is tested only on CIFAR-10, which is a very small and simple data compared to Danbooru, so there's a chance it won't work nearly as well for us as for their vanilla GAN, and the biggest benefits come from adding in BCR and SimCLR (#11 ), which we don't have yet. --- This is a remarkably simple improvement we should add to BigGAN immediately to test out. Since it doesn't modify the dataset or arch, we can even add it to current runs. We probably want (in descending order of how much it helps in Figure 3): ZoomIn, ZoomOut, TranslationX, TranslationY, TranslationX+Y, mirroring/flipping, and CutOut. (CutMix is a lot harder to implement since it requires combing multiple images, and it appears to potentially hurt, so we'll skip that one.) 1. ZoomIn is just a random crop and then a resize up to fit 2. ZoomOut resizes the whole image, and then fills in the padding not with 0s, but with the edge of the image. (This makes it different from [`tf.contrib.image.translate`](https://www.tensorflow.org/versions/r1.15/api_docs/python/tf/contrib/image/translate), which pads with 0s; there is a [`tf.pad`](https://www.tensorflow.org/versions/r1.15/api_docs/python/tf/pad) which has a 'mirror' option which I think may be what is done here, but we need an image version.) 3. TranslationX: shifts the image by a few pixels left/right; the excess pixels get wrapped around to the other side 4. TranslationY: same thing, up/down, wrapping around 5. TranslationX+Y: shift both left/right and up/down (so can be implemented by calling `TranslationX(TranslationY(image))`) 6. CutOut: superimposes a small black square at a random point in the image, erasing the existing pixels; apparently implemented as [`tfa.image.cutout`](https://www.tensorflow.org/addons/api_docs/python/tfa/image/cutout) 7. mirroring: the usual, something like `tf.image.flip_left_right`
non_build
biggan d data augmentations per zhao et al the new paper zhao et al reports data augmentations have been widely studied to improve the accuracy and robustness of classifiers however the potential of image augmentation in improving gan models for image synthesis has not been thoroughly investigated in previous studies in this work we systematically study the effectiveness of various existing augmentation techniques for gan training in a variety of settings we provide insights and guidelines on how to augment images for both vanilla gans and gans with regularizations improving the fidelity of the generated images substantially surprisingly we find that vanilla gans attain generation quality on par with recent state of the art results if we use augmentations on both real and generated images when this gan training is combined with other augmentation based regularization techniques such as contrastive loss and consistency regularization the augmentations further improve the quality of generated images we provide new state of the art results for conditional generation on cifar with both consistency loss and contrastive loss as additional regularizations this follows up zhao et al which we ve discussed at length before the major finding of here is two part no data augmentation helps when applied only to real samples and they often hurt a lot note that progan stylegan mooch biggan compare gan and msg gan either do no data augmentations whatsoever or they apply very limited data augmentation flip mirror and or random cropping to real images as part of the data loading pipeline of reals and never inside the training loop where it could be applied to fakes too we too currently do only random cropping as a hack to deal with non square real images but we seem to get better results when doing rescale black background on reals several of the same data augmentations help when applied to both real fake samples and they often help a lot even for a vanilla gan using the right data augmentation in the right way can improve fid by as much as a quarter to a half with no other change whatsoever the only difference is that the augmentation is done inside the training loop to both real fakes instead of just fakes uses a single data augmentation at a time applied to the entire minibatch there seems to be no reason you couldn t use all of the good data augmentations and pick one at random per iteration to apply to an entire minibatch to get even larger benefits specifically the spatial data augmentations like zooming in out or shifting the image up down left right wrapping around help the most while data augmentations like adding pixel noise or distorting colors do not do anything on net they do not consider flipping or some other possible ones like sobel distortion which the simclr paper considered more details on the helpful augmentation implementations the intuition from is that if you do data augmentation only on reals d will learn that reals are distorted in such a way while the g samples are undistorted and so the data augmentation instead of teaching d more meaningful representations of images instead misleads d resulting in leakage where d makes g start to generate distorted samples too this is visible in generated samples wastes training and apparently is bad enough that it eliminates the benefits that data augmentation should have given you this may account for why so many gans omit data augmentation entirely or use an extremely weaksauce one like flipping flipping can t result in leakage because usually a flipped image is just another valid image and is not distorted if you don t use data augmentation it can t help you but it also can t hurt you when you use it wrong but if you do it on both reals fakes d no longer learns the wrong thing it merely gets a much more challenging task that it can learn from better as it must distinguish real from fake even when the images are distorted in various ways forcing it to generalize more on the downside while the ablations are fairly thorough and they do many runs so it s unlikely to be the usual gan bugaboo of very high variance results and they test on meaningful architectures like biggan so the results look promising the data augmentation is tested only on cifar which is a very small and simple data compared to danbooru so there s a chance it won t work nearly as well for us as for their vanilla gan and the biggest benefits come from adding in bcr and simclr which we don t have yet this is a remarkably simple improvement we should add to biggan immediately to test out since it doesn t modify the dataset or arch we can even add it to current runs we probably want in descending order of how much it helps in figure zoomin zoomout translationx translationy translationx y mirroring flipping and cutout cutmix is a lot harder to implement since it requires combing multiple images and it appears to potentially hurt so we ll skip that one zoomin is just a random crop and then a resize up to fit zoomout resizes the whole image and then fills in the padding not with but with the edge of the image this makes it different from which pads with there is a which has a mirror option which i think may be what is done here but we need an image version translationx shifts the image by a few pixels left right the excess pixels get wrapped around to the other side translationy same thing up down wrapping around translationx y shift both left right and up down so can be implemented by calling translationx translationy image cutout superimposes a small black square at a random point in the image erasing the existing pixels apparently implemented as mirroring the usual something like tf image flip left right
0
82,776
23,874,693,436
IssuesEvent
2022-09-07 17:51:52
mfem/mfem
https://api.github.com/repos/mfem/mfem
closed
Add support for SUNDIALS v6
building WIP libraries todo
Version 6 of SUNDIALS made some interface changes that are breaking the integration with MFEM. Addition: When addressing this, we should also address an issue seen when running `ex16p` when HYPRE is built with GPU support -- there is a repeated registration of the same host pointer which leads to problems. Fix an issue: `SUNDIALS_DIR` is set by default to `@MFEM_DIR@/../sundials-5.0.0/instdir` which does not work since we try to expand the path with `$(wildcard ...)` before replacing `@MFEM_DIR@` with the real path.
1.0
Add support for SUNDIALS v6 - Version 6 of SUNDIALS made some interface changes that are breaking the integration with MFEM. Addition: When addressing this, we should also address an issue seen when running `ex16p` when HYPRE is built with GPU support -- there is a repeated registration of the same host pointer which leads to problems. Fix an issue: `SUNDIALS_DIR` is set by default to `@MFEM_DIR@/../sundials-5.0.0/instdir` which does not work since we try to expand the path with `$(wildcard ...)` before replacing `@MFEM_DIR@` with the real path.
build
add support for sundials version of sundials made some interface changes that are breaking the integration with mfem addition when addressing this we should also address an issue seen when running when hypre is built with gpu support there is a repeated registration of the same host pointer which leads to problems fix an issue sundials dir is set by default to mfem dir sundials instdir which does not work since we try to expand the path with wildcard before replacing mfem dir with the real path
1
17,618
6,485,770,224
IssuesEvent
2017-08-19 13:41:06
ElektraInitiative/libelektra
https://api.github.com/repos/ElektraInitiative/libelektra
closed
debian: errors installing+using
build low priority
The build server finally built https://build.libelektra.org/job/elektra-git-buildpackage-jessie/lastSuccessfulBuild/console :+1: Unfortunately, there are errors in installing: ``` Preparing to unpack .../elektra-qt-gui_0.8.19-1.5721_amd64.deb ... Unpacking elektra-qt-gui (0.8.19-1.5721) ... dpkg: error processing archive /var/cache/apt/archives/elektra-qt-gui_0.8.19-1.5721_amd64.deb (--unpack): trying to overwrite '/usr/share/man/man1/kdb-qt-gui.1.gz', which is also in package elektra-bin 0.8.19-1.5721 Processing triggers for man-db (2.7.0.2-5) ... Errors were encountered while processing: /var/cache/apt/archives/elektra-qt-gui_0.8.19-1.5721_amd64.deb E: Sub-process /usr/bin/dpkg returned an error code (1) Failed to perform requested operation on package. Trying to recover: Press Return to continue. ``` and install-docs fails: ``` install-docs --verbose --check /usr/share/doc-base/elektra-doc Warning in `/usr/share/doc-base/elektra-doc', line 8: file `/usr/share/doc/elektra-doc/html/index.html' does not exist. Error in `/usr/share/doc-base/elektra-doc', line 8: all `Format' sections are invalid. /usr/share/doc-base/elektra-doc: Fatal error found, the file won't be registered. ``` and using it with cmake fails, too (can be tested with `ctest -V -R external` which runs `cmake` in the `examples/external/cmake` project): ``` CMake Error at /usr/lib/x86_64-linux-gnu/cmake/Elektra/ElektraTargetsLibelektra.cmake:98 (message): The imported target "elektra-utility" references the file "/usr/lib/x86_64-linux-gnu/libelektra-utility.so.0.8.19" but this file does not exist. Possible reasons include: * The file was deleted, renamed, or moved to another location. * An install or uninstall procedure did not complete successfully. * The installation package was faulty and contained "/usr/lib/x86_64-linux-gnu/cmake/Elektra/ElektraTargetsLibelektra.cmake" but not all the files it references. Call Stack (most recent call first): /usr/lib/x86_64-linux-gnu/cmake/Elektra/ElektraConfig.cmake:16 (include) CMakeLists.txt:5 (find_package) -- Configuring incomplete, errors occurred ``` Btw. at least aptitude needed a lot of help to upgrade to the new packages. Can you test on PCS? Thank you, great job nevertheless. It is very nice to have debug symbols and hopefully also the qt-gui soon ;)
1.0
debian: errors installing+using - The build server finally built https://build.libelektra.org/job/elektra-git-buildpackage-jessie/lastSuccessfulBuild/console :+1: Unfortunately, there are errors in installing: ``` Preparing to unpack .../elektra-qt-gui_0.8.19-1.5721_amd64.deb ... Unpacking elektra-qt-gui (0.8.19-1.5721) ... dpkg: error processing archive /var/cache/apt/archives/elektra-qt-gui_0.8.19-1.5721_amd64.deb (--unpack): trying to overwrite '/usr/share/man/man1/kdb-qt-gui.1.gz', which is also in package elektra-bin 0.8.19-1.5721 Processing triggers for man-db (2.7.0.2-5) ... Errors were encountered while processing: /var/cache/apt/archives/elektra-qt-gui_0.8.19-1.5721_amd64.deb E: Sub-process /usr/bin/dpkg returned an error code (1) Failed to perform requested operation on package. Trying to recover: Press Return to continue. ``` and install-docs fails: ``` install-docs --verbose --check /usr/share/doc-base/elektra-doc Warning in `/usr/share/doc-base/elektra-doc', line 8: file `/usr/share/doc/elektra-doc/html/index.html' does not exist. Error in `/usr/share/doc-base/elektra-doc', line 8: all `Format' sections are invalid. /usr/share/doc-base/elektra-doc: Fatal error found, the file won't be registered. ``` and using it with cmake fails, too (can be tested with `ctest -V -R external` which runs `cmake` in the `examples/external/cmake` project): ``` CMake Error at /usr/lib/x86_64-linux-gnu/cmake/Elektra/ElektraTargetsLibelektra.cmake:98 (message): The imported target "elektra-utility" references the file "/usr/lib/x86_64-linux-gnu/libelektra-utility.so.0.8.19" but this file does not exist. Possible reasons include: * The file was deleted, renamed, or moved to another location. * An install or uninstall procedure did not complete successfully. * The installation package was faulty and contained "/usr/lib/x86_64-linux-gnu/cmake/Elektra/ElektraTargetsLibelektra.cmake" but not all the files it references. Call Stack (most recent call first): /usr/lib/x86_64-linux-gnu/cmake/Elektra/ElektraConfig.cmake:16 (include) CMakeLists.txt:5 (find_package) -- Configuring incomplete, errors occurred ``` Btw. at least aptitude needed a lot of help to upgrade to the new packages. Can you test on PCS? Thank you, great job nevertheless. It is very nice to have debug symbols and hopefully also the qt-gui soon ;)
build
debian errors installing using the build server finally built unfortunately there are errors in installing preparing to unpack elektra qt gui deb unpacking elektra qt gui dpkg error processing archive var cache apt archives elektra qt gui deb unpack trying to overwrite usr share man kdb qt gui gz which is also in package elektra bin processing triggers for man db errors were encountered while processing var cache apt archives elektra qt gui deb e sub process usr bin dpkg returned an error code failed to perform requested operation on package trying to recover press return to continue and install docs fails install docs verbose check usr share doc base elektra doc warning in usr share doc base elektra doc line file usr share doc elektra doc html index html does not exist error in usr share doc base elektra doc line all format sections are invalid usr share doc base elektra doc fatal error found the file won t be registered and using it with cmake fails too can be tested with ctest v r external which runs cmake in the examples external cmake project cmake error at usr lib linux gnu cmake elektra elektratargetslibelektra cmake message the imported target elektra utility references the file usr lib linux gnu libelektra utility so but this file does not exist possible reasons include the file was deleted renamed or moved to another location an install or uninstall procedure did not complete successfully the installation package was faulty and contained usr lib linux gnu cmake elektra elektratargetslibelektra cmake but not all the files it references call stack most recent call first usr lib linux gnu cmake elektra elektraconfig cmake include cmakelists txt find package configuring incomplete errors occurred btw at least aptitude needed a lot of help to upgrade to the new packages can you test on pcs thank you great job nevertheless it is very nice to have debug symbols and hopefully also the qt gui soon
1
53,917
6,774,515,699
IssuesEvent
2017-10-27 10:40:31
nawissor/stomer
https://api.github.com/repos/nawissor/stomer
closed
Front page banner
Design
Create a stich in photoshop and include a few images for the banner on the front page. See this howto --> https://digital-photography-school.com/how-to-make-a-photo-collage-banner-for-your-photoblog-or-website/
1.0
Front page banner - Create a stich in photoshop and include a few images for the banner on the front page. See this howto --> https://digital-photography-school.com/how-to-make-a-photo-collage-banner-for-your-photoblog-or-website/
non_build
front page banner create a stich in photoshop and include a few images for the banner on the front page see this howto
0
84,811
24,429,745,345
IssuesEvent
2022-10-06 07:12:25
artichoke/artichoke
https://api.github.com/repos/artichoke/artichoke
closed
Implement a build of artichoke that can be used to bootstrap the build with no Ruby dependency
A-core A-build A-ruby-core S-speculative
Once Core is complete, it should be possible to produce a Core-only build of artichoke that can be used to bootstrap artichoke-core and mruby-sys.
1.0
Implement a build of artichoke that can be used to bootstrap the build with no Ruby dependency - Once Core is complete, it should be possible to produce a Core-only build of artichoke that can be used to bootstrap artichoke-core and mruby-sys.
build
implement a build of artichoke that can be used to bootstrap the build with no ruby dependency once core is complete it should be possible to produce a core only build of artichoke that can be used to bootstrap artichoke core and mruby sys
1
21,135
6,986,235,177
IssuesEvent
2017-12-14 02:09:40
openshiftio/openshift.io
https://api.github.com/repos/openshiftio/openshift.io
opened
Pipeline crashes and fails to build
area/pipelines SEV2-high team/build-cd type/bug
The pipelines fail and require a restart of the jenkins node/pod thing or GKE or something. Once whatever it is gets restarted the build passes. This thing stops working a lot. Why is it so fragile? We need this to be up and running all the time. Example PR: https://github.com/fabric8-ui/fabric8-ui/pull/2302 Logs: https://jenkins.cd.test.fabric8.io/job/fabric8-ui/job/fabric8-ui/view/change-requests/job/PR-2302/5/console
1.0
Pipeline crashes and fails to build - The pipelines fail and require a restart of the jenkins node/pod thing or GKE or something. Once whatever it is gets restarted the build passes. This thing stops working a lot. Why is it so fragile? We need this to be up and running all the time. Example PR: https://github.com/fabric8-ui/fabric8-ui/pull/2302 Logs: https://jenkins.cd.test.fabric8.io/job/fabric8-ui/job/fabric8-ui/view/change-requests/job/PR-2302/5/console
build
pipeline crashes and fails to build the pipelines fail and require a restart of the jenkins node pod thing or gke or something once whatever it is gets restarted the build passes this thing stops working a lot why is it so fragile we need this to be up and running all the time example pr logs
1
1,329
2,729,697,462
IssuesEvent
2015-04-16 10:16:08
AdaptiveMe/adaptive-arp-javascript
https://api.github.com/repos/AdaptiveMe/adaptive-arp-javascript
closed
Push Adaptive Javascript Definitions to Bower Repository
build
Every Adaptive ARP new version push the Javascript Library to the repo created on #2 with a new tag in order to used for the new applications. The files has to be: adaptive.js (generated with ts compiler) adaptive.min.js (generated with Uglify) adaptive.min.js.map (generated with Uglify) adaptive.d.ts
1.0
Push Adaptive Javascript Definitions to Bower Repository - Every Adaptive ARP new version push the Javascript Library to the repo created on #2 with a new tag in order to used for the new applications. The files has to be: adaptive.js (generated with ts compiler) adaptive.min.js (generated with Uglify) adaptive.min.js.map (generated with Uglify) adaptive.d.ts
build
push adaptive javascript definitions to bower repository every adaptive arp new version push the javascript library to the repo created on with a new tag in order to used for the new applications the files has to be adaptive js generated with ts compiler adaptive min js generated with uglify adaptive min js map generated with uglify adaptive d ts
1
297,721
25,758,644,523
IssuesEvent
2022-12-08 18:27:07
openedx/build-test-release-wg
https://api.github.com/repos/openedx/build-test-release-wg
closed
olive-testing (TC_AUTHOR_45): Advanced problem types
bug release testing
# Passing Tests - [x] Blank advanced Problem - [x] Circuit Schematic Builder - [x] Custom JavaScript Display and Grading - [x] Custom Python-Evaluated Input - [x] Drag and Drop - [ ] Drag and Drop (Deprecated Version) - [x] Image Mapped Input - [x] Math Expression Input - [x] Problem with Adaptive Hint - [x] Staff Graded Points
1.0
olive-testing (TC_AUTHOR_45): Advanced problem types - # Passing Tests - [x] Blank advanced Problem - [x] Circuit Schematic Builder - [x] Custom JavaScript Display and Grading - [x] Custom Python-Evaluated Input - [x] Drag and Drop - [ ] Drag and Drop (Deprecated Version) - [x] Image Mapped Input - [x] Math Expression Input - [x] Problem with Adaptive Hint - [x] Staff Graded Points
non_build
olive testing tc author advanced problem types passing tests blank advanced problem circuit schematic builder custom javascript display and grading custom python evaluated input drag and drop drag and drop deprecated version image mapped input math expression input problem with adaptive hint staff graded points
0
38,291
10,166,661,977
IssuesEvent
2019-08-07 16:18:14
sactory/sactory
https://api.github.com/repos/sactory/sactory
closed
Custom events, including `append` and `remove` are not triggered on Firefox
bug runtime:builder
```html <div +append=console.log /> ``` does not log anything
1.0
Custom events, including `append` and `remove` are not triggered on Firefox - ```html <div +append=console.log /> ``` does not log anything
build
custom events including append and remove are not triggered on firefox html does not log anything
1
209,672
23,730,739,875
IssuesEvent
2022-08-31 01:18:33
jiw065/jiw065.github.io
https://api.github.com/repos/jiw065/jiw065.github.io
opened
CVE-2020-11022 (Medium) detected in jquery-3.1.1.min.js
security vulnerability
## CVE-2020-11022 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jquery-3.1.1.min.js</b></p></summary> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/3.1.1/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/3.1.1/jquery.min.js</a></p> <p>Path to dependency file: /MyPortfolio/index.html</p> <p>Path to vulnerable library: /MyPortfolio/index.html</p> <p> Dependency Hierarchy: - :x: **jquery-3.1.1.min.js** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In jQuery versions greater than or equal to 1.2 and before 3.5.0, passing HTML from untrusted sources - even after sanitizing it - to one of jQuery's DOM manipulation methods (i.e. .html(), .append(), and others) may execute untrusted code. This problem is patched in jQuery 3.5.0. <p>Publish Date: 2020-04-29 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11022>CVE-2020-11022</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://blog.jquery.com/2020/04/10/jquery-3-5-0-released/">https://blog.jquery.com/2020/04/10/jquery-3-5-0-released/</a></p> <p>Release Date: 2020-04-29</p> <p>Fix Resolution: jQuery - 3.5.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2020-11022 (Medium) detected in jquery-3.1.1.min.js - ## CVE-2020-11022 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jquery-3.1.1.min.js</b></p></summary> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/3.1.1/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/3.1.1/jquery.min.js</a></p> <p>Path to dependency file: /MyPortfolio/index.html</p> <p>Path to vulnerable library: /MyPortfolio/index.html</p> <p> Dependency Hierarchy: - :x: **jquery-3.1.1.min.js** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In jQuery versions greater than or equal to 1.2 and before 3.5.0, passing HTML from untrusted sources - even after sanitizing it - to one of jQuery's DOM manipulation methods (i.e. .html(), .append(), and others) may execute untrusted code. This problem is patched in jQuery 3.5.0. <p>Publish Date: 2020-04-29 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11022>CVE-2020-11022</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://blog.jquery.com/2020/04/10/jquery-3-5-0-released/">https://blog.jquery.com/2020/04/10/jquery-3-5-0-released/</a></p> <p>Release Date: 2020-04-29</p> <p>Fix Resolution: jQuery - 3.5.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_build
cve medium detected in jquery min js cve medium severity vulnerability vulnerable library jquery min js javascript library for dom operations library home page a href path to dependency file myportfolio index html path to vulnerable library myportfolio index html dependency hierarchy x jquery min js vulnerable library vulnerability details in jquery versions greater than or equal to and before passing html from untrusted sources even after sanitizing it to one of jquery s dom manipulation methods i e html append and others may execute untrusted code this problem is patched in jquery publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution jquery step up your open source security game with mend
0
91,257
26,339,746,127
IssuesEvent
2023-01-10 16:46:47
dotnet/arcade
https://api.github.com/repos/dotnet/arcade
closed
Build failed: dotnet-arcade-validation-official/main #20230110.1
Build Failed
Build [#20230110.1](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_build/results?buildId=2084297) partiallySucceeded ## :warning: : internal / dotnet-arcade-validation-official partiallySucceeded ### Summary **Finished** - Tue, 10 Jan 2023 15:47:10 GMT **Duration** - 116 minutes **Requested for** - DotNet Bot **Reason** - batchedCI ### Details #### Promote Arcade to '.NET Eng - Latest' channel - :warning: - [[Log]](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_apis/build/builds/2084297/logs/354) - The latest build on 'main' branch for the 'runtime' repository was not successful. - :warning: - [[Log]](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_apis/build/builds/2084297/logs/354) - The latest build on 'main' branch for the 'aspnetcore' repository was not successful. #### Publish using Darc - :warning: - [[Log]](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_apis/build/builds/2084297/logs/312) - There was a failure in sending the provision message: A timeout occurred while sending request to the remote provider. ### Changes - [07665182](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_git/017fb734-e4b4-4cc1-a90f-98a09ac25cd5/commit/07665182e45fc1aaf5a04ac9fcb73a4053259e5a) - dotnet-maestro[bot] - Update dependencies from https://github.com/dotnet/arcade build 20230110.1 (#3606) - [bbb340e6](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_git/017fb734-e4b4-4cc1-a90f-98a09ac25cd5/commit/bbb340e689757c2320e66379ab5ae2d9dede73e2) - dotnet-maestro[bot] - Update dependencies from https://github.com/dotnet/arcade build 20230109.3 (#3605) - [1d04b3e1](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_git/017fb734-e4b4-4cc1-a90f-98a09ac25cd5/commit/1d04b3e194a01a1fa590deeca7f5d2308874c0df) - dotnet-maestro[bot] - Update dependencies from https://github.com/dotnet/arcade build 20230109.2 (#3604)
1.0
Build failed: dotnet-arcade-validation-official/main #20230110.1 - Build [#20230110.1](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_build/results?buildId=2084297) partiallySucceeded ## :warning: : internal / dotnet-arcade-validation-official partiallySucceeded ### Summary **Finished** - Tue, 10 Jan 2023 15:47:10 GMT **Duration** - 116 minutes **Requested for** - DotNet Bot **Reason** - batchedCI ### Details #### Promote Arcade to '.NET Eng - Latest' channel - :warning: - [[Log]](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_apis/build/builds/2084297/logs/354) - The latest build on 'main' branch for the 'runtime' repository was not successful. - :warning: - [[Log]](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_apis/build/builds/2084297/logs/354) - The latest build on 'main' branch for the 'aspnetcore' repository was not successful. #### Publish using Darc - :warning: - [[Log]](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_apis/build/builds/2084297/logs/312) - There was a failure in sending the provision message: A timeout occurred while sending request to the remote provider. ### Changes - [07665182](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_git/017fb734-e4b4-4cc1-a90f-98a09ac25cd5/commit/07665182e45fc1aaf5a04ac9fcb73a4053259e5a) - dotnet-maestro[bot] - Update dependencies from https://github.com/dotnet/arcade build 20230110.1 (#3606) - [bbb340e6](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_git/017fb734-e4b4-4cc1-a90f-98a09ac25cd5/commit/bbb340e689757c2320e66379ab5ae2d9dede73e2) - dotnet-maestro[bot] - Update dependencies from https://github.com/dotnet/arcade build 20230109.3 (#3605) - [1d04b3e1](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_git/017fb734-e4b4-4cc1-a90f-98a09ac25cd5/commit/1d04b3e194a01a1fa590deeca7f5d2308874c0df) - dotnet-maestro[bot] - Update dependencies from https://github.com/dotnet/arcade build 20230109.2 (#3604)
build
build failed dotnet arcade validation official main build partiallysucceeded warning internal dotnet arcade validation official partiallysucceeded summary finished tue jan gmt duration minutes requested for dotnet bot reason batchedci details promote arcade to net eng latest channel warning the latest build on main branch for the runtime repository was not successful warning the latest build on main branch for the aspnetcore repository was not successful publish using darc warning there was a failure in sending the provision message a timeout occurred while sending request to the remote provider changes dotnet maestro update dependencies from build dotnet maestro update dependencies from build dotnet maestro update dependencies from build
1
10,256
4,028,973,861
IssuesEvent
2016-05-18 08:53:05
zturtleman/spearmint
https://api.github.com/repos/zturtleman/spearmint
opened
Activate entity regression on nodm7
bug gamecode
Activate entity regression on [Claustrophobopolis 2003](http://de.ws.q3df.org/map/nodm7/) reportedly caused by https://github.com/zturtleman/mint-arena/compare/f10b0870f6df...116ec08b3843 changes for #220.
1.0
Activate entity regression on nodm7 - Activate entity regression on [Claustrophobopolis 2003](http://de.ws.q3df.org/map/nodm7/) reportedly caused by https://github.com/zturtleman/mint-arena/compare/f10b0870f6df...116ec08b3843 changes for #220.
non_build
activate entity regression on activate entity regression on reportedly caused by changes for
0
86,116
10,474,314,009
IssuesEvent
2019-09-23 14:17:38
weaveworks/wks-quickstart-firekube
https://api.github.com/repos/weaveworks/wks-quickstart-firekube
closed
Add a step that shows gitops in action
documentation enhancement
Once the cluster is up and running we should be able to demo gitops by telling the user to add a manifest to their fork and observe it being deployed to the cluster.
1.0
Add a step that shows gitops in action - Once the cluster is up and running we should be able to demo gitops by telling the user to add a manifest to their fork and observe it being deployed to the cluster.
non_build
add a step that shows gitops in action once the cluster is up and running we should be able to demo gitops by telling the user to add a manifest to their fork and observe it being deployed to the cluster
0
312,256
26,852,300,169
IssuesEvent
2023-02-03 12:02:40
galaxyproject/training-material
https://api.github.com/repos/galaxyproject/training-material
closed
Rule Based Uploader: Advanced - Outdated Info
bug galaxy-interface galaxy-release-testing
Release Testing 23.0 * “Building URLs from Accession Information”, Hands-on: Uniport Outdated Info and Screenshots - Format can now be found under Download. Format menu point no longer exists - Page now has an api url generator, which may be a better fit for this tutorial. It is found under the parent section, and is formated: `https://rest.uniprot.org/uniprotkb/stream?format=fasta&query=accession%{identifier}` * "Apply Rules to Existing Collections", Step 3: Click `Build` should be Click `Upload` * Outdated Screenshots showing the History
1.0
Rule Based Uploader: Advanced - Outdated Info - Release Testing 23.0 * “Building URLs from Accession Information”, Hands-on: Uniport Outdated Info and Screenshots - Format can now be found under Download. Format menu point no longer exists - Page now has an api url generator, which may be a better fit for this tutorial. It is found under the parent section, and is formated: `https://rest.uniprot.org/uniprotkb/stream?format=fasta&query=accession%{identifier}` * "Apply Rules to Existing Collections", Step 3: Click `Build` should be Click `Upload` * Outdated Screenshots showing the History
non_build
rule based uploader advanced outdated info release testing “building urls from accession information” hands on uniport outdated info and screenshots format can now be found under download format menu point no longer exists page now has an api url generator which may be a better fit for this tutorial it is found under the parent section and is formated apply rules to existing collections step click build should be click upload outdated screenshots showing the history
0
17,886
3,644,828,375
IssuesEvent
2016-02-15 11:47:20
hu-semesterprojekt-genombrowser/Alazen-Middleware
https://api.github.com/repos/hu-semesterprojekt-genombrowser/Alazen-Middleware
closed
Ant File anpassen
enhancement testing
Die Datei build.xml muss an eure Struktur angepasst werden. Einer Erklärung, wie Ant funktioniert, ist unter [https://ant.apache.org/manual/tutorial-HelloWorldWithAnt.html] zu finden. Es fehlt doch: * Tests bauen und ausführen * ~~Dokumentation bauen~~
1.0
Ant File anpassen - Die Datei build.xml muss an eure Struktur angepasst werden. Einer Erklärung, wie Ant funktioniert, ist unter [https://ant.apache.org/manual/tutorial-HelloWorldWithAnt.html] zu finden. Es fehlt doch: * Tests bauen und ausführen * ~~Dokumentation bauen~~
non_build
ant file anpassen die datei build xml muss an eure struktur angepasst werden einer erklärung wie ant funktioniert ist unter zu finden es fehlt doch tests bauen und ausführen dokumentation bauen
0
265,262
23,156,535,719
IssuesEvent
2022-07-29 13:32:07
simpleworks-gmbh/staf
https://api.github.com/repos/simpleworks-gmbh/staf
closed
Proxy can't connect to the expected port
bug framework testing v0.120.13
If one starts several API Testcase, from another Testcase (manually via executeTestcase Method). The configured port, will be used for all Proxy Server instances., Only the first Proxy will run, but the other proxies (for the APITestcases left) won't run. That will fail with the error message 2022-07-11 14:34:16.481 ERROR HttpClient (HttpClient.java:103) - can't set up Response. java.net.ConnectException: Failed to connect to /0.0.0.0:8893 at okhttp3.internal.io.RealConnection.connectSocket(RealConnection.java:187) ~[okhttp-3.3.1.jar:?] at okhttp3.internal.io.RealConnection.buildTunneledConnection(RealConnection.java:151) ~[okhttp-3.3.1.jar:?] at okhttp3.internal.io.RealConnection.connect(RealConnection.java:108) ~[okhttp-3.3.1.jar:?] at okhttp3.internal.http.StreamAllocation.findConnection(StreamAllocation.java:187) ~[okhttp-3.3.1.jar:?] at okhttp3.internal.http.StreamAllocation.findHealthyConnection(StreamAllocation.java:123) ~[okhttp-3.3.1.jar:?] at okhttp3.internal.http.StreamAllocation.newStream(StreamAllocation.java:93) ~[okhttp-3.3.1.jar:?] at okhttp3.internal.http.HttpEngine.connect(HttpEngine.java:296) ~[okhttp-3.3.1.jar:?] at okhttp3.internal.http.HttpEngine.sendRequest(HttpEngine.java:248) ~[okhttp-3.3.1.jar:?] at okhttp3.RealCall.getResponse(RealCall.java:243) ~[okhttp-3.3.1.jar:?] at okhttp3.RealCall$ApplicationInterceptorChain.proceed(RealCall.java:201) ~[okhttp-3.3.1.jar:?] at okhttp3.RealCall.getResponseWithInterceptorChain(RealCall.java:163) ~[okhttp-3.3.1.jar:?] at okhttp3.RealCall.execute(RealCall.java:57) ~[okhttp-3.3.1.jar:?] at de.simpleworks.staf.framework.api.httpclient.HttpClient.doRequest(HttpClient.java:89) [framework-0.120.9.jar:?] at de.simpleworks.staf.framework.elements.api.APITestCase.doRequest(APITestCase.java:323) [framework-0.120.9.jar:?] at de.simpleworks.staf.framework.elements.api.APITestCase.checkRequest(APITestCase.java:258) [framework-0.120.9.jar:?] at de.simpleworks.staf.framework.elements.api.APITestCase.executeTestStep(APITestCase.java:498) [framework-0.120.9.jar:?] at de.simpleworks.staf.framework.elements.commons.TestCase.executeTestcase(TestCase.java:325) [framework-0.120.9.jar:?] at package.executeTestcase(TestCaseIntegration.kt:48) [main/:?] at package.step5(TEST_CASE_ID.kt:112) [test/:?] at jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) ~[?:?] at jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:566) ~[?:?] at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59) [junit-4.13.1.jar:4.13.1] at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) [junit-4.13.1.jar:4.13.1] at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56) [junit-4.13.1.jar:4.13.1] at de.simpleworks.staf.module.junit4.STAFInvokeMethod.evaluate(STAFInvokeMethod.java:88) [junit4-module-0.120.9.jar:?] at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26) [junit-4.13.1.jar:4.13.1] at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) [junit-4.13.1.jar:4.13.1] at org.junit.runners.BlockJUnit4ClassRunner$1.evaluate(BlockJUnit4ClassRunner.java:100) [junit-4.13.1.jar:4.13.1] at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:366) [junit-4.13.1.jar:4.13.1] at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:103) [junit-4.13.1.jar:4.13.1] at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:63) [junit-4.13.1.jar:4.13.1] at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) [junit-4.13.1.jar:4.13.1] at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) [junit-4.13.1.jar:4.13.1] at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) [junit-4.13.1.jar:4.13.1] at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) [junit-4.13.1.jar:4.13.1] at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) [junit-4.13.1.jar:4.13.1] at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) [junit-4.13.1.jar:4.13.1] at org.junit.runners.ParentRunner.run(ParentRunner.java:413) [junit-4.13.1.jar:4.13.1] at org.junit.runner.JUnitCore.run(JUnitCore.java:137) [junit-4.13.1.jar:4.13.1] at com.intellij.junit4.JUnit4IdeaTestRunner.startRunnerWithArgs(JUnit4IdeaTestRunner.java:69) [junit-rt.jar:?] at com.intellij.rt.junit.IdeaTestRunner$Repeater.startRunnerWithArgs(IdeaTestRunner.java:33) [junit-rt.jar:?] at com.intellij.rt.junit.JUnitStarter.prepareStreamsAndStart(JUnitStarter.java:220) [junit-rt.jar:?] at com.intellij.rt.junit.JUnitStarter.main(JUnitStarter.java:53) [junit-rt.jar:?]
1.0
Proxy can't connect to the expected port - If one starts several API Testcase, from another Testcase (manually via executeTestcase Method). The configured port, will be used for all Proxy Server instances., Only the first Proxy will run, but the other proxies (for the APITestcases left) won't run. That will fail with the error message 2022-07-11 14:34:16.481 ERROR HttpClient (HttpClient.java:103) - can't set up Response. java.net.ConnectException: Failed to connect to /0.0.0.0:8893 at okhttp3.internal.io.RealConnection.connectSocket(RealConnection.java:187) ~[okhttp-3.3.1.jar:?] at okhttp3.internal.io.RealConnection.buildTunneledConnection(RealConnection.java:151) ~[okhttp-3.3.1.jar:?] at okhttp3.internal.io.RealConnection.connect(RealConnection.java:108) ~[okhttp-3.3.1.jar:?] at okhttp3.internal.http.StreamAllocation.findConnection(StreamAllocation.java:187) ~[okhttp-3.3.1.jar:?] at okhttp3.internal.http.StreamAllocation.findHealthyConnection(StreamAllocation.java:123) ~[okhttp-3.3.1.jar:?] at okhttp3.internal.http.StreamAllocation.newStream(StreamAllocation.java:93) ~[okhttp-3.3.1.jar:?] at okhttp3.internal.http.HttpEngine.connect(HttpEngine.java:296) ~[okhttp-3.3.1.jar:?] at okhttp3.internal.http.HttpEngine.sendRequest(HttpEngine.java:248) ~[okhttp-3.3.1.jar:?] at okhttp3.RealCall.getResponse(RealCall.java:243) ~[okhttp-3.3.1.jar:?] at okhttp3.RealCall$ApplicationInterceptorChain.proceed(RealCall.java:201) ~[okhttp-3.3.1.jar:?] at okhttp3.RealCall.getResponseWithInterceptorChain(RealCall.java:163) ~[okhttp-3.3.1.jar:?] at okhttp3.RealCall.execute(RealCall.java:57) ~[okhttp-3.3.1.jar:?] at de.simpleworks.staf.framework.api.httpclient.HttpClient.doRequest(HttpClient.java:89) [framework-0.120.9.jar:?] at de.simpleworks.staf.framework.elements.api.APITestCase.doRequest(APITestCase.java:323) [framework-0.120.9.jar:?] at de.simpleworks.staf.framework.elements.api.APITestCase.checkRequest(APITestCase.java:258) [framework-0.120.9.jar:?] at de.simpleworks.staf.framework.elements.api.APITestCase.executeTestStep(APITestCase.java:498) [framework-0.120.9.jar:?] at de.simpleworks.staf.framework.elements.commons.TestCase.executeTestcase(TestCase.java:325) [framework-0.120.9.jar:?] at package.executeTestcase(TestCaseIntegration.kt:48) [main/:?] at package.step5(TEST_CASE_ID.kt:112) [test/:?] at jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) ~[?:?] at jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:566) ~[?:?] at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59) [junit-4.13.1.jar:4.13.1] at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) [junit-4.13.1.jar:4.13.1] at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56) [junit-4.13.1.jar:4.13.1] at de.simpleworks.staf.module.junit4.STAFInvokeMethod.evaluate(STAFInvokeMethod.java:88) [junit4-module-0.120.9.jar:?] at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26) [junit-4.13.1.jar:4.13.1] at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) [junit-4.13.1.jar:4.13.1] at org.junit.runners.BlockJUnit4ClassRunner$1.evaluate(BlockJUnit4ClassRunner.java:100) [junit-4.13.1.jar:4.13.1] at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:366) [junit-4.13.1.jar:4.13.1] at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:103) [junit-4.13.1.jar:4.13.1] at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:63) [junit-4.13.1.jar:4.13.1] at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) [junit-4.13.1.jar:4.13.1] at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) [junit-4.13.1.jar:4.13.1] at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) [junit-4.13.1.jar:4.13.1] at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) [junit-4.13.1.jar:4.13.1] at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) [junit-4.13.1.jar:4.13.1] at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) [junit-4.13.1.jar:4.13.1] at org.junit.runners.ParentRunner.run(ParentRunner.java:413) [junit-4.13.1.jar:4.13.1] at org.junit.runner.JUnitCore.run(JUnitCore.java:137) [junit-4.13.1.jar:4.13.1] at com.intellij.junit4.JUnit4IdeaTestRunner.startRunnerWithArgs(JUnit4IdeaTestRunner.java:69) [junit-rt.jar:?] at com.intellij.rt.junit.IdeaTestRunner$Repeater.startRunnerWithArgs(IdeaTestRunner.java:33) [junit-rt.jar:?] at com.intellij.rt.junit.JUnitStarter.prepareStreamsAndStart(JUnitStarter.java:220) [junit-rt.jar:?] at com.intellij.rt.junit.JUnitStarter.main(JUnitStarter.java:53) [junit-rt.jar:?]
non_build
proxy can t connect to the expected port if one starts several api testcase from another testcase manually via executetestcase method the configured port will be used for all proxy server instances only the first proxy will run but the other proxies for the apitestcases left won t run that will fail with the error message error httpclient httpclient java can t set up response java net connectexception failed to connect to at internal io realconnection connectsocket realconnection java at internal io realconnection buildtunneledconnection realconnection java at internal io realconnection connect realconnection java at internal http streamallocation findconnection streamallocation java at internal http streamallocation findhealthyconnection streamallocation java at internal http streamallocation newstream streamallocation java at internal http httpengine connect httpengine java at internal http httpengine sendrequest httpengine java at realcall getresponse realcall java at realcall applicationinterceptorchain proceed realcall java at realcall getresponsewithinterceptorchain realcall java at realcall execute realcall java at de simpleworks staf framework api httpclient httpclient dorequest httpclient java at de simpleworks staf framework elements api apitestcase dorequest apitestcase java at de simpleworks staf framework elements api apitestcase checkrequest apitestcase java at de simpleworks staf framework elements api apitestcase executeteststep apitestcase java at de simpleworks staf framework elements commons testcase executetestcase testcase java at package executetestcase testcaseintegration kt at package test case id kt at jdk internal reflect nativemethodaccessorimpl native method at jdk internal reflect nativemethodaccessorimpl invoke nativemethodaccessorimpl java at jdk internal reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java lang reflect method invoke method java at org junit runners model frameworkmethod runreflectivecall frameworkmethod java at org junit internal runners model reflectivecallable run reflectivecallable java at org junit runners model frameworkmethod invokeexplosively frameworkmethod java at de simpleworks staf module stafinvokemethod evaluate stafinvokemethod java at org junit internal runners statements runbefores evaluate runbefores java at org junit runners parentrunner evaluate parentrunner java at org junit runners evaluate java at org junit runners parentrunner runleaf parentrunner java at org junit runners runchild java at org junit runners runchild java at org junit runners parentrunner run parentrunner java at org junit runners parentrunner schedule parentrunner java at org junit runners parentrunner runchildren parentrunner java at org junit runners parentrunner access parentrunner java at org junit runners parentrunner evaluate parentrunner java at org junit runners parentrunner evaluate parentrunner java at org junit runners parentrunner run parentrunner java at org junit runner junitcore run junitcore java at com intellij startrunnerwithargs java at com intellij rt junit ideatestrunner repeater startrunnerwithargs ideatestrunner java at com intellij rt junit junitstarter preparestreamsandstart junitstarter java at com intellij rt junit junitstarter main junitstarter java
0
159,537
20,068,329,610
IssuesEvent
2022-02-04 01:12:23
panasalap/frameworks_av
https://api.github.com/repos/panasalap/frameworks_av
closed
CVE-2020-0169 (Medium) detected in avandroid-10.0.0_r46 - autoclosed
security vulnerability
## CVE-2020-0169 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>avandroid-10.0.0_r46</b></p></summary> <p> <p>Library home page: <a href=https://android.googlesource.com/platform/frameworks/av>https://android.googlesource.com/platform/frameworks/av</a></p> <p>Found in HEAD commit: <a href="https://github.com/panasalap/frameworks_av/commit/0fef4b48931b6a0ff4bfc61cfbc4e16710e0f08b">0fef4b48931b6a0ff4bfc61cfbc4e16710e0f08b</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/media/extractors/midi/MidiExtractor.cpp</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In RTTTL_Event of eas_rtttl.c, there is possible resource exhaustion due to a missing bounds check. This could lead to remote denial of service with no additional execution privileges needed. User interaction is needed for exploitation.Product: AndroidVersions: Android-10Android ID: A-123700383 <p>Publish Date: 2020-06-11 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-0169>CVE-2020-0169</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://android.googlesource.com/platform/frameworks/av/+/refs/tags/android-10.0.0_r37">https://android.googlesource.com/platform/frameworks/av/+/refs/tags/android-10.0.0_r37</a></p> <p>Release Date: 2020-06-11</p> <p>Fix Resolution: android-10.0.0_r37</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2020-0169 (Medium) detected in avandroid-10.0.0_r46 - autoclosed - ## CVE-2020-0169 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>avandroid-10.0.0_r46</b></p></summary> <p> <p>Library home page: <a href=https://android.googlesource.com/platform/frameworks/av>https://android.googlesource.com/platform/frameworks/av</a></p> <p>Found in HEAD commit: <a href="https://github.com/panasalap/frameworks_av/commit/0fef4b48931b6a0ff4bfc61cfbc4e16710e0f08b">0fef4b48931b6a0ff4bfc61cfbc4e16710e0f08b</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/media/extractors/midi/MidiExtractor.cpp</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In RTTTL_Event of eas_rtttl.c, there is possible resource exhaustion due to a missing bounds check. This could lead to remote denial of service with no additional execution privileges needed. User interaction is needed for exploitation.Product: AndroidVersions: Android-10Android ID: A-123700383 <p>Publish Date: 2020-06-11 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-0169>CVE-2020-0169</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://android.googlesource.com/platform/frameworks/av/+/refs/tags/android-10.0.0_r37">https://android.googlesource.com/platform/frameworks/av/+/refs/tags/android-10.0.0_r37</a></p> <p>Release Date: 2020-06-11</p> <p>Fix Resolution: android-10.0.0_r37</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_build
cve medium detected in avandroid autoclosed cve medium severity vulnerability vulnerable library avandroid library home page a href found in head commit a href found in base branch master vulnerable source files media extractors midi midiextractor cpp vulnerability details in rtttl event of eas rtttl c there is possible resource exhaustion due to a missing bounds check this could lead to remote denial of service with no additional execution privileges needed user interaction is needed for exploitation product androidversions android id a publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution android step up your open source security game with whitesource
0
458,045
13,167,428,812
IssuesEvent
2020-08-11 10:15:15
StrangeLoopGames/EcoIssues
https://api.github.com/repos/StrangeLoopGames/EcoIssues
closed
[0.9.0.0 - dev-23] Modding - Server crash on config manipulation
Category: Modkit Priority: Medium
While developing DiscordLink is noticed the server would crash when any element (that is not a collection) was changed in the config via the server UI. For booleans, this happens right away and for other types it crashes upon hitting enter while the input field has focus. This is a huge issue for the 9.0 update of the mod as we recommend our users to do configuration via the server interface. I've made a minimal project that reproduces the issue. I used DLLs fetched from the temp directory of the dev-23 server for this. [EcoServerCrashMVP.zip](https://github.com/StrangeLoopGames/EcoIssues/files/5033963/EcoServerCrashMVP.zip) Here is the DLL output of the above project (compiled with debug flags) [EcoServerCrashMVP.zip](https://github.com/StrangeLoopGames/EcoIssues/files/5033974/EcoServerCrashMVP.zip) Steps to reproduce: 1. Add the mod DLL to the server. 2. Start the server. 3. In the server UI, navigate to EcoServerCrashMVP.MVP (probably did something odd to get that name but that shouldn't be related to the crash). 4. Change the value of the "The Bool" field. 5. Observe NullReferenceException outside of the plugin code. -------------------------- Exception + Stack trace -------------------------- ``` System.NullReferenceException: Object reference not set to an instance of an object. at Eco.Server.UI.ServerMainForm.<>c__DisplayClass7_0.<BuildEditor>b__0(Object o, PropertyValueChangedEventArgs e) at System.Windows.Forms.PropertyGrid.OnPropertyValueChanged(PropertyValueChangedEventArgs e) at System.Windows.Forms.PropertyGrid.OnPropertyValueSet(GridItem changedItem, Object oldValue) at System.Windows.Forms.PropertyGridInternal.PropertyGridView.CommitValue(GridEntry ipeCur, Object value, Boolean closeDropDown) at System.Windows.Forms.PropertyGridInternal.PropertyGridView.ProcessEnumUpAndDown(GridEntry gridEntry, Keys keyCode, Boolean closeDropDown) at System.Windows.Forms.PropertyGridInternal.PropertyGridView.OnEditKeyDown(Object sender, KeyEventArgs ke) at System.Windows.Forms.Control.OnKeyDown(KeyEventArgs e) at System.Windows.Forms.PropertyGridInternal.PropertyGridView.GridViewEdit.OnKeyDown(KeyEventArgs ke) at System.Windows.Forms.Control.ProcessKeyEventArgs(Message& m) at System.Windows.Forms.Control.ProcessKeyMessage(Message& m) at System.Windows.Forms.Control.WndProc(Message& m) at System.Windows.Forms.TextBox.WndProc(Message& m) at System.Windows.Forms.PropertyGridInternal.PropertyGridView.GridViewEdit.WndProc(Message& m) at System.Windows.Forms.NativeWindow.Callback(IntPtr hWnd, Int32 msg, IntPtr wparam, IntPtr lparam) ``` -------------------------- Plugin code reference -------------------------- ``` using System.ComponentModel; using Eco.Core.Plugins; using Eco.Core.Plugins.Interfaces; using Eco.Core.Utils; namespace EcoServerCrashMVP { public class MVP : IModKitPlugin, IConfigurablePlugin, IInitializablePlugin { public ThreadSafeAction<object, string> ParamChanged { get; set; } private PluginConfig<TinyConfig> _configOptions; public string GetStatus() { return _configOptions.Config.TheBool ? "Bool is True" : "Bool is False"; } public IPluginConfig PluginConfig { get { return _configOptions; } } public object GetEditObject() { return _configOptions.Config; } public void OnEditObjectChanged(object o, string param) { } public void Initialize(TimedTask timer) { _configOptions = new PluginConfig<TinyConfig>("CrashMVP"); } } public class TinyConfig { [Description("Flip to crash!"), Category("Crashing")] public bool TheBool { get; set; } = false; } } ```
1.0
[0.9.0.0 - dev-23] Modding - Server crash on config manipulation - While developing DiscordLink is noticed the server would crash when any element (that is not a collection) was changed in the config via the server UI. For booleans, this happens right away and for other types it crashes upon hitting enter while the input field has focus. This is a huge issue for the 9.0 update of the mod as we recommend our users to do configuration via the server interface. I've made a minimal project that reproduces the issue. I used DLLs fetched from the temp directory of the dev-23 server for this. [EcoServerCrashMVP.zip](https://github.com/StrangeLoopGames/EcoIssues/files/5033963/EcoServerCrashMVP.zip) Here is the DLL output of the above project (compiled with debug flags) [EcoServerCrashMVP.zip](https://github.com/StrangeLoopGames/EcoIssues/files/5033974/EcoServerCrashMVP.zip) Steps to reproduce: 1. Add the mod DLL to the server. 2. Start the server. 3. In the server UI, navigate to EcoServerCrashMVP.MVP (probably did something odd to get that name but that shouldn't be related to the crash). 4. Change the value of the "The Bool" field. 5. Observe NullReferenceException outside of the plugin code. -------------------------- Exception + Stack trace -------------------------- ``` System.NullReferenceException: Object reference not set to an instance of an object. at Eco.Server.UI.ServerMainForm.<>c__DisplayClass7_0.<BuildEditor>b__0(Object o, PropertyValueChangedEventArgs e) at System.Windows.Forms.PropertyGrid.OnPropertyValueChanged(PropertyValueChangedEventArgs e) at System.Windows.Forms.PropertyGrid.OnPropertyValueSet(GridItem changedItem, Object oldValue) at System.Windows.Forms.PropertyGridInternal.PropertyGridView.CommitValue(GridEntry ipeCur, Object value, Boolean closeDropDown) at System.Windows.Forms.PropertyGridInternal.PropertyGridView.ProcessEnumUpAndDown(GridEntry gridEntry, Keys keyCode, Boolean closeDropDown) at System.Windows.Forms.PropertyGridInternal.PropertyGridView.OnEditKeyDown(Object sender, KeyEventArgs ke) at System.Windows.Forms.Control.OnKeyDown(KeyEventArgs e) at System.Windows.Forms.PropertyGridInternal.PropertyGridView.GridViewEdit.OnKeyDown(KeyEventArgs ke) at System.Windows.Forms.Control.ProcessKeyEventArgs(Message& m) at System.Windows.Forms.Control.ProcessKeyMessage(Message& m) at System.Windows.Forms.Control.WndProc(Message& m) at System.Windows.Forms.TextBox.WndProc(Message& m) at System.Windows.Forms.PropertyGridInternal.PropertyGridView.GridViewEdit.WndProc(Message& m) at System.Windows.Forms.NativeWindow.Callback(IntPtr hWnd, Int32 msg, IntPtr wparam, IntPtr lparam) ``` -------------------------- Plugin code reference -------------------------- ``` using System.ComponentModel; using Eco.Core.Plugins; using Eco.Core.Plugins.Interfaces; using Eco.Core.Utils; namespace EcoServerCrashMVP { public class MVP : IModKitPlugin, IConfigurablePlugin, IInitializablePlugin { public ThreadSafeAction<object, string> ParamChanged { get; set; } private PluginConfig<TinyConfig> _configOptions; public string GetStatus() { return _configOptions.Config.TheBool ? "Bool is True" : "Bool is False"; } public IPluginConfig PluginConfig { get { return _configOptions; } } public object GetEditObject() { return _configOptions.Config; } public void OnEditObjectChanged(object o, string param) { } public void Initialize(TimedTask timer) { _configOptions = new PluginConfig<TinyConfig>("CrashMVP"); } } public class TinyConfig { [Description("Flip to crash!"), Category("Crashing")] public bool TheBool { get; set; } = false; } } ```
non_build
modding server crash on config manipulation while developing discordlink is noticed the server would crash when any element that is not a collection was changed in the config via the server ui for booleans this happens right away and for other types it crashes upon hitting enter while the input field has focus this is a huge issue for the update of the mod as we recommend our users to do configuration via the server interface i ve made a minimal project that reproduces the issue i used dlls fetched from the temp directory of the dev server for this here is the dll output of the above project compiled with debug flags steps to reproduce add the mod dll to the server start the server in the server ui navigate to ecoservercrashmvp mvp probably did something odd to get that name but that shouldn t be related to the crash change the value of the the bool field observe nullreferenceexception outside of the plugin code exception stack trace system nullreferenceexception object reference not set to an instance of an object at eco server ui servermainform c b object o propertyvaluechangedeventargs e at system windows forms propertygrid onpropertyvaluechanged propertyvaluechangedeventargs e at system windows forms propertygrid onpropertyvalueset griditem changeditem object oldvalue at system windows forms propertygridinternal propertygridview commitvalue gridentry ipecur object value boolean closedropdown at system windows forms propertygridinternal propertygridview processenumupanddown gridentry gridentry keys keycode boolean closedropdown at system windows forms propertygridinternal propertygridview oneditkeydown object sender keyeventargs ke at system windows forms control onkeydown keyeventargs e at system windows forms propertygridinternal propertygridview gridviewedit onkeydown keyeventargs ke at system windows forms control processkeyeventargs message m at system windows forms control processkeymessage message m at system windows forms control wndproc message m at system windows forms textbox wndproc message m at system windows forms propertygridinternal propertygridview gridviewedit wndproc message m at system windows forms nativewindow callback intptr hwnd msg intptr wparam intptr lparam plugin code reference using system componentmodel using eco core plugins using eco core plugins interfaces using eco core utils namespace ecoservercrashmvp public class mvp imodkitplugin iconfigurableplugin iinitializableplugin public threadsafeaction paramchanged get set private pluginconfig configoptions public string getstatus return configoptions config thebool bool is true bool is false public ipluginconfig pluginconfig get return configoptions public object geteditobject return configoptions config public void oneditobjectchanged object o string param public void initialize timedtask timer configoptions new pluginconfig crashmvp public class tinyconfig public bool thebool get set false
0
5,081
7,603,776,161
IssuesEvent
2018-04-29 17:59:29
hawthorne-git/hawthorne-admin
https://api.github.com/repos/hawthorne-git/hawthorne-admin
opened
Checkup Job
Bug Priority - High Effort Required - Medium v2 Requirement
Validate that each product has an image, and its image is valid (EX: for re-sizing)
1.0
Checkup Job - Validate that each product has an image, and its image is valid (EX: for re-sizing)
non_build
checkup job validate that each product has an image and its image is valid ex for re sizing
0
104,415
4,211,085,495
IssuesEvent
2016-06-29 12:25:29
play2-maven-plugin/play2-maven-plugin
https://api.github.com/repos/play2-maven-plugin/play2-maven-plugin
closed
Upgrade minimum supported Maven version to 3.0
Component-Maven-Plugin Priority-Medium
It will be required for running Play! in development mode with hot reloading (issue #6).
1.0
Upgrade minimum supported Maven version to 3.0 - It will be required for running Play! in development mode with hot reloading (issue #6).
non_build
upgrade minimum supported maven version to it will be required for running play in development mode with hot reloading issue
0
85,167
10,429,827,736
IssuesEvent
2019-09-17 04:26:56
StarChart-Labs/operations
https://api.github.com/repos/StarChart-Labs/operations
opened
Standard Release Scope Handling
documentation enhancement
Document a standard practice for determining the scope a change requires in release versioning (major, minor, micro), and a standardized implementation for specifying it. Currently, flare-plugins defines a decent starting implementation where there are labels for major, minor, and micro release scopes, and the "next-release" milestone always includes the minimum scope requried
1.0
Standard Release Scope Handling - Document a standard practice for determining the scope a change requires in release versioning (major, minor, micro), and a standardized implementation for specifying it. Currently, flare-plugins defines a decent starting implementation where there are labels for major, minor, and micro release scopes, and the "next-release" milestone always includes the minimum scope requried
non_build
standard release scope handling document a standard practice for determining the scope a change requires in release versioning major minor micro and a standardized implementation for specifying it currently flare plugins defines a decent starting implementation where there are labels for major minor and micro release scopes and the next release milestone always includes the minimum scope requried
0
52,114
12,878,892,334
IssuesEvent
2020-07-11 19:00:11
PowerShell/PowerShell
https://api.github.com/repos/PowerShell/PowerShell
closed
Azure Pipelines clock issues
Area-Build Issue-Question Resolution-Answered Resolution-External
In build 55800 the clock keeps jumping back and forth about 16 hours: 2020-06-18T16:46:46.1993748Z 2020-06-19T09:28:05.5534482Z https://dev.azure.com/powershell/2972bb5c-f20c-4a60-8bd9-00ffe9987edc/_apis/build/builds/55800/logs/35
1.0
Azure Pipelines clock issues - In build 55800 the clock keeps jumping back and forth about 16 hours: 2020-06-18T16:46:46.1993748Z 2020-06-19T09:28:05.5534482Z https://dev.azure.com/powershell/2972bb5c-f20c-4a60-8bd9-00ffe9987edc/_apis/build/builds/55800/logs/35
build
azure pipelines clock issues in build the clock keeps jumping back and forth about hours
1
27,989
8,058,156,988
IssuesEvent
2018-08-02 17:34:22
PowerShell/PowerShell
https://api.github.com/repos/PowerShell/PowerShell
closed
Microsoft.PowerShell.SDK version 6.1.0-preview.4 fails at InitialSessionState.CreateDefault()
Area-Build
I'm unable to use preview.4 of the SDK with my test project. preview.3 worked fine. The test project targets .net core 2.1. Steps to reproduce ------------------ 1. Clone https://github.com/jherby2k/AudioWorks.git. 1. Run the tests in the AudioWorks.Commands.Tests project. All should pass. 1. Change the project to reference Microsoft.PowerShell.SDK version 6.1.0-preview.4 (from preview.3) 1. Run the tests again. Most will fail. Expected behavior ----------------- all tests pass Actual behavior --------------- > One or more errors occurred. (Cannot load PowerShell snap-in Microsoft.PowerShell.Diagnostics because of the following error: Could not load file or assembly '..\Source\Repos\AudioWorks\AudioWorks\tests\AudioWorks.Commands.Tests\bin\Release-Windows\netcoreapp2.1\Microsoft.PowerShell.Commands'. The system cannot find the file specified.
1.0
Microsoft.PowerShell.SDK version 6.1.0-preview.4 fails at InitialSessionState.CreateDefault() - I'm unable to use preview.4 of the SDK with my test project. preview.3 worked fine. The test project targets .net core 2.1. Steps to reproduce ------------------ 1. Clone https://github.com/jherby2k/AudioWorks.git. 1. Run the tests in the AudioWorks.Commands.Tests project. All should pass. 1. Change the project to reference Microsoft.PowerShell.SDK version 6.1.0-preview.4 (from preview.3) 1. Run the tests again. Most will fail. Expected behavior ----------------- all tests pass Actual behavior --------------- > One or more errors occurred. (Cannot load PowerShell snap-in Microsoft.PowerShell.Diagnostics because of the following error: Could not load file or assembly '..\Source\Repos\AudioWorks\AudioWorks\tests\AudioWorks.Commands.Tests\bin\Release-Windows\netcoreapp2.1\Microsoft.PowerShell.Commands'. The system cannot find the file specified.
build
microsoft powershell sdk version preview fails at initialsessionstate createdefault i m unable to use preview of the sdk with my test project preview worked fine the test project targets net core steps to reproduce clone run the tests in the audioworks commands tests project all should pass change the project to reference microsoft powershell sdk version preview from preview run the tests again most will fail expected behavior all tests pass actual behavior one or more errors occurred cannot load powershell snap in microsoft powershell diagnostics because of the following error could not load file or assembly source repos audioworks audioworks tests audioworks commands tests bin release windows microsoft powershell commands the system cannot find the file specified
1
11,132
4,892,462,668
IssuesEvent
2016-11-18 19:47:59
numenta/nupic.core
https://api.github.com/repos/numenta/nupic.core
closed
Reuse VERSION file in Doxygen
priority:4 triage type:build type:cleanup type:deployment
Currently, the version is specified in both places. The Doxygen file should reuse the VERSION file.
1.0
Reuse VERSION file in Doxygen - Currently, the version is specified in both places. The Doxygen file should reuse the VERSION file.
build
reuse version file in doxygen currently the version is specified in both places the doxygen file should reuse the version file
1
78,676
22,345,789,526
IssuesEvent
2022-06-15 07:40:50
isl-org/Open3D
https://api.github.com/repos/isl-org/Open3D
opened
Summarize the issue and your environments (e.g., "Cannot compile on Ubuntu 20.04 with OpenBLAS")
build/install issue
### Checklist - [X] I have searched for [similar issues](https://github.com/isl-org/Open3D/issues). - [X] For Python issues, I have tested with the [latest development wheel](http://www.open3d.org/docs/latest/getting_started.html#development-version-pip). - [X] I have checked the [release documentation](http://www.open3d.org/docs/release/) and the [latest documentation](http://www.open3d.org/docs/latest/) (for `master` branch). ### Steps to reproduce the issue #### I first cloned Open3D by: ``` git clone https://github.com/isl-org/Open3D.git cd Open3D ``` #### Then, I build Open3D (on Ubuntu 20.04, with CUDA 11.5) with: ``` mkdir build cd build mkdir build cd build cmake .. -G "Visual Studio 15 2017" -A x64 -DBUILD_CUDA_MODULE=ON -DBUILD_COMMON_CUDA_ARCHS=ON -DBUILD_LIBREALSENSE=ON -DBUILD_AZURE_KINECT=ON -DCMAKE_INSTALL_PREFIX="<open3d_install_directory>" ``` ### Error message ```shell D:\Program Files\Open3D\cpp\open3d/core/Tensor.h(1297): error C2220: 警告被视为错误 - 没有生成“object”文件 (编译源文件 D:\Program Files\Open3D\cpp\open3d\visualization\gui\UIImage.cpp) [D:\Program Files\Open3D\build\cpp\open3d\visualization\gui\GUI.vcxproj] D:\Program Files\Open3D\cpp\open3d/core/Tensor.h(1297): warning C4522: “open3d::core::Tensor”: 指定了多个赋值运算符 (编译源文件 D:\Program Files\Open3D\cpp\open3d\visualization\gui\UIImage.cpp) [D:\Program Files\Open3D\build\cpp\open3d\visualization\gui\GUI.vcxproj] D:\Program Files\Open3D\cpp\open3d/core/Tensor.h(1297): error C2220: 警告被视为错误 - 没有生成“object”文件 (编译源文件 D:\Program Files\Open3D\cpp\open3d\visualization\gui\PickPointsInteractor.cpp) [D:\Program Files\Open3D\build\cpp\open3d\visualization\gui\GUI.vcxproj] and many following same erros: D:\Program Files\Open3D\cpp\open3d/core/Tensor.h(1297): error C2220: 警告被视为错误 - 没有生成“object”文件 [D:\Program Files\Open3D\build\cpp\open3d\core\core.vcxproj] ``` ### Open3D, Python and System information ```markdown - Operating system: Windows 10 64-bit - Python version: Python 3.8` - Open3D version: 15.9.21+g9802d43bc3 - System architecture: x86 - Is this a remote workstation?: no - How did you install Open3D?: build from source - Compiler version (if built from source): vs2017 NVIDIA GPU Computing Toolkit/CUDA/v11.0/include (found version "11.0.221") ``` ### Additional information _No response_
1.0
Summarize the issue and your environments (e.g., "Cannot compile on Ubuntu 20.04 with OpenBLAS") - ### Checklist - [X] I have searched for [similar issues](https://github.com/isl-org/Open3D/issues). - [X] For Python issues, I have tested with the [latest development wheel](http://www.open3d.org/docs/latest/getting_started.html#development-version-pip). - [X] I have checked the [release documentation](http://www.open3d.org/docs/release/) and the [latest documentation](http://www.open3d.org/docs/latest/) (for `master` branch). ### Steps to reproduce the issue #### I first cloned Open3D by: ``` git clone https://github.com/isl-org/Open3D.git cd Open3D ``` #### Then, I build Open3D (on Ubuntu 20.04, with CUDA 11.5) with: ``` mkdir build cd build mkdir build cd build cmake .. -G "Visual Studio 15 2017" -A x64 -DBUILD_CUDA_MODULE=ON -DBUILD_COMMON_CUDA_ARCHS=ON -DBUILD_LIBREALSENSE=ON -DBUILD_AZURE_KINECT=ON -DCMAKE_INSTALL_PREFIX="<open3d_install_directory>" ``` ### Error message ```shell D:\Program Files\Open3D\cpp\open3d/core/Tensor.h(1297): error C2220: 警告被视为错误 - 没有生成“object”文件 (编译源文件 D:\Program Files\Open3D\cpp\open3d\visualization\gui\UIImage.cpp) [D:\Program Files\Open3D\build\cpp\open3d\visualization\gui\GUI.vcxproj] D:\Program Files\Open3D\cpp\open3d/core/Tensor.h(1297): warning C4522: “open3d::core::Tensor”: 指定了多个赋值运算符 (编译源文件 D:\Program Files\Open3D\cpp\open3d\visualization\gui\UIImage.cpp) [D:\Program Files\Open3D\build\cpp\open3d\visualization\gui\GUI.vcxproj] D:\Program Files\Open3D\cpp\open3d/core/Tensor.h(1297): error C2220: 警告被视为错误 - 没有生成“object”文件 (编译源文件 D:\Program Files\Open3D\cpp\open3d\visualization\gui\PickPointsInteractor.cpp) [D:\Program Files\Open3D\build\cpp\open3d\visualization\gui\GUI.vcxproj] and many following same erros: D:\Program Files\Open3D\cpp\open3d/core/Tensor.h(1297): error C2220: 警告被视为错误 - 没有生成“object”文件 [D:\Program Files\Open3D\build\cpp\open3d\core\core.vcxproj] ``` ### Open3D, Python and System information ```markdown - Operating system: Windows 10 64-bit - Python version: Python 3.8` - Open3D version: 15.9.21+g9802d43bc3 - System architecture: x86 - Is this a remote workstation?: no - How did you install Open3D?: build from source - Compiler version (if built from source): vs2017 NVIDIA GPU Computing Toolkit/CUDA/v11.0/include (found version "11.0.221") ``` ### Additional information _No response_
build
summarize the issue and your environments e g cannot compile on ubuntu with openblas checklist i have searched for for python issues i have tested with the i have checked the and the for master branch steps to reproduce the issue i first cloned by git clone cd then i build on ubuntu with cuda with mkdir build cd build mkdir build cd build cmake g visual studio a dbuild cuda module on dbuild common cuda archs on dbuild librealsense on dbuild azure kinect on dcmake install prefix error message shell d program files cpp core tensor h error 警告被视为错误 没有生成“object”文件 编译源文件 d program files cpp visualization gui uiimage cpp d program files cpp core tensor h warning “ core tensor” 指定了多个赋值运算符 编译源文件 d program files cpp visualization gui uiimage cpp d program files cpp core tensor h error 警告被视为错误 没有生成“object”文件 编译源文件 d program files cpp visualization gui pickpointsinteractor cpp and many following same erros d program files cpp core tensor h error 警告被视为错误 没有生成“object”文件 python and system information markdown operating system windows bit python version python version system architecture is this a remote workstation no how did you install build from source compiler version if built from source nvidia gpu computing toolkit cuda include found version additional information no response
1
423,741
28,931,685,301
IssuesEvent
2023-05-09 00:16:03
OpenINF/wg-a-team
https://api.github.com/repos/OpenINF/wg-a-team
opened
meta: the new Product Integrity WG meta-team
documentation help wanted question
Our readme states something about a _new Product Integrity WG meta-team_ @jorgebucaran &amp; @smorimoto would probably know more about this as QA &amp; PM, so just opening an issue about it for us to see what needs to be done here.
1.0
meta: the new Product Integrity WG meta-team - Our readme states something about a _new Product Integrity WG meta-team_ @jorgebucaran &amp; @smorimoto would probably know more about this as QA &amp; PM, so just opening an issue about it for us to see what needs to be done here.
non_build
meta the new product integrity wg meta team our readme states something about a new product integrity wg meta team jorgebucaran amp smorimoto would probably know more about this as qa amp pm so just opening an issue about it for us to see what needs to be done here
0
77,972
22,060,897,136
IssuesEvent
2022-05-30 17:39:58
rust-lang/rust
https://api.github.com/repos/rust-lang/rust
closed
Add back `help_on_error` for download-ci-llvm
E-easy E-mentor A-rustbuild
I accidentally missed this when porting it to rustbuild: https://github.com/rust-lang/rust/pull/95170/files#diff-9e25a089f077eae8c8cefe9586cc07498e64cd950be4c954cb90e4245e5f9fc3L634-L651 It would be nice to add it back. Mentoring instructions: Add a new `help_on_error` parameter to `Builder::download_component_helper` and pass it down from `download_ci_llvm`: https://github.com/rust-lang/rust/blob/00bb4df1fbfdcc9249f8fd5e009b0e18d30c96bf/src/bootstrap/builder.rs#L872 https://github.com/rust-lang/rust/blob/00bb4df1fbfdcc9249f8fd5e009b0e18d30c96bf/src/bootstrap/native.rs#L182 This code moved around in https://github.com/rust-lang/rust/pull/96687 so you may want to base your changes off that PR until it lands. @rustbot label +E-easy +E-mentor +A-rustbuild <!-- TRIAGEBOT_START --> <!-- TRIAGEBOT_ASSIGN_START --> <!-- TRIAGEBOT_ASSIGN_DATA_START$${"user":"binggh"}$$TRIAGEBOT_ASSIGN_DATA_END --> <!-- TRIAGEBOT_ASSIGN_END --> <!-- TRIAGEBOT_END -->
1.0
Add back `help_on_error` for download-ci-llvm - I accidentally missed this when porting it to rustbuild: https://github.com/rust-lang/rust/pull/95170/files#diff-9e25a089f077eae8c8cefe9586cc07498e64cd950be4c954cb90e4245e5f9fc3L634-L651 It would be nice to add it back. Mentoring instructions: Add a new `help_on_error` parameter to `Builder::download_component_helper` and pass it down from `download_ci_llvm`: https://github.com/rust-lang/rust/blob/00bb4df1fbfdcc9249f8fd5e009b0e18d30c96bf/src/bootstrap/builder.rs#L872 https://github.com/rust-lang/rust/blob/00bb4df1fbfdcc9249f8fd5e009b0e18d30c96bf/src/bootstrap/native.rs#L182 This code moved around in https://github.com/rust-lang/rust/pull/96687 so you may want to base your changes off that PR until it lands. @rustbot label +E-easy +E-mentor +A-rustbuild <!-- TRIAGEBOT_START --> <!-- TRIAGEBOT_ASSIGN_START --> <!-- TRIAGEBOT_ASSIGN_DATA_START$${"user":"binggh"}$$TRIAGEBOT_ASSIGN_DATA_END --> <!-- TRIAGEBOT_ASSIGN_END --> <!-- TRIAGEBOT_END -->
build
add back help on error for download ci llvm i accidentally missed this when porting it to rustbuild it would be nice to add it back mentoring instructions add a new help on error parameter to builder download component helper and pass it down from download ci llvm this code moved around in so you may want to base your changes off that pr until it lands rustbot label e easy e mentor a rustbuild
1
176,976
28,308,294,995
IssuesEvent
2023-04-10 13:12:29
bounswe/bounswe2023group5
https://api.github.com/repos/bounswe/bounswe2023group5
closed
Designing Sequence Diagram: Moderator
Priority: Critical Type: Design Status: Done
### Description The sequence diagrams are shared among the team members, I will design the use case diagram of moderator. Moderator checks the reports and evaluates them. Moderators also can delete posts, comments and ban users from the forum. ### 👮‍♀️ Reviewer Bilal Atım ### ⏰ Deadline 10.04.2023
1.0
Designing Sequence Diagram: Moderator - ### Description The sequence diagrams are shared among the team members, I will design the use case diagram of moderator. Moderator checks the reports and evaluates them. Moderators also can delete posts, comments and ban users from the forum. ### 👮‍♀️ Reviewer Bilal Atım ### ⏰ Deadline 10.04.2023
non_build
designing sequence diagram moderator description the sequence diagrams are shared among the team members i will design the use case diagram of moderator moderator checks the reports and evaluates them moderators also can delete posts comments and ban users from the forum 👮‍♀️ reviewer bilal atım ⏰ deadline
0
181,195
14,008,236,570
IssuesEvent
2020-10-28 23:11:56
alteryx/evalml
https://api.github.com/repos/alteryx/evalml
closed
CircleCI: auth needed for Docker Hub as of Nov 1st
task testing
Message which @jeremyliweishih posted in slack today (thanks Jeremy!): > Hi there, > On November 1st, Docker Hub will begin limiting anonymous image pulls. We want to make sure you know how you might be impacted and what you can do to avoid interruptions to your workflow. > Adding Docker authentication to your pipeline config is the easiest way to avoid any service disruptions. If you use the Docker executor or pull Docker images when using the machine executor on CircleCI, we encourage you to authenticate. Because the anonymous API rate limits are based on IP addresses, they will impact CircleCI cloud customers. Authenticated users get higher per-user rate limits, regardless of IP. > We are currently working on a partnership with Docker to minimize the impact of this change for our users and will share more details as we get them. > For more information or to leave a question for us, please head over to Discuss or contact support. > Thanks and happy building, > - The CircleCI team Next step: determine what if any action is needed in order for our CI jobs to continue uninterrupted. @rwedge @gsheni FYI this may affect CI for featuretools/compose/woodwork/others as well
1.0
CircleCI: auth needed for Docker Hub as of Nov 1st - Message which @jeremyliweishih posted in slack today (thanks Jeremy!): > Hi there, > On November 1st, Docker Hub will begin limiting anonymous image pulls. We want to make sure you know how you might be impacted and what you can do to avoid interruptions to your workflow. > Adding Docker authentication to your pipeline config is the easiest way to avoid any service disruptions. If you use the Docker executor or pull Docker images when using the machine executor on CircleCI, we encourage you to authenticate. Because the anonymous API rate limits are based on IP addresses, they will impact CircleCI cloud customers. Authenticated users get higher per-user rate limits, regardless of IP. > We are currently working on a partnership with Docker to minimize the impact of this change for our users and will share more details as we get them. > For more information or to leave a question for us, please head over to Discuss or contact support. > Thanks and happy building, > - The CircleCI team Next step: determine what if any action is needed in order for our CI jobs to continue uninterrupted. @rwedge @gsheni FYI this may affect CI for featuretools/compose/woodwork/others as well
non_build
circleci auth needed for docker hub as of nov message which jeremyliweishih posted in slack today thanks jeremy hi there on november docker hub will begin limiting anonymous image pulls we want to make sure you know how you might be impacted and what you can do to avoid interruptions to your workflow adding docker authentication to your pipeline config is the easiest way to avoid any service disruptions if you use the docker executor or pull docker images when using the machine executor on circleci we encourage you to authenticate because the anonymous api rate limits are based on ip addresses they will impact circleci cloud customers authenticated users get higher per user rate limits regardless of ip we are currently working on a partnership with docker to minimize the impact of this change for our users and will share more details as we get them for more information or to leave a question for us please head over to discuss or contact support thanks and happy building the circleci team next step determine what if any action is needed in order for our ci jobs to continue uninterrupted rwedge gsheni fyi this may affect ci for featuretools compose woodwork others as well
0
45,461
11,672,816,364
IssuesEvent
2020-03-04 07:38:38
Polymer/tools
https://api.github.com/repos/Polymer/tools
closed
Add glob support for fragments
Package: build wontfix
It would be nice to pass fragments into `resolveGlob` as well as `sourceGlobs` are processed: ``` "fragments": "/src/pages/**/*" ```
1.0
Add glob support for fragments - It would be nice to pass fragments into `resolveGlob` as well as `sourceGlobs` are processed: ``` "fragments": "/src/pages/**/*" ```
build
add glob support for fragments it would be nice to pass fragments into resolveglob as well as sourceglobs are processed fragments src pages
1
54,520
13,383,522,510
IssuesEvent
2020-09-02 10:27:56
lkeegan/spatial-model-editor
https://api.github.com/repos/lkeegan/spatial-model-editor
closed
Pixel OpenMP support
CLI build system enhancement
Currently the Pixel simulator supports multithreading using tbb, which is statically linked in the distributed binaries. It has the advantage of being very flexible & has a nice c++ interface. But for parallelisation of simple loops, which is all that we are currently using it for, its implementation seems to have significant overhead costs compared to e.g. openMP. It would be nice to add OpenMP as a compile-time alternative option for users who compile the code themselves, e.g. for running the CLI version on a cluster.
1.0
Pixel OpenMP support - Currently the Pixel simulator supports multithreading using tbb, which is statically linked in the distributed binaries. It has the advantage of being very flexible & has a nice c++ interface. But for parallelisation of simple loops, which is all that we are currently using it for, its implementation seems to have significant overhead costs compared to e.g. openMP. It would be nice to add OpenMP as a compile-time alternative option for users who compile the code themselves, e.g. for running the CLI version on a cluster.
build
pixel openmp support currently the pixel simulator supports multithreading using tbb which is statically linked in the distributed binaries it has the advantage of being very flexible has a nice c interface but for parallelisation of simple loops which is all that we are currently using it for its implementation seems to have significant overhead costs compared to e g openmp it would be nice to add openmp as a compile time alternative option for users who compile the code themselves e g for running the cli version on a cluster
1
11,277
9,301,759,160
IssuesEvent
2019-03-24 01:35:49
dimigoin/dimigoin-front
https://api.github.com/repos/dimigoin/dimigoin-front
closed
Feature request in teacher afterschool management page
feature request service
- 엑셀 파일 다운로드 버튼이 없습니다 (학년별로 달라야합니다.) - ~~세부 관리 페이지가 존재하지 않습니다. 세부관리 페이지는 해당 방과후 신청자 학번 이름을 띄어주는걸 클라이언트가 원하시는듯 싶습니다~~ #54 를 참고해서 구현해야 합니다.
1.0
Feature request in teacher afterschool management page - - 엑셀 파일 다운로드 버튼이 없습니다 (학년별로 달라야합니다.) - ~~세부 관리 페이지가 존재하지 않습니다. 세부관리 페이지는 해당 방과후 신청자 학번 이름을 띄어주는걸 클라이언트가 원하시는듯 싶습니다~~ #54 를 참고해서 구현해야 합니다.
non_build
feature request in teacher afterschool management page 엑셀 파일 다운로드 버튼이 없습니다 학년별로 달라야합니다 세부 관리 페이지가 존재하지 않습니다 세부관리 페이지는 해당 방과후 신청자 학번 이름을 띄어주는걸 클라이언트가 원하시는듯 싶습니다 를 참고해서 구현해야 합니다
0
169,055
26,740,743,190
IssuesEvent
2023-01-30 12:48:35
codestates-seb/seb41_main_028
https://api.github.com/repos/codestates-seb/seb41_main_028
closed
[FE] Feat: certificationModal 데이터 바인딩 작업
FE Feat Design
## To do List <!--해당 Issue를 해결하기 위해 수행해야 할 To do list를 작성해주세요.--> - [x] certification 모달창의 데이터를 받아와서 뿌려주기 ## Deadline <!--작업 완료 기한을 입력해주세요. (e.g. 221219 / 15:00)--> 230129 / 23:59 ## Issue Checklist 아래 체크리스트를 확인해주세요. - [x] Issue Title을 다음 예시와 같이 작성했습니다. (e.g. [FE] Feat: 회원가입 기능 구현 ) - [x] 우측의 Assignees, Labels, Projects, Milestone을 적절하게 선택했습니다.
1.0
[FE] Feat: certificationModal 데이터 바인딩 작업 - ## To do List <!--해당 Issue를 해결하기 위해 수행해야 할 To do list를 작성해주세요.--> - [x] certification 모달창의 데이터를 받아와서 뿌려주기 ## Deadline <!--작업 완료 기한을 입력해주세요. (e.g. 221219 / 15:00)--> 230129 / 23:59 ## Issue Checklist 아래 체크리스트를 확인해주세요. - [x] Issue Title을 다음 예시와 같이 작성했습니다. (e.g. [FE] Feat: 회원가입 기능 구현 ) - [x] 우측의 Assignees, Labels, Projects, Milestone을 적절하게 선택했습니다.
non_build
feat certificationmodal 데이터 바인딩 작업 to do list certification 모달창의 데이터를 받아와서 뿌려주기 deadline issue checklist 아래 체크리스트를 확인해주세요 issue title을 다음 예시와 같이 작성했습니다 e g feat 회원가입 기능 구현 우측의 assignees labels projects milestone을 적절하게 선택했습니다
0
244,526
7,876,155,042
IssuesEvent
2018-06-25 23:21:58
bleehu/Compound_X
https://api.github.com/repos/bleehu/Compound_X
closed
Weapon Proposal: Staff-sling
Feature Request Greenlit - Awaiting Pull Request Priority - Low
It's like a sling, but gigantic. Probably requires very high Dex to use, can load with bullets (which means lead lumps) or grenades. More normal slings also includeable.
1.0
Weapon Proposal: Staff-sling - It's like a sling, but gigantic. Probably requires very high Dex to use, can load with bullets (which means lead lumps) or grenades. More normal slings also includeable.
non_build
weapon proposal staff sling it s like a sling but gigantic probably requires very high dex to use can load with bullets which means lead lumps or grenades more normal slings also includeable
0
5,956
13,391,049,326
IssuesEvent
2020-09-02 21:42:09
pirate/ArchiveBox
https://api.github.com/repos/pirate/ArchiveBox
closed
Bugfix: django.db.utils.IntegrityError: UNIQUE constraint failed: core_snapshot.timestamp
bug changes: architecture complexity: medium status: wip
#### Describe the bug Y'all helped me with upgrading my super old archive to the django branch before official 0.4.9 release. I recently upgraded to the newest version, so I could start adding links. archivebox said I had to re-init. `archivebox init` gives me following error, and will not let me add new links. ``` django.db.utils.IntegrityError: UNIQUE constraint failed: core_snapshot.timestamp ``` Full log/error below. #### Steps to reproduce 1. `git checkout master` to switch from django branch. 2. `git pull origin master` to pull new release. 3. `pip install -e .` (also tried with `pip uninstall archivebox && pip install .`) 4. Navigate to archivebox-output directory. 5. Run `archivebox init`. 6. error. #### Screenshots or log output ``` [i] [2020-07-31 17:34:44] ArchiveBox v0.4.9: archivebox init > /.archivebox-output/archive-working [*] Updating existing ArchiveBox collection in this folder... /.archivebox-output/archive-working ------------------------------------------------------------------ [*] Verifying archive folder structure... √ /.archivebox-output/archive-working/sources √ /.archivebox-output/archive-working/archive √ /.archivebox-output/archive-working/logs √ /.archivebox-output/archive-working/ArchiveBox.conf [*] Verifying main SQL index and running migrations... √ /.archivebox-output/archive-working/index.sqlite3 Operations to perform: Apply all migrations: admin, auth, contenttypes, core, sessions Running migrations: Applying core.0005_auto_20200728_0326... OK [*] Collecting links from any existing indexes and archive folders... √ Loaded 1376 links from existing main index. √ Added 347 orphaned links from existing archive directories. ! Skipped adding 239 invalid link data directories. X /* SNIP A BUNCH OF BROKEN ARCHIVES /* Hint: For more information about the link data directories that were skipped, run: archivebox status archivebox list --status=invalid [*] [2020-07-31 18:01:50] Writing 1723 links to main index... Traceback (most recent call last): File "/home/USERNAME/.local/lib/python3.7/site-packages/django/db/models/query.py", line 575, in update_or_create obj = self.select_for_update().get(**kwargs) File "/home/USERNAME/.local/lib/python3.7/site-packages/django/db/models/query.py", line 417, in get self.model._meta.object_name core.models.DoesNotExist: Snapshot matching query does not exist. During handling of the above exception, another exception occurred: Traceback (most recent call last): File "/home/USERNAME/.local/lib/python3.7/site-packages/django/db/backends/utils.py", line 86, in _execute return self.cursor.execute(sql, params) File "/home/USERNAME/.local/lib/python3.7/site-packages/django/db/backends/sqlite3/base.py", line 396, in execute return Database.Cursor.execute(self, query, params) sqlite3.IntegrityError: UNIQUE constraint failed: core_snapshot.timestamp The above exception was the direct cause of the following exception: Traceback (most recent call last): File "/home/USERNAME/.local/bin/archivebox", line 33, in <module> sys.exit(load_entry_point('archivebox', 'console_scripts', 'archivebox')()) File "/home/USERNAME/datahoard/ArchiveBox/archivebox/cli/__init__.py", line 126, in main pwd=pwd or OUTPUT_DIR, File "/home/USERNAME/datahoard/ArchiveBox/archivebox/cli/__init__.py", line 62, in run_subcommand module.main(args=subcommand_args, stdin=stdin, pwd=pwd) # type: ignore File "/home/USERNAME/datahoard/ArchiveBox/archivebox/cli/archivebox_init.py", line 35, in main out_dir=pwd or OUTPUT_DIR, File "/home/USERNAME/datahoard/ArchiveBox/archivebox/util.py", line 109, in typechecked_function return func(*args, **kwargs) File "/home/USERNAME/datahoard/ArchiveBox/archivebox/main.py", line 369, in init write_main_index(list(all_links.values()), out_dir=out_dir) File "/home/USERNAME/datahoard/ArchiveBox/archivebox/util.py", line 109, in typechecked_function return func(*args, **kwargs) File "/home/USERNAME/datahoard/ArchiveBox/archivebox/index/__init__.py", line 235, in write_main_index write_sql_main_index(links, out_dir=out_dir) File "/home/USERNAME/datahoard/ArchiveBox/archivebox/util.py", line 109, in typechecked_function return func(*args, **kwargs) File "/home/USERNAME/datahoard/ArchiveBox/archivebox/index/sql.py", line 42, in write_sql_main_index Snapshot.objects.update_or_create(url=link.url, defaults=info) File "/home/USERNAME/.local/lib/python3.7/site-packages/django/db/models/manager.py", line 82, in manager_method return getattr(self.get_queryset(), name)(*args, **kwargs) File "/home/USERNAME/.local/lib/python3.7/site-packages/django/db/models/query.py", line 580, in update_or_create obj, created = self._create_object_from_params(kwargs, params, lock=True) File "/home/USERNAME/.local/lib/python3.7/site-packages/django/db/models/query.py", line 604, in _create_object_from_params raise e File "/home/USERNAME/.local/lib/python3.7/site-packages/django/db/models/query.py", line 596, in _create_object_from_params obj = self.create(**params) File "/home/USERNAME/.local/lib/python3.7/site-packages/django/db/models/query.py", line 433, in create obj.save(force_insert=True, using=self.db) File "/home/USERNAME/.local/lib/python3.7/site-packages/django/db/models/base.py", line 746, in save force_update=force_update, update_fields=update_fields) File "/home/USERNAME/.local/lib/python3.7/site-packages/django/db/models/base.py", line 784, in save_base force_update, using, update_fields, File "/home/USERNAME/.local/lib/python3.7/site-packages/django/db/models/base.py", line 887, in _save_table results = self._do_insert(cls._base_manager, using, fields, returning_fields, raw) File "/home/USERNAME/.local/lib/python3.7/site-packages/django/db/models/base.py", line 926, in _do_insert using=using, raw=raw, File "/home/USERNAME/.local/lib/python3.7/site-packages/django/db/models/manager.py", line 82, in manager_method return getattr(self.get_queryset(), name)(*args, **kwargs) File "/home/USERNAME/.local/lib/python3.7/site-packages/django/db/models/query.py", line 1204, in _insert return query.get_compiler(using=using).execute_sql(returning_fields) File "/home/USERNAME/.local/lib/python3.7/site-packages/django/db/models/sql/compiler.py", line 1392, in execute_sql cursor.execute(sql, params) File "/home/USERNAME/.local/lib/python3.7/site-packages/django/db/backends/utils.py", line 68, in execute return self._execute_with_wrappers(sql, params, many=False, executor=self._execute) File "/home/USERNAME/.local/lib/python3.7/site-packages/django/db/backends/utils.py", line 77, in _execute_with_wrappers return executor(sql, params, many, context) File "/home/USERNAME/.local/lib/python3.7/site-packages/django/db/backends/utils.py", line 86, in _execute return self.cursor.execute(sql, params) File "/home/USERNAME/.local/lib/python3.7/site-packages/django/db/utils.py", line 90, in __exit__ raise dj_exc_value.with_traceback(traceback) from exc_value File "/home/USERNAME/.local/lib/python3.7/site-packages/django/db/backends/utils.py", line 86, in _execute return self.cursor.execute(sql, params) File "/home/USERNAME/.local/lib/python3.7/site-packages/django/db/backends/sqlite3/base.py", line 396, in execute return Database.Cursor.execute(self, query, params) django.db.utils.IntegrityError: UNIQUE constraint failed: core_snapshot.timestamp ``` #### Software versions - OS: Ubuntu 18.04 - ArchiveBox version: 0.4.9 (0ac4e12) - Python version: Python 3.7.8
1.0
Bugfix: django.db.utils.IntegrityError: UNIQUE constraint failed: core_snapshot.timestamp - #### Describe the bug Y'all helped me with upgrading my super old archive to the django branch before official 0.4.9 release. I recently upgraded to the newest version, so I could start adding links. archivebox said I had to re-init. `archivebox init` gives me following error, and will not let me add new links. ``` django.db.utils.IntegrityError: UNIQUE constraint failed: core_snapshot.timestamp ``` Full log/error below. #### Steps to reproduce 1. `git checkout master` to switch from django branch. 2. `git pull origin master` to pull new release. 3. `pip install -e .` (also tried with `pip uninstall archivebox && pip install .`) 4. Navigate to archivebox-output directory. 5. Run `archivebox init`. 6. error. #### Screenshots or log output ``` [i] [2020-07-31 17:34:44] ArchiveBox v0.4.9: archivebox init > /.archivebox-output/archive-working [*] Updating existing ArchiveBox collection in this folder... /.archivebox-output/archive-working ------------------------------------------------------------------ [*] Verifying archive folder structure... √ /.archivebox-output/archive-working/sources √ /.archivebox-output/archive-working/archive √ /.archivebox-output/archive-working/logs √ /.archivebox-output/archive-working/ArchiveBox.conf [*] Verifying main SQL index and running migrations... √ /.archivebox-output/archive-working/index.sqlite3 Operations to perform: Apply all migrations: admin, auth, contenttypes, core, sessions Running migrations: Applying core.0005_auto_20200728_0326... OK [*] Collecting links from any existing indexes and archive folders... √ Loaded 1376 links from existing main index. √ Added 347 orphaned links from existing archive directories. ! Skipped adding 239 invalid link data directories. X /* SNIP A BUNCH OF BROKEN ARCHIVES /* Hint: For more information about the link data directories that were skipped, run: archivebox status archivebox list --status=invalid [*] [2020-07-31 18:01:50] Writing 1723 links to main index... Traceback (most recent call last): File "/home/USERNAME/.local/lib/python3.7/site-packages/django/db/models/query.py", line 575, in update_or_create obj = self.select_for_update().get(**kwargs) File "/home/USERNAME/.local/lib/python3.7/site-packages/django/db/models/query.py", line 417, in get self.model._meta.object_name core.models.DoesNotExist: Snapshot matching query does not exist. During handling of the above exception, another exception occurred: Traceback (most recent call last): File "/home/USERNAME/.local/lib/python3.7/site-packages/django/db/backends/utils.py", line 86, in _execute return self.cursor.execute(sql, params) File "/home/USERNAME/.local/lib/python3.7/site-packages/django/db/backends/sqlite3/base.py", line 396, in execute return Database.Cursor.execute(self, query, params) sqlite3.IntegrityError: UNIQUE constraint failed: core_snapshot.timestamp The above exception was the direct cause of the following exception: Traceback (most recent call last): File "/home/USERNAME/.local/bin/archivebox", line 33, in <module> sys.exit(load_entry_point('archivebox', 'console_scripts', 'archivebox')()) File "/home/USERNAME/datahoard/ArchiveBox/archivebox/cli/__init__.py", line 126, in main pwd=pwd or OUTPUT_DIR, File "/home/USERNAME/datahoard/ArchiveBox/archivebox/cli/__init__.py", line 62, in run_subcommand module.main(args=subcommand_args, stdin=stdin, pwd=pwd) # type: ignore File "/home/USERNAME/datahoard/ArchiveBox/archivebox/cli/archivebox_init.py", line 35, in main out_dir=pwd or OUTPUT_DIR, File "/home/USERNAME/datahoard/ArchiveBox/archivebox/util.py", line 109, in typechecked_function return func(*args, **kwargs) File "/home/USERNAME/datahoard/ArchiveBox/archivebox/main.py", line 369, in init write_main_index(list(all_links.values()), out_dir=out_dir) File "/home/USERNAME/datahoard/ArchiveBox/archivebox/util.py", line 109, in typechecked_function return func(*args, **kwargs) File "/home/USERNAME/datahoard/ArchiveBox/archivebox/index/__init__.py", line 235, in write_main_index write_sql_main_index(links, out_dir=out_dir) File "/home/USERNAME/datahoard/ArchiveBox/archivebox/util.py", line 109, in typechecked_function return func(*args, **kwargs) File "/home/USERNAME/datahoard/ArchiveBox/archivebox/index/sql.py", line 42, in write_sql_main_index Snapshot.objects.update_or_create(url=link.url, defaults=info) File "/home/USERNAME/.local/lib/python3.7/site-packages/django/db/models/manager.py", line 82, in manager_method return getattr(self.get_queryset(), name)(*args, **kwargs) File "/home/USERNAME/.local/lib/python3.7/site-packages/django/db/models/query.py", line 580, in update_or_create obj, created = self._create_object_from_params(kwargs, params, lock=True) File "/home/USERNAME/.local/lib/python3.7/site-packages/django/db/models/query.py", line 604, in _create_object_from_params raise e File "/home/USERNAME/.local/lib/python3.7/site-packages/django/db/models/query.py", line 596, in _create_object_from_params obj = self.create(**params) File "/home/USERNAME/.local/lib/python3.7/site-packages/django/db/models/query.py", line 433, in create obj.save(force_insert=True, using=self.db) File "/home/USERNAME/.local/lib/python3.7/site-packages/django/db/models/base.py", line 746, in save force_update=force_update, update_fields=update_fields) File "/home/USERNAME/.local/lib/python3.7/site-packages/django/db/models/base.py", line 784, in save_base force_update, using, update_fields, File "/home/USERNAME/.local/lib/python3.7/site-packages/django/db/models/base.py", line 887, in _save_table results = self._do_insert(cls._base_manager, using, fields, returning_fields, raw) File "/home/USERNAME/.local/lib/python3.7/site-packages/django/db/models/base.py", line 926, in _do_insert using=using, raw=raw, File "/home/USERNAME/.local/lib/python3.7/site-packages/django/db/models/manager.py", line 82, in manager_method return getattr(self.get_queryset(), name)(*args, **kwargs) File "/home/USERNAME/.local/lib/python3.7/site-packages/django/db/models/query.py", line 1204, in _insert return query.get_compiler(using=using).execute_sql(returning_fields) File "/home/USERNAME/.local/lib/python3.7/site-packages/django/db/models/sql/compiler.py", line 1392, in execute_sql cursor.execute(sql, params) File "/home/USERNAME/.local/lib/python3.7/site-packages/django/db/backends/utils.py", line 68, in execute return self._execute_with_wrappers(sql, params, many=False, executor=self._execute) File "/home/USERNAME/.local/lib/python3.7/site-packages/django/db/backends/utils.py", line 77, in _execute_with_wrappers return executor(sql, params, many, context) File "/home/USERNAME/.local/lib/python3.7/site-packages/django/db/backends/utils.py", line 86, in _execute return self.cursor.execute(sql, params) File "/home/USERNAME/.local/lib/python3.7/site-packages/django/db/utils.py", line 90, in __exit__ raise dj_exc_value.with_traceback(traceback) from exc_value File "/home/USERNAME/.local/lib/python3.7/site-packages/django/db/backends/utils.py", line 86, in _execute return self.cursor.execute(sql, params) File "/home/USERNAME/.local/lib/python3.7/site-packages/django/db/backends/sqlite3/base.py", line 396, in execute return Database.Cursor.execute(self, query, params) django.db.utils.IntegrityError: UNIQUE constraint failed: core_snapshot.timestamp ``` #### Software versions - OS: Ubuntu 18.04 - ArchiveBox version: 0.4.9 (0ac4e12) - Python version: Python 3.7.8
non_build
bugfix django db utils integrityerror unique constraint failed core snapshot timestamp describe the bug y all helped me with upgrading my super old archive to the django branch before official release i recently upgraded to the newest version so i could start adding links archivebox said i had to re init archivebox init gives me following error and will not let me add new links django db utils integrityerror unique constraint failed core snapshot timestamp full log error below steps to reproduce git checkout master to switch from django branch git pull origin master to pull new release pip install e also tried with pip uninstall archivebox pip install navigate to archivebox output directory run archivebox init error screenshots or log output archivebox archivebox init archivebox output archive working updating existing archivebox collection in this folder archivebox output archive working verifying archive folder structure √ archivebox output archive working sources √ archivebox output archive working archive √ archivebox output archive working logs √ archivebox output archive working archivebox conf verifying main sql index and running migrations √ archivebox output archive working index operations to perform apply all migrations admin auth contenttypes core sessions running migrations applying core auto ok collecting links from any existing indexes and archive folders √ loaded links from existing main index √ added orphaned links from existing archive directories skipped adding invalid link data directories x snip a bunch of broken archives hint for more information about the link data directories that were skipped run archivebox status archivebox list status invalid writing links to main index traceback most recent call last file home username local lib site packages django db models query py line in update or create obj self select for update get kwargs file home username local lib site packages django db models query py line in get self model meta object name core models doesnotexist snapshot matching query does not exist during handling of the above exception another exception occurred traceback most recent call last file home username local lib site packages django db backends utils py line in execute return self cursor execute sql params file home username local lib site packages django db backends base py line in execute return database cursor execute self query params integrityerror unique constraint failed core snapshot timestamp the above exception was the direct cause of the following exception traceback most recent call last file home username local bin archivebox line in sys exit load entry point archivebox console scripts archivebox file home username datahoard archivebox archivebox cli init py line in main pwd pwd or output dir file home username datahoard archivebox archivebox cli init py line in run subcommand module main args subcommand args stdin stdin pwd pwd type ignore file home username datahoard archivebox archivebox cli archivebox init py line in main out dir pwd or output dir file home username datahoard archivebox archivebox util py line in typechecked function return func args kwargs file home username datahoard archivebox archivebox main py line in init write main index list all links values out dir out dir file home username datahoard archivebox archivebox util py line in typechecked function return func args kwargs file home username datahoard archivebox archivebox index init py line in write main index write sql main index links out dir out dir file home username datahoard archivebox archivebox util py line in typechecked function return func args kwargs file home username datahoard archivebox archivebox index sql py line in write sql main index snapshot objects update or create url link url defaults info file home username local lib site packages django db models manager py line in manager method return getattr self get queryset name args kwargs file home username local lib site packages django db models query py line in update or create obj created self create object from params kwargs params lock true file home username local lib site packages django db models query py line in create object from params raise e file home username local lib site packages django db models query py line in create object from params obj self create params file home username local lib site packages django db models query py line in create obj save force insert true using self db file home username local lib site packages django db models base py line in save force update force update update fields update fields file home username local lib site packages django db models base py line in save base force update using update fields file home username local lib site packages django db models base py line in save table results self do insert cls base manager using fields returning fields raw file home username local lib site packages django db models base py line in do insert using using raw raw file home username local lib site packages django db models manager py line in manager method return getattr self get queryset name args kwargs file home username local lib site packages django db models query py line in insert return query get compiler using using execute sql returning fields file home username local lib site packages django db models sql compiler py line in execute sql cursor execute sql params file home username local lib site packages django db backends utils py line in execute return self execute with wrappers sql params many false executor self execute file home username local lib site packages django db backends utils py line in execute with wrappers return executor sql params many context file home username local lib site packages django db backends utils py line in execute return self cursor execute sql params file home username local lib site packages django db utils py line in exit raise dj exc value with traceback traceback from exc value file home username local lib site packages django db backends utils py line in execute return self cursor execute sql params file home username local lib site packages django db backends base py line in execute return database cursor execute self query params django db utils integrityerror unique constraint failed core snapshot timestamp software versions os ubuntu archivebox version python version python
0
80,798
15,589,001,517
IssuesEvent
2021-03-18 07:23:03
soumya132/pomscan
https://api.github.com/repos/soumya132/pomscan
closed
CVE-2017-5664 (High) detected in tomcat-embed-core-8.5.4.jar
security vulnerability
## CVE-2017-5664 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tomcat-embed-core-8.5.4.jar</b></p></summary> <p>Core Tomcat implementation</p> <p>Library home page: <a href="http://tomcat.apache.org/">http://tomcat.apache.org/</a></p> <p>Path to dependency file: pomscan/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/apache/tomcat/embed/tomcat-embed-core/8.5.4/tomcat-embed-core-8.5.4.jar</p> <p> Dependency Hierarchy: - spring-boot-starter-jersey-1.4.0.RELEASE.jar (Root Library) - spring-boot-starter-tomcat-1.4.0.RELEASE.jar - :x: **tomcat-embed-core-8.5.4.jar** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The error page mechanism of the Java Servlet Specification requires that, when an error occurs and an error page is configured for the error that occurred, the original request and response are forwarded to the error page. This means that the request is presented to the error page with the original HTTP method. If the error page is a static file, expected behaviour is to serve content of the file as if processing a GET request, regardless of the actual HTTP method. The Default Servlet in Apache Tomcat 9.0.0.M1 to 9.0.0.M20, 8.5.0 to 8.5.14, 8.0.0.RC1 to 8.0.43 and 7.0.0 to 7.0.77 did not do this. Depending on the original request this could lead to unexpected and undesirable results for static error pages including, if the DefaultServlet is configured to permit writes, the replacement or removal of the custom error page. Notes for other user provided error pages: (1) Unless explicitly coded otherwise, JSPs ignore the HTTP method. JSPs used as error pages must must ensure that they handle any error dispatch as a GET request, regardless of the actual method. (2) By default, the response generated by a Servlet does depend on the HTTP method. Custom Servlets used as error pages must ensure that they handle any error dispatch as a GET request, regardless of the actual method. <p>Publish Date: 2017-06-06 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-5664>CVE-2017-5664</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: High - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2017-5664">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2017-5664</a></p> <p>Release Date: 2017-06-06</p> <p>Fix Resolution: org.apache.tomcat.embed:tomcat-embed-core:9.0.0.M21,8.5.15,8.0.44,7.0.78,org.apache.tomcat:tomcat-catalina:9.0.0.M21,8.5.15,8.0.44,7.0.78</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2017-5664 (High) detected in tomcat-embed-core-8.5.4.jar - ## CVE-2017-5664 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tomcat-embed-core-8.5.4.jar</b></p></summary> <p>Core Tomcat implementation</p> <p>Library home page: <a href="http://tomcat.apache.org/">http://tomcat.apache.org/</a></p> <p>Path to dependency file: pomscan/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/apache/tomcat/embed/tomcat-embed-core/8.5.4/tomcat-embed-core-8.5.4.jar</p> <p> Dependency Hierarchy: - spring-boot-starter-jersey-1.4.0.RELEASE.jar (Root Library) - spring-boot-starter-tomcat-1.4.0.RELEASE.jar - :x: **tomcat-embed-core-8.5.4.jar** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The error page mechanism of the Java Servlet Specification requires that, when an error occurs and an error page is configured for the error that occurred, the original request and response are forwarded to the error page. This means that the request is presented to the error page with the original HTTP method. If the error page is a static file, expected behaviour is to serve content of the file as if processing a GET request, regardless of the actual HTTP method. The Default Servlet in Apache Tomcat 9.0.0.M1 to 9.0.0.M20, 8.5.0 to 8.5.14, 8.0.0.RC1 to 8.0.43 and 7.0.0 to 7.0.77 did not do this. Depending on the original request this could lead to unexpected and undesirable results for static error pages including, if the DefaultServlet is configured to permit writes, the replacement or removal of the custom error page. Notes for other user provided error pages: (1) Unless explicitly coded otherwise, JSPs ignore the HTTP method. JSPs used as error pages must must ensure that they handle any error dispatch as a GET request, regardless of the actual method. (2) By default, the response generated by a Servlet does depend on the HTTP method. Custom Servlets used as error pages must ensure that they handle any error dispatch as a GET request, regardless of the actual method. <p>Publish Date: 2017-06-06 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-5664>CVE-2017-5664</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: High - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2017-5664">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2017-5664</a></p> <p>Release Date: 2017-06-06</p> <p>Fix Resolution: org.apache.tomcat.embed:tomcat-embed-core:9.0.0.M21,8.5.15,8.0.44,7.0.78,org.apache.tomcat:tomcat-catalina:9.0.0.M21,8.5.15,8.0.44,7.0.78</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_build
cve high detected in tomcat embed core jar cve high severity vulnerability vulnerable library tomcat embed core jar core tomcat implementation library home page a href path to dependency file pomscan pom xml path to vulnerable library home wss scanner repository org apache tomcat embed tomcat embed core tomcat embed core jar dependency hierarchy spring boot starter jersey release jar root library spring boot starter tomcat release jar x tomcat embed core jar vulnerable library found in base branch master vulnerability details the error page mechanism of the java servlet specification requires that when an error occurs and an error page is configured for the error that occurred the original request and response are forwarded to the error page this means that the request is presented to the error page with the original http method if the error page is a static file expected behaviour is to serve content of the file as if processing a get request regardless of the actual http method the default servlet in apache tomcat to to to and to did not do this depending on the original request this could lead to unexpected and undesirable results for static error pages including if the defaultservlet is configured to permit writes the replacement or removal of the custom error page notes for other user provided error pages unless explicitly coded otherwise jsps ignore the http method jsps used as error pages must must ensure that they handle any error dispatch as a get request regardless of the actual method by default the response generated by a servlet does depend on the http method custom servlets used as error pages must ensure that they handle any error dispatch as a get request regardless of the actual method publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact high availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org apache tomcat embed tomcat embed core org apache tomcat tomcat catalina step up your open source security game with whitesource
0
End of preview. Expand in Data Studio

Dataset Card for "binary-10IQR-build"

More Information needed

Downloads last month
18

Collection including karths/binary-10IQR-build