Technical Debt and it Types Datasets
Collection
24 items
•
Updated
•
1
Unnamed: 0
int64 0
832k
| id
float64 2.49B
32.1B
| type
stringclasses 1
value | created_at
stringlengths 19
19
| repo
stringlengths 7
112
| repo_url
stringlengths 36
141
| action
stringclasses 3
values | title
stringlengths 1
900
| labels
stringlengths 4
522
| body
stringlengths 5
218k
| index
stringclasses 6
values | text_combine
stringlengths 96
219k
| label
stringclasses 2
values | text
stringlengths 96
102k
| binary_label
int64 0
1
|
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
318,683 | 23,734,541,985 | IssuesEvent | 2022-08-31 06:50:08 | codeing999/CLIPs-backend | https://api.github.com/repos/codeing999/CLIPs-backend | closed | git 협업 방법 변경 | documentation | 기존 브랜치
main : 최종적으로 배포할 브랜치
submain : 각자의 작업 브랜치에서 작업하다가 에러 없이 작업 완료될 때마다 수시로 푸쉬할 브랜치
dev/닉네임 : 각자의 작업 브랜치.
변경된 브랜치
main : 배포할 브랜치
develop : 개발 단계에서 각각의 기능이 완성될 때 마다 합칠 브랜치
feature/기능명, fix/에러 : 각 기능 단위로 브랜치를 새로 만들고 완성 시마다 develop에 합치고 이 브랜치는 삭제.
이 내용은 동의 하시면 docs에도 반영하여 수정하도록 하겠습니다. | 1.0 | git 협업 방법 변경 - 기존 브랜치
main : 최종적으로 배포할 브랜치
submain : 각자의 작업 브랜치에서 작업하다가 에러 없이 작업 완료될 때마다 수시로 푸쉬할 브랜치
dev/닉네임 : 각자의 작업 브랜치.
변경된 브랜치
main : 배포할 브랜치
develop : 개발 단계에서 각각의 기능이 완성될 때 마다 합칠 브랜치
feature/기능명, fix/에러 : 각 기능 단위로 브랜치를 새로 만들고 완성 시마다 develop에 합치고 이 브랜치는 삭제.
이 내용은 동의 하시면 docs에도 반영하여 수정하도록 하겠습니다. | non_architecture | git 협업 방법 변경 기존 브랜치 main 최종적으로 배포할 브랜치 submain 각자의 작업 브랜치에서 작업하다가 에러 없이 작업 완료될 때마다 수시로 푸쉬할 브랜치 dev 닉네임 각자의 작업 브랜치 변경된 브랜치 main 배포할 브랜치 develop 개발 단계에서 각각의 기능이 완성될 때 마다 합칠 브랜치 feature 기능명 fix 에러 각 기능 단위로 브랜치를 새로 만들고 완성 시마다 develop에 합치고 이 브랜치는 삭제 이 내용은 동의 하시면 docs에도 반영하여 수정하도록 하겠습니다 | 0 |
11,270 | 14,060,158,658 | IssuesEvent | 2020-11-03 05:16:59 | gfx-rs/naga | https://api.github.com/repos/gfx-rs/naga | opened | Typifier -> Classifier | area: processing help wanted kind: feature kind: question | Currently, we have the typifier module that assigns expression types. We need more information induced from the expressions though. One such bit is what is the variability of an expression: global, uniform, or local.
We could go at least two ways from here:
1. Introduce an expression visitor, which will be used by the typifier as well as other things, potentially in user space as well.
2. Rename the typifier to "classifier" and make it derive the variability together with the type.
The (2) seems more straightforward to me, although I do wonder about a way to let users easily process our IR modules. Maybe it's just too early for that, and we'll need to pick the easier solution. | 1.0 | Typifier -> Classifier - Currently, we have the typifier module that assigns expression types. We need more information induced from the expressions though. One such bit is what is the variability of an expression: global, uniform, or local.
We could go at least two ways from here:
1. Introduce an expression visitor, which will be used by the typifier as well as other things, potentially in user space as well.
2. Rename the typifier to "classifier" and make it derive the variability together with the type.
The (2) seems more straightforward to me, although I do wonder about a way to let users easily process our IR modules. Maybe it's just too early for that, and we'll need to pick the easier solution. | non_architecture | typifier classifier currently we have the typifier module that assigns expression types we need more information induced from the expressions though one such bit is what is the variability of an expression global uniform or local we could go at least two ways from here introduce an expression visitor which will be used by the typifier as well as other things potentially in user space as well rename the typifier to classifier and make it derive the variability together with the type the seems more straightforward to me although i do wonder about a way to let users easily process our ir modules maybe it s just too early for that and we ll need to pick the easier solution | 0 |
1,423 | 5,892,275,610 | IssuesEvent | 2017-05-17 19:04:08 | gctools-outilsgc/gcconnex | https://api.github.com/repos/gctools-outilsgc/gcconnex | closed | design prompt for auto suggesting tags / audience on new content creation | enhancement high-level design Information Architecture - Controlled Vocabulary | design mock up for user prompt suggesting tags / audiences for content. | 1.0 | design prompt for auto suggesting tags / audience on new content creation - design mock up for user prompt suggesting tags / audiences for content. | architecture | design prompt for auto suggesting tags audience on new content creation design mock up for user prompt suggesting tags audiences for content | 1 |
6,669 | 15,014,277,621 | IssuesEvent | 2021-02-01 06:18:02 | burespe1/FRAME | https://api.github.com/repos/burespe1/FRAME | closed | Functional View to Physical View | EA Development architecture methodologies automation on hold physical view | following are the steps to be followed to build a physical view.
1)

2)

3)

can we automate this process within EA? | 1.0 | Functional View to Physical View - following are the steps to be followed to build a physical view.
1)

2)

3)

can we automate this process within EA? | architecture | functional view to physical view following are the steps to be followed to build a physical view can we automate this process within ea | 1 |
7,842 | 19,649,865,557 | IssuesEvent | 2022-01-10 04:56:42 | Vector35/binaryninja-api | https://api.github.com/repos/Vector35/binaryninja-api | closed | arm64 `fcmp` doesn't work with many condition codes | enhancement architecture ARM64 Effort: Low Impact: Medium | **Version and Platform (required):**
- Binary Ninja Version: Version 2.5.3140-dev (Build ID 532595b6)
- OS: macOS
- OS Version: 11.6
**Bug Description:**
[This binary](https://github.com/Vector35/binaryninja-api/files/7790320/fcmp.zip) demonstrates using all 14 ARM condition codes with `fcmp`. Among these:
- `eq`, `ne`, `cs`, and `cc` look good.
- `ge`, `lt`, `gt`, and `le` are decompiled using `unimplemented`, e.g.:
```
000000f0 int64_t test_ge(int32_t arg1 @ v0, int32_t arg2 @ v1)
000000f0 int64_t x0 = 0
000000f4 arg1 f- arg2
000000f4 bool v = unimplemented {fcmp s0, s1}
000000f4 bool n = unimplemented {fcmp s0, s1}
000000f8 if (n == v)
00000100 x0 = 1
000000fc return x0
```
This is unfortunate, since those condition codes are very common.
- `ls` and its inverse `hi` are decompiled correctly but suboptimally. For `ls`:
```
000000d8 int64_t test_ls(float arg1 @ v0, float arg2 @ v1)
000000d8 int64_t x0 = 0
000000dc arg1 - arg2
000000e0 if (arg1 == arg2 || arg1 < arg2)
000000e8 x0 = 1
000000e4 return x0
```
This could be `arg1 <= arg2`.
- `pl` is decompiled incorrectly:
```
00000078 int64_t test_pl(int32_t arg1 @ v0, int32_t arg2 @ v1)
00000078 int64_t x0 = 0
00000080 if (arg1 f- arg2 s>= 0)
00000088 x0 = 1
00000084 return x0
```
In reality it should be `not(arg1 < arg2)` (see below table), which is not the same as `arg1 - arg2 >= 0`. For example, if one or both arguments is NaN, `not(arg1 < arg2)` is true, but `arg1 - arg2 >= 0` is false.
- `mi` is similarly oddly decompiled as `arg1 f- arg2 s< 0` when it should be `arg1 < arg2`. I can't think of any cases within standard IEEE floating point where these expressions aren't equivalent, but under IEEE floating point with subnormals disabled (common in games), [there are pairs of floats](https://stackoverflow.com/a/54532647) `a`, `b` such that `a < b` but `a - b == 0.0`.
Here is the relevant table from the ARM manual:
<img width="819" alt="image" src="https://user-images.githubusercontent.com/47517/147696978-bd8616b1-e459-4cbd-bdaf-7db3e19bffda.png">
| 1.0 | arm64 `fcmp` doesn't work with many condition codes - **Version and Platform (required):**
- Binary Ninja Version: Version 2.5.3140-dev (Build ID 532595b6)
- OS: macOS
- OS Version: 11.6
**Bug Description:**
[This binary](https://github.com/Vector35/binaryninja-api/files/7790320/fcmp.zip) demonstrates using all 14 ARM condition codes with `fcmp`. Among these:
- `eq`, `ne`, `cs`, and `cc` look good.
- `ge`, `lt`, `gt`, and `le` are decompiled using `unimplemented`, e.g.:
```
000000f0 int64_t test_ge(int32_t arg1 @ v0, int32_t arg2 @ v1)
000000f0 int64_t x0 = 0
000000f4 arg1 f- arg2
000000f4 bool v = unimplemented {fcmp s0, s1}
000000f4 bool n = unimplemented {fcmp s0, s1}
000000f8 if (n == v)
00000100 x0 = 1
000000fc return x0
```
This is unfortunate, since those condition codes are very common.
- `ls` and its inverse `hi` are decompiled correctly but suboptimally. For `ls`:
```
000000d8 int64_t test_ls(float arg1 @ v0, float arg2 @ v1)
000000d8 int64_t x0 = 0
000000dc arg1 - arg2
000000e0 if (arg1 == arg2 || arg1 < arg2)
000000e8 x0 = 1
000000e4 return x0
```
This could be `arg1 <= arg2`.
- `pl` is decompiled incorrectly:
```
00000078 int64_t test_pl(int32_t arg1 @ v0, int32_t arg2 @ v1)
00000078 int64_t x0 = 0
00000080 if (arg1 f- arg2 s>= 0)
00000088 x0 = 1
00000084 return x0
```
In reality it should be `not(arg1 < arg2)` (see below table), which is not the same as `arg1 - arg2 >= 0`. For example, if one or both arguments is NaN, `not(arg1 < arg2)` is true, but `arg1 - arg2 >= 0` is false.
- `mi` is similarly oddly decompiled as `arg1 f- arg2 s< 0` when it should be `arg1 < arg2`. I can't think of any cases within standard IEEE floating point where these expressions aren't equivalent, but under IEEE floating point with subnormals disabled (common in games), [there are pairs of floats](https://stackoverflow.com/a/54532647) `a`, `b` such that `a < b` but `a - b == 0.0`.
Here is the relevant table from the ARM manual:
<img width="819" alt="image" src="https://user-images.githubusercontent.com/47517/147696978-bd8616b1-e459-4cbd-bdaf-7db3e19bffda.png">
| architecture | fcmp doesn t work with many condition codes version and platform required binary ninja version version dev build id os macos os version bug description demonstrates using all arm condition codes with fcmp among these eq ne cs and cc look good ge lt gt and le are decompiled using unimplemented e g t test ge t t t f bool v unimplemented fcmp bool n unimplemented fcmp if n v return this is unfortunate since those condition codes are very common ls and its inverse hi are decompiled correctly but suboptimally for ls t test ls float float t if return this could be pl is decompiled incorrectly t test pl t t t if f s return in reality it should be not for example if one or both arguments is nan not is false mi is similarly oddly decompiled as f s when it should be i can t think of any cases within standard ieee floating point where these expressions aren t equivalent but under ieee floating point with subnormals disabled common in games a b such that a b but a b here is the relevant table from the arm manual img width alt image src | 1 |
4,840 | 11,757,762,280 | IssuesEvent | 2020-03-13 14:16:01 | kubernetes/kubernetes | https://api.github.com/repos/kubernetes/kubernetes | closed | Investigate Transitive Deps from docker/libnetwork | area/code-organization area/dependency kind/feature lifecycle/rotten sig/architecture sig/network | **What would you like to be added**:
github.com/docker/libnetwork has a lot of transitive deps but we only use its (relatively small) `ipvs` package for kube-proxy. Luckily because it doesn't use go modules yet, updating it doesn't actually update it's transitive deps, but if/when it does, managing it's transitive deps will be a pain because of its overlapping transitive deps to our other dependencies.
From doing a quick search of the Kubernetes repo, we only use the `ipvs` package from github.com/docker/libnetwork. Some options going forward would be:
* fork that repo with only the `ipvs` package
* copy the `ipvs` package to k8s.io/kubernetes
* ask the docker maintainers to put the ipvs package into a separate repo
* ???
Open to other options I haven't considered yet.
**Why is this needed**:
Will significantly improve the maintainability of our dependencies once github.com/docker/libnetwork uses go modules.
| 1.0 | Investigate Transitive Deps from docker/libnetwork - **What would you like to be added**:
github.com/docker/libnetwork has a lot of transitive deps but we only use its (relatively small) `ipvs` package for kube-proxy. Luckily because it doesn't use go modules yet, updating it doesn't actually update it's transitive deps, but if/when it does, managing it's transitive deps will be a pain because of its overlapping transitive deps to our other dependencies.
From doing a quick search of the Kubernetes repo, we only use the `ipvs` package from github.com/docker/libnetwork. Some options going forward would be:
* fork that repo with only the `ipvs` package
* copy the `ipvs` package to k8s.io/kubernetes
* ask the docker maintainers to put the ipvs package into a separate repo
* ???
Open to other options I haven't considered yet.
**Why is this needed**:
Will significantly improve the maintainability of our dependencies once github.com/docker/libnetwork uses go modules.
| architecture | investigate transitive deps from docker libnetwork what would you like to be added github com docker libnetwork has a lot of transitive deps but we only use its relatively small ipvs package for kube proxy luckily because it doesn t use go modules yet updating it doesn t actually update it s transitive deps but if when it does managing it s transitive deps will be a pain because of its overlapping transitive deps to our other dependencies from doing a quick search of the kubernetes repo we only use the ipvs package from github com docker libnetwork some options going forward would be fork that repo with only the ipvs package copy the ipvs package to io kubernetes ask the docker maintainers to put the ipvs package into a separate repo open to other options i haven t considered yet why is this needed will significantly improve the maintainability of our dependencies once github com docker libnetwork uses go modules | 1 |
134,670 | 30,113,591,798 | IssuesEvent | 2023-06-30 09:40:04 | FerretDB/FerretDB | https://api.github.com/repos/FerretDB/FerretDB | opened | Implement `$tsSecond` timestamp expression operator | code/feature not ready area/aggregations | ### What should be done?
It should be supported in all pipeline stages that support raw expressions and other pipelines that allow the `$expr` operator.
* https://www.mongodb.com/docs/manual/reference/operator/aggregation/tsSecond/#mongodb-expression-exp.-tsSecond | 1.0 | Implement `$tsSecond` timestamp expression operator - ### What should be done?
It should be supported in all pipeline stages that support raw expressions and other pipelines that allow the `$expr` operator.
* https://www.mongodb.com/docs/manual/reference/operator/aggregation/tsSecond/#mongodb-expression-exp.-tsSecond | non_architecture | implement tssecond timestamp expression operator what should be done it should be supported in all pipeline stages that support raw expressions and other pipelines that allow the expr operator | 0 |
10,997 | 27,734,774,944 | IssuesEvent | 2023-03-15 10:28:09 | OasisLMF/OasisPlatform | https://api.github.com/repos/OasisLMF/OasisPlatform | opened | Fix Helm customization Readme | bug Documentation scalable architecture | ## Issue Description
> **carlfischerjba:** Helm appears to be combining the default values files from `OasisPlatform/kubernetes/charts/oasis-models/values.yaml` with the file I specify on the command line.
>
> Apparently, the `workers` (a mapping) get merged so we have the default `piwind-demo` as well as the New workers I've defined, but the `modelVolumes` (a sequence) are overwritten so we only have the New volumes and not `piwind-model-data-pv`, this leads to the error.
>
> This means [the methods recommended in the readme](https://github.com/OasisLMF/OasisPlatform/blob/6dd90eb3ced94e48464de158af88417de3b49b9a/kubernetes/charts/README.md?plain=1#L242) don't work. I guess it's not been spotted until now because everyone has kept the PiWind model in place. Trouble starts once you decide you don't need it. Merging values vs maps vs lists is a problem with docker-compose.yml and other types of config files too, including json. It's surprising there's no way to tell Helm to ignore the defaults. Without such an option, the ways to keep everything working are not very satisfactory:
> * edit the values files in place
> * rename or delete the default values files
> * or copy the template as suggested but only add to it, never remove anything
>
> Another alternative would be for you to rename models/values.yaml to models/values_sample.yaml so it's ignored by Helm and update the instructions in the readme accordingly.
>
> The same could occur with the platform and monitoring charts but I guess that's less likely because in general a few values are modified and override the defaults instead of adding extra items to mappings or sequences (are those the correct YAML terms?).
At a minimum the documentation instructions should be updated to note/fix this problem | 1.0 | Fix Helm customization Readme - ## Issue Description
> **carlfischerjba:** Helm appears to be combining the default values files from `OasisPlatform/kubernetes/charts/oasis-models/values.yaml` with the file I specify on the command line.
>
> Apparently, the `workers` (a mapping) get merged so we have the default `piwind-demo` as well as the New workers I've defined, but the `modelVolumes` (a sequence) are overwritten so we only have the New volumes and not `piwind-model-data-pv`, this leads to the error.
>
> This means [the methods recommended in the readme](https://github.com/OasisLMF/OasisPlatform/blob/6dd90eb3ced94e48464de158af88417de3b49b9a/kubernetes/charts/README.md?plain=1#L242) don't work. I guess it's not been spotted until now because everyone has kept the PiWind model in place. Trouble starts once you decide you don't need it. Merging values vs maps vs lists is a problem with docker-compose.yml and other types of config files too, including json. It's surprising there's no way to tell Helm to ignore the defaults. Without such an option, the ways to keep everything working are not very satisfactory:
> * edit the values files in place
> * rename or delete the default values files
> * or copy the template as suggested but only add to it, never remove anything
>
> Another alternative would be for you to rename models/values.yaml to models/values_sample.yaml so it's ignored by Helm and update the instructions in the readme accordingly.
>
> The same could occur with the platform and monitoring charts but I guess that's less likely because in general a few values are modified and override the defaults instead of adding extra items to mappings or sequences (are those the correct YAML terms?).
At a minimum the documentation instructions should be updated to note/fix this problem | architecture | fix helm customization readme issue description carlfischerjba helm appears to be combining the default values files from oasisplatform kubernetes charts oasis models values yaml with the file i specify on the command line apparently the workers a mapping get merged so we have the default piwind demo as well as the new workers i ve defined but the modelvolumes a sequence are overwritten so we only have the new volumes and not piwind model data pv this leads to the error this means don t work i guess it s not been spotted until now because everyone has kept the piwind model in place trouble starts once you decide you don t need it merging values vs maps vs lists is a problem with docker compose yml and other types of config files too including json it s surprising there s no way to tell helm to ignore the defaults without such an option the ways to keep everything working are not very satisfactory edit the values files in place rename or delete the default values files or copy the template as suggested but only add to it never remove anything another alternative would be for you to rename models values yaml to models values sample yaml so it s ignored by helm and update the instructions in the readme accordingly the same could occur with the platform and monitoring charts but i guess that s less likely because in general a few values are modified and override the defaults instead of adding extra items to mappings or sequences are those the correct yaml terms at a minimum the documentation instructions should be updated to note fix this problem | 1 |
10,129 | 26,364,651,544 | IssuesEvent | 2023-01-11 15:43:36 | mehab/DTKafkaPOC | https://api.github.com/repos/mehab/DTKafkaPOC | opened | Add architecture diagrams | documentation 📃 architecture 🔮 | In order for us (and others) to better understand what we're building here, we should have architecture diagrams.
Preferably there should be multiple "resolutions" from a high-level overview to individual services. The topology diagrams we can generate using Kafka Streams will be helpful for the latter. | 1.0 | Add architecture diagrams - In order for us (and others) to better understand what we're building here, we should have architecture diagrams.
Preferably there should be multiple "resolutions" from a high-level overview to individual services. The topology diagrams we can generate using Kafka Streams will be helpful for the latter. | architecture | add architecture diagrams in order for us and others to better understand what we re building here we should have architecture diagrams preferably there should be multiple resolutions from a high level overview to individual services the topology diagrams we can generate using kafka streams will be helpful for the latter | 1 |
445,871 | 12,837,462,604 | IssuesEvent | 2020-07-07 15:49:31 | code-ready/crc | https://api.github.com/repos/code-ready/crc | closed | Add 'Experimental' messages when `podman-env` command is used. | priority/critical status/stale | We need to add output to the `podman-env` command, as for the time being no changes will happen to this functionality
Note: adding an `echo`, so when `podman-env` is `eval`-ed it would still show | 1.0 | Add 'Experimental' messages when `podman-env` command is used. - We need to add output to the `podman-env` command, as for the time being no changes will happen to this functionality
Note: adding an `echo`, so when `podman-env` is `eval`-ed it would still show | non_architecture | add experimental messages when podman env command is used we need to add output to the podman env command as for the time being no changes will happen to this functionality note adding an echo so when podman env is eval ed it would still show | 0 |
4,114 | 10,584,831,316 | IssuesEvent | 2019-10-08 16:11:07 | fga-eps-mds/2019.2-Over26 | https://api.github.com/repos/fga-eps-mds/2019.2-Over26 | closed | Elaborar Plano de Qualidade | Architecture Documentation EPS | ## Descrição da Mudança *
<!--- Forneça um resumo geral da _issue_ -->
Criar a primeira versão do plano de qualidade para o projeto.
## Checklist *
<!-- Essa checklist propõe a criação de uma boa issue -->
<!-- Se a issue é sobre uma história de usuário, seu nome deve ser "USXX - Nome da história-->
<!-- Se a issue é sobre um bug, seu nome deve ser "BF - Nome curto do bug"-->
<!-- Se a issue é sobre outra tarefa o nome deve ser uma simples descrição da tarefa-->
- [x] Esta issue tem um nome significativo.
- [x] O nome da issue está no padrão.
- [x] Esta issue tem uma descrição de fácil entendimento.
- [x] Esta issue tem uma boa definição de critérios de aceitação.
- [x] Esta issue tem labels associadas.
- [ ] Esta issue está associada à uma milestone.
- [ ] Esta issue tem uma pontuação estimada.
## Tarefas *
<!-- Adicione aqui as tarefas necessárias para concluir a issue -->
- [ ] Criar plano de qualidade
## Critérios de Aceitação *
<!-- Liste aqui o conjunto de aspectos mecessários para considerar a atividade como completa-->
<!-- Os itens serão adicionados pelo Product Owner -->
- [ ] A primeira versão do plano de qualidade deve estar elaborada | 1.0 | Elaborar Plano de Qualidade - ## Descrição da Mudança *
<!--- Forneça um resumo geral da _issue_ -->
Criar a primeira versão do plano de qualidade para o projeto.
## Checklist *
<!-- Essa checklist propõe a criação de uma boa issue -->
<!-- Se a issue é sobre uma história de usuário, seu nome deve ser "USXX - Nome da história-->
<!-- Se a issue é sobre um bug, seu nome deve ser "BF - Nome curto do bug"-->
<!-- Se a issue é sobre outra tarefa o nome deve ser uma simples descrição da tarefa-->
- [x] Esta issue tem um nome significativo.
- [x] O nome da issue está no padrão.
- [x] Esta issue tem uma descrição de fácil entendimento.
- [x] Esta issue tem uma boa definição de critérios de aceitação.
- [x] Esta issue tem labels associadas.
- [ ] Esta issue está associada à uma milestone.
- [ ] Esta issue tem uma pontuação estimada.
## Tarefas *
<!-- Adicione aqui as tarefas necessárias para concluir a issue -->
- [ ] Criar plano de qualidade
## Critérios de Aceitação *
<!-- Liste aqui o conjunto de aspectos mecessários para considerar a atividade como completa-->
<!-- Os itens serão adicionados pelo Product Owner -->
- [ ] A primeira versão do plano de qualidade deve estar elaborada | architecture | elaborar plano de qualidade descrição da mudança criar a primeira versão do plano de qualidade para o projeto checklist esta issue tem um nome significativo o nome da issue está no padrão esta issue tem uma descrição de fácil entendimento esta issue tem uma boa definição de critérios de aceitação esta issue tem labels associadas esta issue está associada à uma milestone esta issue tem uma pontuação estimada tarefas criar plano de qualidade critérios de aceitação a primeira versão do plano de qualidade deve estar elaborada | 1 |
1,694 | 6,553,962,218 | IssuesEvent | 2017-09-06 02:15:13 | City-Bureau/documenters-aggregator | https://api.github.com/repos/City-Bureau/documenters-aggregator | opened | What geocoder service should we use? | architecture: spiders priority: high (must have) | See https://github.com/City-Bureau/documenters-aggregator/pull/85#issuecomment-327325384
Leaning towards Mapbox for now and will use to close #85. | 1.0 | What geocoder service should we use? - See https://github.com/City-Bureau/documenters-aggregator/pull/85#issuecomment-327325384
Leaning towards Mapbox for now and will use to close #85. | architecture | what geocoder service should we use see leaning towards mapbox for now and will use to close | 1 |
2,097 | 7,276,508,762 | IssuesEvent | 2018-02-21 16:34:13 | AnalyticalGraphicsInc/cesium | https://api.github.com/repos/AnalyticalGraphicsInc/cesium | closed | CesiumMath vs Math naming ambiguity | category - architecture / api category - doc | It's not clear that the `CesiumMath` class is included in the namespace as `Cesium.Math`. This is also inconsistent with other classes that have the Cesium prefix, like `Cesium3DTileset`. If this is not something we want to change in the API, this should be made clear in the documentation.
Relevant forum thread: https://groups.google.com/forum/#!topic/cesium-dev/icpMxc_bea8 | 1.0 | CesiumMath vs Math naming ambiguity - It's not clear that the `CesiumMath` class is included in the namespace as `Cesium.Math`. This is also inconsistent with other classes that have the Cesium prefix, like `Cesium3DTileset`. If this is not something we want to change in the API, this should be made clear in the documentation.
Relevant forum thread: https://groups.google.com/forum/#!topic/cesium-dev/icpMxc_bea8 | architecture | cesiummath vs math naming ambiguity it s not clear that the cesiummath class is included in the namespace as cesium math this is also inconsistent with other classes that have the cesium prefix like if this is not something we want to change in the api this should be made clear in the documentation relevant forum thread | 1 |
179,282 | 21,557,595,193 | IssuesEvent | 2022-04-30 17:37:48 | NixOS/nixpkgs | https://api.github.com/repos/NixOS/nixpkgs | closed | Vulnerability roundup 113: ffmpeg-5.0.1: 1 advisory [7.5] | 1.severity: security | [search](https://search.nix.gsc.io/?q=ffmpeg&i=fosho&repos=NixOS-nixpkgs), [files](https://github.com/NixOS/nixpkgs/search?utf8=%E2%9C%93&q=ffmpeg+in%3Apath&type=Code)
* [ ] [CVE-2021-38291](https://nvd.nist.gov/vuln/detail/CVE-2021-38291) CVSSv3=7.5 (nixos-unstable)
## CVE details
### CVE-2021-38291
FFmpeg version (git commit de8e6e67e7523e48bb27ac224a0b446df05e1640) suffers from a an assertion failure at src/libavutil/mathematics.c.
-----
Scanned versions: nixos-unstable: ff9efb0724d.
Cc @codyopel
| True | Vulnerability roundup 113: ffmpeg-5.0.1: 1 advisory [7.5] - [search](https://search.nix.gsc.io/?q=ffmpeg&i=fosho&repos=NixOS-nixpkgs), [files](https://github.com/NixOS/nixpkgs/search?utf8=%E2%9C%93&q=ffmpeg+in%3Apath&type=Code)
* [ ] [CVE-2021-38291](https://nvd.nist.gov/vuln/detail/CVE-2021-38291) CVSSv3=7.5 (nixos-unstable)
## CVE details
### CVE-2021-38291
FFmpeg version (git commit de8e6e67e7523e48bb27ac224a0b446df05e1640) suffers from a an assertion failure at src/libavutil/mathematics.c.
-----
Scanned versions: nixos-unstable: ff9efb0724d.
Cc @codyopel
| non_architecture | vulnerability roundup ffmpeg advisory nixos unstable cve details cve ffmpeg version git commit suffers from a an assertion failure at src libavutil mathematics c scanned versions nixos unstable cc codyopel | 0 |
4,072 | 10,552,476,500 | IssuesEvent | 2019-10-03 15:14:11 | dotnet/docs | https://api.github.com/repos/dotnet/docs | closed | Multuple IHostedService registration | :book: guide - .NET Microservices :books: Area - .NET Architecture Guide Source - Docs.ms | If i try to register two or more services, only one could work properly.
For example:
```
services.AddSingleton<IHostedService, ServiceA>();
services.AddSingleton<IHostedService, ServiceB>();
```
Implementations are simplest as possible:
```
public class ServiceA: IHostedService
{
public Task StartAsync(CancellationToken cancellationToken)
{
DoWork();
return Task.CompletedTask;
}
public Task StopAsync(CancellationToken cancellationToken)
{
return Task.CompletedTask;
}
private void DoWork()
{
while (true)
{
Console.WriteLine("ServiceA");
Thread.Sleep(2000);
}
}
}
```
and
```
public class ServiceB: IHostedService
{
public Task StartAsync(CancellationToken cancellationToken)
{
DoWork();
return Task.CompletedTask;
}
public Task StopAsync(CancellationToken cancellationToken)
{
return Task.CompletedTask;
}
private void DoWork()
{
while (true)
{
Console.WriteLine("ServiceB");
Thread.Sleep(1000);
}
}
}
```
In output getting messages only from ServiceA
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: d49a03c0-a844-26eb-48a5-33a612dd3ead
* Version Independent ID: 0707865f-9db7-0d71-42a5-bc1a1e89680a
* Content: [Implement background tasks in microservices with IHostedService and the BackgroundService class](https://docs.microsoft.com/en-us/dotnet/architecture/microservices/multi-container-microservice-net-applications/background-tasks-with-ihostedservice#feedback)
* Content Source: [docs/architecture/microservices/multi-container-microservice-net-applications/background-tasks-with-ihostedservice.md](https://github.com/dotnet/docs/blob/master/docs/architecture/microservices/multi-container-microservice-net-applications/background-tasks-with-ihostedservice.md)
* Product: **dotnet**
* Technology: **dotnet-ebooks**
* GitHub Login: @nishanil
* Microsoft Alias: **nanil** | 1.0 | Multuple IHostedService registration - If i try to register two or more services, only one could work properly.
For example:
```
services.AddSingleton<IHostedService, ServiceA>();
services.AddSingleton<IHostedService, ServiceB>();
```
Implementations are simplest as possible:
```
public class ServiceA: IHostedService
{
public Task StartAsync(CancellationToken cancellationToken)
{
DoWork();
return Task.CompletedTask;
}
public Task StopAsync(CancellationToken cancellationToken)
{
return Task.CompletedTask;
}
private void DoWork()
{
while (true)
{
Console.WriteLine("ServiceA");
Thread.Sleep(2000);
}
}
}
```
and
```
public class ServiceB: IHostedService
{
public Task StartAsync(CancellationToken cancellationToken)
{
DoWork();
return Task.CompletedTask;
}
public Task StopAsync(CancellationToken cancellationToken)
{
return Task.CompletedTask;
}
private void DoWork()
{
while (true)
{
Console.WriteLine("ServiceB");
Thread.Sleep(1000);
}
}
}
```
In output getting messages only from ServiceA
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: d49a03c0-a844-26eb-48a5-33a612dd3ead
* Version Independent ID: 0707865f-9db7-0d71-42a5-bc1a1e89680a
* Content: [Implement background tasks in microservices with IHostedService and the BackgroundService class](https://docs.microsoft.com/en-us/dotnet/architecture/microservices/multi-container-microservice-net-applications/background-tasks-with-ihostedservice#feedback)
* Content Source: [docs/architecture/microservices/multi-container-microservice-net-applications/background-tasks-with-ihostedservice.md](https://github.com/dotnet/docs/blob/master/docs/architecture/microservices/multi-container-microservice-net-applications/background-tasks-with-ihostedservice.md)
* Product: **dotnet**
* Technology: **dotnet-ebooks**
* GitHub Login: @nishanil
* Microsoft Alias: **nanil** | architecture | multuple ihostedservice registration if i try to register two or more services only one could work properly for example services addsingleton services addsingleton implementations are simplest as possible public class servicea ihostedservice public task startasync cancellationtoken cancellationtoken dowork return task completedtask public task stopasync cancellationtoken cancellationtoken return task completedtask private void dowork while true console writeline servicea thread sleep and public class serviceb ihostedservice public task startasync cancellationtoken cancellationtoken dowork return task completedtask public task stopasync cancellationtoken cancellationtoken return task completedtask private void dowork while true console writeline serviceb thread sleep in output getting messages only from servicea document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source product dotnet technology dotnet ebooks github login nishanil microsoft alias nanil | 1 |
255,598 | 21,939,757,136 | IssuesEvent | 2022-05-23 16:49:11 | ooni/probe | https://api.github.com/repos/ooni/probe | closed | oonimkall: setRunType: cannot find symbol | bug testing ooni/probe-mobile priority/high platform/android platform/ios ooni/probe-engine | Linking with the oonimkall.aar engine fails with this error:
```
> Task :engine:compileExperimentalReleaseJavaWithJavac FAILED
$monorepo/repo/probe-android/engine/src/main/java/org/openobservatory/engine/OONICheckInConfig.java:67: error: cannot find symbol
c.setRunType(runType);
^
symbol: method setRunType(String)
location: variable c of type CheckInConfig
1 error
FAILURE: Build failed with an exception.
```
We need to fix this error in order to release 3.15.0.
(cc: @hellais @aanorbel) | 1.0 | oonimkall: setRunType: cannot find symbol - Linking with the oonimkall.aar engine fails with this error:
```
> Task :engine:compileExperimentalReleaseJavaWithJavac FAILED
$monorepo/repo/probe-android/engine/src/main/java/org/openobservatory/engine/OONICheckInConfig.java:67: error: cannot find symbol
c.setRunType(runType);
^
symbol: method setRunType(String)
location: variable c of type CheckInConfig
1 error
FAILURE: Build failed with an exception.
```
We need to fix this error in order to release 3.15.0.
(cc: @hellais @aanorbel) | non_architecture | oonimkall setruntype cannot find symbol linking with the oonimkall aar engine fails with this error task engine compileexperimentalreleasejavawithjavac failed monorepo repo probe android engine src main java org openobservatory engine oonicheckinconfig java error cannot find symbol c setruntype runtype symbol method setruntype string location variable c of type checkinconfig error failure build failed with an exception we need to fix this error in order to release cc hellais aanorbel | 0 |
1,559 | 6,335,238,819 | IssuesEvent | 2017-07-26 18:24:54 | LearnersGuild/echo | https://api.github.com/repos/LearnersGuild/echo | reopened | Move changefeed listeners to web service | architecture chore | Currently, the workers set up listening to db changefeeds and, effectively, queue items for their _own_ work queues. This is problematic because if changes to the database occur that should be handled by a worker while the worker is not running, that change event is lost and the job is never processed. Instead of connecting the changefeed listeners in the worker process, we should do it in the standing web service. This also makes it easier to choose to alternative mechanisms for background task execution (instead of having always-running worker dynos). | 1.0 | Move changefeed listeners to web service - Currently, the workers set up listening to db changefeeds and, effectively, queue items for their _own_ work queues. This is problematic because if changes to the database occur that should be handled by a worker while the worker is not running, that change event is lost and the job is never processed. Instead of connecting the changefeed listeners in the worker process, we should do it in the standing web service. This also makes it easier to choose to alternative mechanisms for background task execution (instead of having always-running worker dynos). | architecture | move changefeed listeners to web service currently the workers set up listening to db changefeeds and effectively queue items for their own work queues this is problematic because if changes to the database occur that should be handled by a worker while the worker is not running that change event is lost and the job is never processed instead of connecting the changefeed listeners in the worker process we should do it in the standing web service this also makes it easier to choose to alternative mechanisms for background task execution instead of having always running worker dynos | 1 |
796,666 | 28,123,191,979 | IssuesEvent | 2023-03-31 15:32:50 | thetrevorharmon/thetrevorharmon.com | https://api.github.com/repos/thetrevorharmon/thetrevorharmon.com | closed | Add landing page for mailing list | enhancement low priority | Now that there is a mailing list on the site, It would be good to have a landing page for the signup. Something like `/signup` with a simple form and nice explanation of what the signup gets them. | 1.0 | Add landing page for mailing list - Now that there is a mailing list on the site, It would be good to have a landing page for the signup. Something like `/signup` with a simple form and nice explanation of what the signup gets them. | non_architecture | add landing page for mailing list now that there is a mailing list on the site it would be good to have a landing page for the signup something like signup with a simple form and nice explanation of what the signup gets them | 0 |
5,105 | 12,098,281,050 | IssuesEvent | 2020-04-20 10:02:29 | stsrki/Blazorise | https://api.github.com/repos/stsrki/Blazorise | closed | Unit testing of components | Status: Investigate Type: Architecture | Investigate more about the new unit testing made by Steve Sanderson, after the release of Blazor preview 9.
Sources:
http://blog.stevensanderson.com/2019/08/29/blazor-unit-testing-prototype/
https://github.com/SteveSandersonMS/BlazorUnitTestingPrototype
| 1.0 | Unit testing of components - Investigate more about the new unit testing made by Steve Sanderson, after the release of Blazor preview 9.
Sources:
http://blog.stevensanderson.com/2019/08/29/blazor-unit-testing-prototype/
https://github.com/SteveSandersonMS/BlazorUnitTestingPrototype
| architecture | unit testing of components investigate more about the new unit testing made by steve sanderson after the release of blazor preview sources | 1 |
357,018 | 10,600,740,255 | IssuesEvent | 2019-10-10 10:43:48 | robotology/whole-body-controllers | https://api.github.com/repos/robotology/whole-body-controllers | opened | Investigate if it makes sense to port matlab-multi-body-sim in wbc | feature priority: normal | I would like to port `matlab-multi-body-sim` in the wbc, but before I need to understand the ratio between effort and benefits | 1.0 | Investigate if it makes sense to port matlab-multi-body-sim in wbc - I would like to port `matlab-multi-body-sim` in the wbc, but before I need to understand the ratio between effort and benefits | non_architecture | investigate if it makes sense to port matlab multi body sim in wbc i would like to port matlab multi body sim in the wbc but before i need to understand the ratio between effort and benefits | 0 |
6,035 | 13,541,185,291 | IssuesEvent | 2020-09-16 15:33:34 | MicrosoftDocs/architecture-center | https://api.github.com/repos/MicrosoftDocs/architecture-center | closed | Naming conventions of Icons | Pri2 architecture-center/svc assigned-to-author doc-enhancement triaged |
We are doing some work with ARM templates and attempting to use the latest Icon sets.
The problem we are seeing is there is no consistence way in which the icons are named which can map to templates or schemas.
Please could it be considered .
e.g Template schemas refer to Virtual scale sets as 'virtualMachineScaleSets', but the icon is buried in the Compute directory called "10034-icon-service-VM-Scalte-Sets".
How are we meant to quick / auto identify between the too. Before the naming convention did link back to the recourses types
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 6624d44c-f9f6-02f8-2c67-0f0637fed62e
* Version Independent ID: 9346a8ff-bab5-be2c-9dc1-d546fda1efa9
* Content: [Azure Icons - Azure Architecture Center](https://docs.microsoft.com/en-us/azure/architecture/icons/)
* Content Source: [docs/icons/index.md](https://github.com/microsoftdocs/architecture-center/blob/master/docs/icons/index.md)
* Service: **architecture-center**
* GitHub Login: @doodlemania2
* Microsoft Alias: **pnp** | 1.0 | Naming conventions of Icons -
We are doing some work with ARM templates and attempting to use the latest Icon sets.
The problem we are seeing is there is no consistence way in which the icons are named which can map to templates or schemas.
Please could it be considered .
e.g Template schemas refer to Virtual scale sets as 'virtualMachineScaleSets', but the icon is buried in the Compute directory called "10034-icon-service-VM-Scalte-Sets".
How are we meant to quick / auto identify between the too. Before the naming convention did link back to the recourses types
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 6624d44c-f9f6-02f8-2c67-0f0637fed62e
* Version Independent ID: 9346a8ff-bab5-be2c-9dc1-d546fda1efa9
* Content: [Azure Icons - Azure Architecture Center](https://docs.microsoft.com/en-us/azure/architecture/icons/)
* Content Source: [docs/icons/index.md](https://github.com/microsoftdocs/architecture-center/blob/master/docs/icons/index.md)
* Service: **architecture-center**
* GitHub Login: @doodlemania2
* Microsoft Alias: **pnp** | architecture | naming conventions of icons we are doing some work with arm templates and attempting to use the latest icon sets the problem we are seeing is there is no consistence way in which the icons are named which can map to templates or schemas please could it be considered e g template schemas refer to virtual scale sets as virtualmachinescalesets but the icon is buried in the compute directory called icon service vm scalte sets how are we meant to quick auto identify between the too before the naming convention did link back to the recourses types document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source service architecture center github login microsoft alias pnp | 1 |
42,411 | 5,444,055,594 | IssuesEvent | 2017-03-07 01:15:24 | dotnet/roslyn | https://api.github.com/repos/dotnet/roslyn | closed | Alternative approach to introduction of variables into scope. | Area-Language Design Discussion | There have a been a few raised voices that the current and suggested approach to how the new language features introduce variable into scope. He is my considered approach.
---
**Default to nearest outer scope**
The variable being introduce should be to **nearest outer scope** by default.
``` c#
if( !(o is int i) )
{
/* i is in scope and definitely not assigned */
throw new ArgumentException("Not an int", nameof(o));
}
/* i is in scope and definitely assigned. */
```
It simplifies one of the initial use cases, and I think the most likely usage.
``` c#
int value;
if( int.TryParse( text , out value )
```
into
``` c#
if( int.TryParse( text, out int value )
```
---
It is the more eccentric usage of localizing to the nearest inner scope (of this particular code block). eg the `then` and `else` blocks. That is cause of tension for the community. especially around where it is introduce in "patterns". I propose that in these cases the change from the default (nearest outer) should made explicit and require the coder to specify a change to the **nearest inner scope**.
For example a lightweight approach to this is by prefixing `~` on the variable identifier.
``` c#
if (o is int ~i)
{
/* i is scope and definitely assigned */
}
/* i is not in scope. /*
```
``` c#
if( !(o is int ~i) )
{
/* i is scope finitely not assigned */
}
else
{
/* I is scope and definitely assigned */
}
/* i is not in scope. /*
```
---
Then there are the case where the coder would like the variable introduced in one of nearest blocks and not the other. Use case: Guards.
``` c#
if( !int.TryParst( text, value ) )
{
/* value not in scope. */
}
else
{
/* value is in scope and assigned */
}
```
In pattern I propose we borrow from VB.net and use `IsNot` to indicate the negation of the pattern.
``` c#
if (o isnot int ~i)
{
/* i is scope not assigned a new value*/
/* also is an error as i is being reused for a variable declaration */
}
else
{
/* i is scope and definitely assigned a value.*/
/* also is an error as i is being reused for a variable declaration */
}
/* i is in scope. /*
```
This should be easier to read ( ie not missing the easily missed `!` at the start. ), and allow us to know that the usage of variable introduction could be different.
---
**When the variable identifier prior exists**
What about the case where the variable already exist prior the usage?
With the default ( nearest outer scope) the variable is reused.
``` c#
int i;
...
if( !(o is int i) )
{
/* i is in scope and definitely maybe assigned */
throw new ArgumentException("Not an int", nameof(o));
}
/* i is in scope and definitely assigned. */
```
And in the nearest inner scope the variable can not reused and thus produces an compile-time error.
Forcing the coder to make a explicit choice, and either make scope outer or use a different variable identifier.
``` c#
int i;
...
if (o is int ~i)
{
/* i is scope and definitely assigned */
/* also is and error as i is be reused for a variable declaration */
}
/* i is in scope. /*
```
``` c#
int i;
...
if (o isnot int ~i)
{
/* i is scope not assigned a new value*/
/* also is an error as i is being reused for a variable declaration */
}
else
{
/* i is scope and definitely assigned a value.*/
/* also is an error as i is being reused for a variable declaration */
}
/* i is in scope. /*
```
if the type of `i` is incompatible with the one in the pattern, it is an error in all cases.
| 1.0 | Alternative approach to introduction of variables into scope. - There have a been a few raised voices that the current and suggested approach to how the new language features introduce variable into scope. He is my considered approach.
---
**Default to nearest outer scope**
The variable being introduce should be to **nearest outer scope** by default.
``` c#
if( !(o is int i) )
{
/* i is in scope and definitely not assigned */
throw new ArgumentException("Not an int", nameof(o));
}
/* i is in scope and definitely assigned. */
```
It simplifies one of the initial use cases, and I think the most likely usage.
``` c#
int value;
if( int.TryParse( text , out value )
```
into
``` c#
if( int.TryParse( text, out int value )
```
---
It is the more eccentric usage of localizing to the nearest inner scope (of this particular code block). eg the `then` and `else` blocks. That is cause of tension for the community. especially around where it is introduce in "patterns". I propose that in these cases the change from the default (nearest outer) should made explicit and require the coder to specify a change to the **nearest inner scope**.
For example a lightweight approach to this is by prefixing `~` on the variable identifier.
``` c#
if (o is int ~i)
{
/* i is scope and definitely assigned */
}
/* i is not in scope. /*
```
``` c#
if( !(o is int ~i) )
{
/* i is scope finitely not assigned */
}
else
{
/* I is scope and definitely assigned */
}
/* i is not in scope. /*
```
---
Then there are the case where the coder would like the variable introduced in one of nearest blocks and not the other. Use case: Guards.
``` c#
if( !int.TryParst( text, value ) )
{
/* value not in scope. */
}
else
{
/* value is in scope and assigned */
}
```
In pattern I propose we borrow from VB.net and use `IsNot` to indicate the negation of the pattern.
``` c#
if (o isnot int ~i)
{
/* i is scope not assigned a new value*/
/* also is an error as i is being reused for a variable declaration */
}
else
{
/* i is scope and definitely assigned a value.*/
/* also is an error as i is being reused for a variable declaration */
}
/* i is in scope. /*
```
This should be easier to read ( ie not missing the easily missed `!` at the start. ), and allow us to know that the usage of variable introduction could be different.
---
**When the variable identifier prior exists**
What about the case where the variable already exist prior the usage?
With the default ( nearest outer scope) the variable is reused.
``` c#
int i;
...
if( !(o is int i) )
{
/* i is in scope and definitely maybe assigned */
throw new ArgumentException("Not an int", nameof(o));
}
/* i is in scope and definitely assigned. */
```
And in the nearest inner scope the variable can not reused and thus produces an compile-time error.
Forcing the coder to make a explicit choice, and either make scope outer or use a different variable identifier.
``` c#
int i;
...
if (o is int ~i)
{
/* i is scope and definitely assigned */
/* also is and error as i is be reused for a variable declaration */
}
/* i is in scope. /*
```
``` c#
int i;
...
if (o isnot int ~i)
{
/* i is scope not assigned a new value*/
/* also is an error as i is being reused for a variable declaration */
}
else
{
/* i is scope and definitely assigned a value.*/
/* also is an error as i is being reused for a variable declaration */
}
/* i is in scope. /*
```
if the type of `i` is incompatible with the one in the pattern, it is an error in all cases.
| non_architecture | alternative approach to introduction of variables into scope there have a been a few raised voices that the current and suggested approach to how the new language features introduce variable into scope he is my considered approach default to nearest outer scope the variable being introduce should be to nearest outer scope by default c if o is int i i is in scope and definitely not assigned throw new argumentexception not an int nameof o i is in scope and definitely assigned it simplifies one of the initial use cases and i think the most likely usage c int value if int tryparse text out value into c if int tryparse text out int value it is the more eccentric usage of localizing to the nearest inner scope of this particular code block eg the then and else blocks that is cause of tension for the community especially around where it is introduce in patterns i propose that in these cases the change from the default nearest outer should made explicit and require the coder to specify a change to the nearest inner scope for example a lightweight approach to this is by prefixing on the variable identifier c if o is int i i is scope and definitely assigned i is not in scope c if o is int i i is scope finitely not assigned else i is scope and definitely assigned i is not in scope then there are the case where the coder would like the variable introduced in one of nearest blocks and not the other use case guards c if int tryparst text value value not in scope else value is in scope and assigned in pattern i propose we borrow from vb net and use isnot to indicate the negation of the pattern c if o isnot int i i is scope not assigned a new value also is an error as i is being reused for a variable declaration else i is scope and definitely assigned a value also is an error as i is being reused for a variable declaration i is in scope this should be easier to read ie not missing the easily missed at the start and allow us to know that the usage of variable introduction could be different when the variable identifier prior exists what about the case where the variable already exist prior the usage with the default nearest outer scope the variable is reused c int i if o is int i i is in scope and definitely maybe assigned throw new argumentexception not an int nameof o i is in scope and definitely assigned and in the nearest inner scope the variable can not reused and thus produces an compile time error forcing the coder to make a explicit choice and either make scope outer or use a different variable identifier c int i if o is int i i is scope and definitely assigned also is and error as i is be reused for a variable declaration i is in scope c int i if o isnot int i i is scope not assigned a new value also is an error as i is being reused for a variable declaration else i is scope and definitely assigned a value also is an error as i is being reused for a variable declaration i is in scope if the type of i is incompatible with the one in the pattern it is an error in all cases | 0 |
9,595 | 24,873,438,277 | IssuesEvent | 2022-10-27 16:59:01 | Azure/azure-sdk | https://api.github.com/repos/Azure/azure-sdk | opened | Board Review: <client library name> | architecture board-review | Thank you for submitting this review request. Thorough review of your client library ensures that your APIs are consistent with the guidelines and the consumers of your client library have a consistently good experience when using Azure.
**The Architecture Board reviews [Track 2 libraries](https://azure.github.io/azure-sdk/general_introduction.html) only.** If your library does not meet this requirement, please reach out to [Architecture Board]([email protected]) before creating the issue.
Please reference our [review process guidelines](https://azure.github.io/azure-sdk/policies_reviewprocess.html) to understand what is being asked for in the issue template.
To ensure consistency, all Tier-1 languages (C#, TypeScript, Java, Python) will generally be reviewed together. In expansive libraries, we will pair dynamic languages (Python, TypeScript) together, and strongly typed languages (C#, Java) together in separate meetings.
For Tier-2 languages (C, C++, Go, Android, iOS), the review will be on an as-needed basis.
**Before submitting, ensure you adjust the title of the issue appropriately.**
**Note that the required material must be included before a meeting can be scheduled.**
## Contacts and Timeline
* Responsible service team: Liftr Nginx
* Main contacts: @SpencerOfwiti @limingu
* Expected code complete date: Not Applicable
* Expected release date:
## About the Service
* Link to documentation introducing/describing the service: https://learn.microsoft.com/en-us/azure/partner-solutions/nginx/
* Link to the service REST APIs: https://github.com/Azure/azure-rest-api-specs/tree/main/specification/nginx/resource-manager/NGINX.NGINXPLUS/stable/2022-08-01
* Link to GitHub issue for previous review sessions, if applicable:
## About the client library
* Name of the client library:
* Languages for this review:
The SDKs are autogenerated from the swagger, this review is only for namespace approval.
## Thank you!
| 1.0 | Board Review: <client library name> - Thank you for submitting this review request. Thorough review of your client library ensures that your APIs are consistent with the guidelines and the consumers of your client library have a consistently good experience when using Azure.
**The Architecture Board reviews [Track 2 libraries](https://azure.github.io/azure-sdk/general_introduction.html) only.** If your library does not meet this requirement, please reach out to [Architecture Board]([email protected]) before creating the issue.
Please reference our [review process guidelines](https://azure.github.io/azure-sdk/policies_reviewprocess.html) to understand what is being asked for in the issue template.
To ensure consistency, all Tier-1 languages (C#, TypeScript, Java, Python) will generally be reviewed together. In expansive libraries, we will pair dynamic languages (Python, TypeScript) together, and strongly typed languages (C#, Java) together in separate meetings.
For Tier-2 languages (C, C++, Go, Android, iOS), the review will be on an as-needed basis.
**Before submitting, ensure you adjust the title of the issue appropriately.**
**Note that the required material must be included before a meeting can be scheduled.**
## Contacts and Timeline
* Responsible service team: Liftr Nginx
* Main contacts: @SpencerOfwiti @limingu
* Expected code complete date: Not Applicable
* Expected release date:
## About the Service
* Link to documentation introducing/describing the service: https://learn.microsoft.com/en-us/azure/partner-solutions/nginx/
* Link to the service REST APIs: https://github.com/Azure/azure-rest-api-specs/tree/main/specification/nginx/resource-manager/NGINX.NGINXPLUS/stable/2022-08-01
* Link to GitHub issue for previous review sessions, if applicable:
## About the client library
* Name of the client library:
* Languages for this review:
The SDKs are autogenerated from the swagger, this review is only for namespace approval.
## Thank you!
| architecture | board review thank you for submitting this review request thorough review of your client library ensures that your apis are consistent with the guidelines and the consumers of your client library have a consistently good experience when using azure the architecture board reviews only if your library does not meet this requirement please reach out to adparch microsoft com before creating the issue please reference our to understand what is being asked for in the issue template to ensure consistency all tier languages c typescript java python will generally be reviewed together in expansive libraries we will pair dynamic languages python typescript together and strongly typed languages c java together in separate meetings for tier languages c c go android ios the review will be on an as needed basis before submitting ensure you adjust the title of the issue appropriately note that the required material must be included before a meeting can be scheduled contacts and timeline responsible service team liftr nginx main contacts spencerofwiti limingu expected code complete date not applicable expected release date about the service link to documentation introducing describing the service link to the service rest apis link to github issue for previous review sessions if applicable about the client library name of the client library languages for this review the sdks are autogenerated from the swagger this review is only for namespace approval thank you | 1 |
213,119 | 23,966,109,535 | IssuesEvent | 2022-09-13 01:12:33 | DavidSpek/kubeflow | https://api.github.com/repos/DavidSpek/kubeflow | opened | CVE-2022-36083 (Medium) detected in jose-2.0.5.tgz | security vulnerability | ## CVE-2022-36083 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jose-2.0.5.tgz</b></p></summary>
<p>JSON Web Almost Everything - JWA, JWS, JWE, JWK, JWT, JWKS for Node.js with minimal dependencies</p>
<p>Library home page: <a href="https://registry.npmjs.org/jose/-/jose-2.0.5.tgz">https://registry.npmjs.org/jose/-/jose-2.0.5.tgz</a></p>
<p>Path to dependency file: /components/crud-web-apps/volumes/frontend/package.json</p>
<p>Path to vulnerable library: /components/crud-web-apps/volumes/frontend/node_modules/jose/package.json,/components/crud-web-apps/jupyter/frontend/node_modules/jose/package.json,/components/crud-web-apps/common/frontend/kubeflow-common-lib/node_modules/jose/package.json,/components/crud-web-apps/tensorboards/frontend/node_modules/jose/package.json</p>
<p>
Dependency Hierarchy:
- client-node-0.12.2.tgz (Root Library)
- openid-client-4.2.2.tgz
- :x: **jose-2.0.5.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/DavidSpek/kubeflow/commit/00cbc9d11a3306fed1e979d79dff6ae36749d4bd">00cbc9d11a3306fed1e979d79dff6ae36749d4bd</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
JOSE is "JSON Web Almost Everything" - JWA, JWS, JWE, JWT, JWK, JWKS with no dependencies using runtime's native crypto in Node.js, Browser, Cloudflare Workers, Electron, and Deno. The PBKDF2-based JWE key management algorithms expect a JOSE Header Parameter named `p2c` PBES2 Count, which determines how many PBKDF2 iterations must be executed in order to derive a CEK wrapping key. The purpose of this parameter is to intentionally slow down the key derivation function in order to make password brute-force and dictionary attacks more expensive. This makes the PBES2 algorithms unsuitable for situations where the JWE is coming from an untrusted source: an adversary can intentionally pick an extremely high PBES2 Count value, that will initiate a CPU-bound computation that may take an unreasonable amount of time to finish. Under certain conditions, it is possible to have the user's environment consume unreasonable amount of CPU time. The impact is limited only to users utilizing the JWE decryption APIs with symmetric secrets to decrypt JWEs from untrusted parties who do not limit the accepted JWE Key Management Algorithms (`alg` Header Parameter) using the `keyManagementAlgorithms` (or `algorithms` in v1.x) decryption option or through other means. The `v1.28.2`, `v2.0.6`, `v3.20.4`, and `v4.9.2` releases limit the maximum PBKDF2 iteration count to `10000` by default. It is possible to adjust this limit with a newly introduced `maxPBES2Count` decryption option. If users are unable to upgrade their required library version, they have two options depending on whether they expect to receive JWEs using any of the three PBKDF2-based JWE key management algorithms. They can use the `keyManagementAlgorithms` decryption option to disable accepting PBKDF2 altogether, or they can inspect the JOSE Header prior to using the decryption API and limit the PBKDF2 iteration count (`p2c` Header Parameter).
<p>Publish Date: 2022-09-07
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-36083>CVE-2022-36083</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/panva/jose/security/advisories/GHSA-jv3g-j58f-9mq9">https://github.com/panva/jose/security/advisories/GHSA-jv3g-j58f-9mq9</a></p>
<p>Release Date: 2022-09-07</p>
<p>Fix Resolution (jose): 2.0.6</p>
<p>Direct dependency fix Resolution (@kubernetes/client-node): 0.12.3</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2022-36083 (Medium) detected in jose-2.0.5.tgz - ## CVE-2022-36083 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jose-2.0.5.tgz</b></p></summary>
<p>JSON Web Almost Everything - JWA, JWS, JWE, JWK, JWT, JWKS for Node.js with minimal dependencies</p>
<p>Library home page: <a href="https://registry.npmjs.org/jose/-/jose-2.0.5.tgz">https://registry.npmjs.org/jose/-/jose-2.0.5.tgz</a></p>
<p>Path to dependency file: /components/crud-web-apps/volumes/frontend/package.json</p>
<p>Path to vulnerable library: /components/crud-web-apps/volumes/frontend/node_modules/jose/package.json,/components/crud-web-apps/jupyter/frontend/node_modules/jose/package.json,/components/crud-web-apps/common/frontend/kubeflow-common-lib/node_modules/jose/package.json,/components/crud-web-apps/tensorboards/frontend/node_modules/jose/package.json</p>
<p>
Dependency Hierarchy:
- client-node-0.12.2.tgz (Root Library)
- openid-client-4.2.2.tgz
- :x: **jose-2.0.5.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/DavidSpek/kubeflow/commit/00cbc9d11a3306fed1e979d79dff6ae36749d4bd">00cbc9d11a3306fed1e979d79dff6ae36749d4bd</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
JOSE is "JSON Web Almost Everything" - JWA, JWS, JWE, JWT, JWK, JWKS with no dependencies using runtime's native crypto in Node.js, Browser, Cloudflare Workers, Electron, and Deno. The PBKDF2-based JWE key management algorithms expect a JOSE Header Parameter named `p2c` PBES2 Count, which determines how many PBKDF2 iterations must be executed in order to derive a CEK wrapping key. The purpose of this parameter is to intentionally slow down the key derivation function in order to make password brute-force and dictionary attacks more expensive. This makes the PBES2 algorithms unsuitable for situations where the JWE is coming from an untrusted source: an adversary can intentionally pick an extremely high PBES2 Count value, that will initiate a CPU-bound computation that may take an unreasonable amount of time to finish. Under certain conditions, it is possible to have the user's environment consume unreasonable amount of CPU time. The impact is limited only to users utilizing the JWE decryption APIs with symmetric secrets to decrypt JWEs from untrusted parties who do not limit the accepted JWE Key Management Algorithms (`alg` Header Parameter) using the `keyManagementAlgorithms` (or `algorithms` in v1.x) decryption option or through other means. The `v1.28.2`, `v2.0.6`, `v3.20.4`, and `v4.9.2` releases limit the maximum PBKDF2 iteration count to `10000` by default. It is possible to adjust this limit with a newly introduced `maxPBES2Count` decryption option. If users are unable to upgrade their required library version, they have two options depending on whether they expect to receive JWEs using any of the three PBKDF2-based JWE key management algorithms. They can use the `keyManagementAlgorithms` decryption option to disable accepting PBKDF2 altogether, or they can inspect the JOSE Header prior to using the decryption API and limit the PBKDF2 iteration count (`p2c` Header Parameter).
<p>Publish Date: 2022-09-07
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-36083>CVE-2022-36083</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/panva/jose/security/advisories/GHSA-jv3g-j58f-9mq9">https://github.com/panva/jose/security/advisories/GHSA-jv3g-j58f-9mq9</a></p>
<p>Release Date: 2022-09-07</p>
<p>Fix Resolution (jose): 2.0.6</p>
<p>Direct dependency fix Resolution (@kubernetes/client-node): 0.12.3</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_architecture | cve medium detected in jose tgz cve medium severity vulnerability vulnerable library jose tgz json web almost everything jwa jws jwe jwk jwt jwks for node js with minimal dependencies library home page a href path to dependency file components crud web apps volumes frontend package json path to vulnerable library components crud web apps volumes frontend node modules jose package json components crud web apps jupyter frontend node modules jose package json components crud web apps common frontend kubeflow common lib node modules jose package json components crud web apps tensorboards frontend node modules jose package json dependency hierarchy client node tgz root library openid client tgz x jose tgz vulnerable library found in head commit a href found in base branch master vulnerability details jose is json web almost everything jwa jws jwe jwt jwk jwks with no dependencies using runtime s native crypto in node js browser cloudflare workers electron and deno the based jwe key management algorithms expect a jose header parameter named count which determines how many iterations must be executed in order to derive a cek wrapping key the purpose of this parameter is to intentionally slow down the key derivation function in order to make password brute force and dictionary attacks more expensive this makes the algorithms unsuitable for situations where the jwe is coming from an untrusted source an adversary can intentionally pick an extremely high count value that will initiate a cpu bound computation that may take an unreasonable amount of time to finish under certain conditions it is possible to have the user s environment consume unreasonable amount of cpu time the impact is limited only to users utilizing the jwe decryption apis with symmetric secrets to decrypt jwes from untrusted parties who do not limit the accepted jwe key management algorithms alg header parameter using the keymanagementalgorithms or algorithms in x decryption option or through other means the and releases limit the maximum iteration count to by default it is possible to adjust this limit with a newly introduced decryption option if users are unable to upgrade their required library version they have two options depending on whether they expect to receive jwes using any of the three based jwe key management algorithms they can use the keymanagementalgorithms decryption option to disable accepting altogether or they can inspect the jose header prior to using the decryption api and limit the iteration count header parameter publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution jose direct dependency fix resolution kubernetes client node step up your open source security game with mend | 0 |
221,426 | 24,630,302,413 | IssuesEvent | 2022-10-17 01:00:10 | MendDemo-josh/moby | https://api.github.com/repos/MendDemo-josh/moby | closed | libiberty9.1.0: 1 vulnerabilities (highest severity is: 7.5) - autoclosed | security vulnerability | <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>libiberty9.1.0</b></p></summary>
<p>
<p>Library home page: <a href=https://ftp.gnu.org/pub/gnu/gcc/gcc-9.1.0/?wsslib=libiberty>https://ftp.gnu.org/pub/gnu/gcc/gcc-9.1.0/?wsslib=libiberty</a></p>
</p>
</p></p>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
</p>
<p></p></details>
## Vulnerabilities
| CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in | Remediation Available |
| ------------- | ------------- | ----- | ----- | ----- | --- | --- |
| [CVE-2022-2879](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-2879) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | libiberty9.1.0 | Direct | go1.18.7,go1.19.2 | ❌ |
## Details
<details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2022-2879</summary>
### Vulnerable Library - <b>libiberty9.1.0</b></p>
<p>
<p>Library home page: <a href=https://ftp.gnu.org/pub/gnu/gcc/gcc-9.1.0/?wsslib=libiberty>https://ftp.gnu.org/pub/gnu/gcc/gcc-9.1.0/?wsslib=libiberty</a></p>
<p>Found in base branch: <b>master</b></p></p>
</p></p>
### Vulnerable Source Files (1)
<p></p>
<p>
</p>
<p></p>
</p>
<p></p>
### Vulnerability Details
<p>
Reader.Read does not set a limit on the maximum size of file headers. A maliciously crafted archive could cause Read to allocate unbounded amounts of memory, potentially causing resource exhaustion or panics. After fix, Reader.Read limits the maximum size of header blocks to 1 MiB.
<p>Publish Date: 2022-10-14
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-2879>CVE-2022-2879</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>7.5</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://pkg.go.dev/vuln/GO-2022-1037">https://pkg.go.dev/vuln/GO-2022-1037</a></p>
<p>Release Date: 2022-10-14</p>
<p>Fix Resolution: go1.18.7,go1.19.2</p>
</p>
<p></p>
</details> | True | libiberty9.1.0: 1 vulnerabilities (highest severity is: 7.5) - autoclosed - <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>libiberty9.1.0</b></p></summary>
<p>
<p>Library home page: <a href=https://ftp.gnu.org/pub/gnu/gcc/gcc-9.1.0/?wsslib=libiberty>https://ftp.gnu.org/pub/gnu/gcc/gcc-9.1.0/?wsslib=libiberty</a></p>
</p>
</p></p>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
</p>
<p></p></details>
## Vulnerabilities
| CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in | Remediation Available |
| ------------- | ------------- | ----- | ----- | ----- | --- | --- |
| [CVE-2022-2879](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-2879) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | libiberty9.1.0 | Direct | go1.18.7,go1.19.2 | ❌ |
## Details
<details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2022-2879</summary>
### Vulnerable Library - <b>libiberty9.1.0</b></p>
<p>
<p>Library home page: <a href=https://ftp.gnu.org/pub/gnu/gcc/gcc-9.1.0/?wsslib=libiberty>https://ftp.gnu.org/pub/gnu/gcc/gcc-9.1.0/?wsslib=libiberty</a></p>
<p>Found in base branch: <b>master</b></p></p>
</p></p>
### Vulnerable Source Files (1)
<p></p>
<p>
</p>
<p></p>
</p>
<p></p>
### Vulnerability Details
<p>
Reader.Read does not set a limit on the maximum size of file headers. A maliciously crafted archive could cause Read to allocate unbounded amounts of memory, potentially causing resource exhaustion or panics. After fix, Reader.Read limits the maximum size of header blocks to 1 MiB.
<p>Publish Date: 2022-10-14
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-2879>CVE-2022-2879</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>7.5</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://pkg.go.dev/vuln/GO-2022-1037">https://pkg.go.dev/vuln/GO-2022-1037</a></p>
<p>Release Date: 2022-10-14</p>
<p>Fix Resolution: go1.18.7,go1.19.2</p>
</p>
<p></p>
</details> | non_architecture | vulnerabilities highest severity is autoclosed vulnerable library library home page a href vulnerable source files vulnerabilities cve severity cvss dependency type fixed in remediation available high direct details cve vulnerable library library home page a href found in base branch master vulnerable source files vulnerability details reader read does not set a limit on the maximum size of file headers a maliciously crafted archive could cause read to allocate unbounded amounts of memory potentially causing resource exhaustion or panics after fix reader read limits the maximum size of header blocks to mib publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution | 0 |
619 | 3,862,146,688 | IssuesEvent | 2016-04-08 00:43:28 | rails-oceania/melbourne-ruby | https://api.github.com/repos/rails-oceania/melbourne-ruby | closed | Event Sourcing, part 2 | 40min architecture code intermediate presentation | I'd like to follow up my RubyConf talk which was meant to make people curious about Event Sourcing, with a talk about some of the practicalities and where to start.
For those who missed my RubyConf talk, I'll quickly recap that before launching into particulars.
I'd allow for 45 mins for this talk. | 1.0 | Event Sourcing, part 2 - I'd like to follow up my RubyConf talk which was meant to make people curious about Event Sourcing, with a talk about some of the practicalities and where to start.
For those who missed my RubyConf talk, I'll quickly recap that before launching into particulars.
I'd allow for 45 mins for this talk. | architecture | event sourcing part i d like to follow up my rubyconf talk which was meant to make people curious about event sourcing with a talk about some of the practicalities and where to start for those who missed my rubyconf talk i ll quickly recap that before launching into particulars i d allow for mins for this talk | 1 |
10,480 | 27,022,476,040 | IssuesEvent | 2023-02-11 06:36:04 | jsolly/blogthedata | https://api.github.com/repos/jsolly/blogthedata | closed | Separate Porfolio page into its own page, disconnecting it from being a category | Architecture SEO | #### Context
The current implementation of the portfolio page is quite hacky. It's technically a 'category' which causes all kinds of wonkiness because I am having to add lots of conditional logic inside categories.html in order to handle special cases on the portfolio page.
I also have to have more logic for meta tags as the portfolio page has it's own tags that are unique from the other category pages.
#### Ideal behavior
The portfolio page is on it's own. It's not a 'category' of posts.
#### Things to consider
Will have to refactor the how posts are brought onto a page that is not a category
Might need to do a database migration to remove the 'category' attribute from portfolio posts.
Will need to refactor the category templates to remove the special portfolio logic. | 1.0 | Separate Porfolio page into its own page, disconnecting it from being a category - #### Context
The current implementation of the portfolio page is quite hacky. It's technically a 'category' which causes all kinds of wonkiness because I am having to add lots of conditional logic inside categories.html in order to handle special cases on the portfolio page.
I also have to have more logic for meta tags as the portfolio page has it's own tags that are unique from the other category pages.
#### Ideal behavior
The portfolio page is on it's own. It's not a 'category' of posts.
#### Things to consider
Will have to refactor the how posts are brought onto a page that is not a category
Might need to do a database migration to remove the 'category' attribute from portfolio posts.
Will need to refactor the category templates to remove the special portfolio logic. | architecture | separate porfolio page into its own page disconnecting it from being a category context the current implementation of the portfolio page is quite hacky it s technically a category which causes all kinds of wonkiness because i am having to add lots of conditional logic inside categories html in order to handle special cases on the portfolio page i also have to have more logic for meta tags as the portfolio page has it s own tags that are unique from the other category pages ideal behavior the portfolio page is on it s own it s not a category of posts things to consider will have to refactor the how posts are brought onto a page that is not a category might need to do a database migration to remove the category attribute from portfolio posts will need to refactor the category templates to remove the special portfolio logic | 1 |
1,169 | 5,221,420,375 | IssuesEvent | 2017-01-27 01:28:11 | jung-digital/ringa | https://api.github.com/repos/jung-digital/ringa | opened | Command: add no wait operator [[]] | architecture | Target Code:
```
controller.addListener([
Command1,
[[Command2]],
Command3
]);
```
Command 1 should run, Command2 should be started, but then Command3 should be run immediately without waiting for Command2 to finish. | 1.0 | Command: add no wait operator [[]] - Target Code:
```
controller.addListener([
Command1,
[[Command2]],
Command3
]);
```
Command 1 should run, Command2 should be started, but then Command3 should be run immediately without waiting for Command2 to finish. | architecture | command add no wait operator target code controller addlistener command should run should be started but then should be run immediately without waiting for to finish | 1 |
9,819 | 25,289,125,063 | IssuesEvent | 2022-11-16 22:07:07 | spring-projects/sts4 | https://api.github.com/repos/spring-projects/sts4 | closed | exception thrown in VSCode when using latest snapshots | type: bug status: needs-investigation theme: internal-architecture for: vscode theme: refactoring | I am using VSCode with the latest pre-releases from:
- Language Support for Java
- Spring Boot Dashboard
- Spring Boot Tools (from the latest VSIX file)
I have a project open in my workspace, created from initializr (web + actuator), on Spring Boot 2.6.12.
After a little while, I see an error popup showing up, complaining about a problem when asking for `textDocument/codeAction`, and showing this exception in the log output:
```
java.util.concurrent.CompletionException: com.google.gson.JsonSyntaxException: java.lang.IllegalStateException: Expected BEGIN_ARRAY but was BEGIN_OBJECT at path $
at java.base/java.util.concurrent.CompletableFuture.encodeThrowable(Unknown Source)
at java.base/java.util.concurrent.CompletableFuture.completeThrowable(Unknown Source)
at java.base/java.util.concurrent.CompletableFuture$UniApply.tryFire(Unknown Source)
at java.base/java.util.concurrent.CompletableFuture$Completion.run(Unknown Source)
at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source)
at java.base/java.lang.Thread.run(Unknown Source)
Caused by: com.google.gson.JsonSyntaxException: java.lang.IllegalStateException: Expected BEGIN_ARRAY but was BEGIN_OBJECT at path $
at com.google.gson.Gson.fromJson(Gson.java:1070)
at com.google.gson.Gson.fromJson(Gson.java:1129)
at org.springframework.ide.vscode.commons.languageserver.util.SimpleTextDocumentService.lambda$computeCodeActions$10(SimpleTextDocumentService.java:445)
at java.base/java.util.ArrayList.forEach(Unknown Source)
at org.springframework.ide.vscode.commons.languageserver.util.SimpleTextDocumentService.computeCodeActions(SimpleTextDocumentService.java:442)
at org.springframework.ide.vscode.commons.languageserver.util.SimpleTextDocumentService.lambda$codeAction$11(SimpleTextDocumentService.java:499)
... 5 more
Caused by: java.lang.IllegalStateException: Expected BEGIN_ARRAY but was BEGIN_OBJECT at path $
at com.google.gson.internal.bind.JsonTreeReader.expect(JsonTreeReader.java:163)
at com.google.gson.internal.bind.JsonTreeReader.beginArray(JsonTreeReader.java:72)
at com.google.gson.internal.bind.CollectionTypeAdapterFactory$Adapter.read(CollectionTypeAdapterFactory.java:80)
at com.google.gson.internal.bind.CollectionTypeAdapterFactory$Adapter.read(CollectionTypeAdapterFactory.java:61)
at com.google.gson.Gson.fromJson(Gson.java:1058)
... 10 more
``` | 1.0 | exception thrown in VSCode when using latest snapshots - I am using VSCode with the latest pre-releases from:
- Language Support for Java
- Spring Boot Dashboard
- Spring Boot Tools (from the latest VSIX file)
I have a project open in my workspace, created from initializr (web + actuator), on Spring Boot 2.6.12.
After a little while, I see an error popup showing up, complaining about a problem when asking for `textDocument/codeAction`, and showing this exception in the log output:
```
java.util.concurrent.CompletionException: com.google.gson.JsonSyntaxException: java.lang.IllegalStateException: Expected BEGIN_ARRAY but was BEGIN_OBJECT at path $
at java.base/java.util.concurrent.CompletableFuture.encodeThrowable(Unknown Source)
at java.base/java.util.concurrent.CompletableFuture.completeThrowable(Unknown Source)
at java.base/java.util.concurrent.CompletableFuture$UniApply.tryFire(Unknown Source)
at java.base/java.util.concurrent.CompletableFuture$Completion.run(Unknown Source)
at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source)
at java.base/java.lang.Thread.run(Unknown Source)
Caused by: com.google.gson.JsonSyntaxException: java.lang.IllegalStateException: Expected BEGIN_ARRAY but was BEGIN_OBJECT at path $
at com.google.gson.Gson.fromJson(Gson.java:1070)
at com.google.gson.Gson.fromJson(Gson.java:1129)
at org.springframework.ide.vscode.commons.languageserver.util.SimpleTextDocumentService.lambda$computeCodeActions$10(SimpleTextDocumentService.java:445)
at java.base/java.util.ArrayList.forEach(Unknown Source)
at org.springframework.ide.vscode.commons.languageserver.util.SimpleTextDocumentService.computeCodeActions(SimpleTextDocumentService.java:442)
at org.springframework.ide.vscode.commons.languageserver.util.SimpleTextDocumentService.lambda$codeAction$11(SimpleTextDocumentService.java:499)
... 5 more
Caused by: java.lang.IllegalStateException: Expected BEGIN_ARRAY but was BEGIN_OBJECT at path $
at com.google.gson.internal.bind.JsonTreeReader.expect(JsonTreeReader.java:163)
at com.google.gson.internal.bind.JsonTreeReader.beginArray(JsonTreeReader.java:72)
at com.google.gson.internal.bind.CollectionTypeAdapterFactory$Adapter.read(CollectionTypeAdapterFactory.java:80)
at com.google.gson.internal.bind.CollectionTypeAdapterFactory$Adapter.read(CollectionTypeAdapterFactory.java:61)
at com.google.gson.Gson.fromJson(Gson.java:1058)
... 10 more
``` | architecture | exception thrown in vscode when using latest snapshots i am using vscode with the latest pre releases from language support for java spring boot dashboard spring boot tools from the latest vsix file i have a project open in my workspace created from initializr web actuator on spring boot after a little while i see an error popup showing up complaining about a problem when asking for textdocument codeaction and showing this exception in the log output java util concurrent completionexception com google gson jsonsyntaxexception java lang illegalstateexception expected begin array but was begin object at path at java base java util concurrent completablefuture encodethrowable unknown source at java base java util concurrent completablefuture completethrowable unknown source at java base java util concurrent completablefuture uniapply tryfire unknown source at java base java util concurrent completablefuture completion run unknown source at java base java util concurrent threadpoolexecutor runworker unknown source at java base java util concurrent threadpoolexecutor worker run unknown source at java base java lang thread run unknown source caused by com google gson jsonsyntaxexception java lang illegalstateexception expected begin array but was begin object at path at com google gson gson fromjson gson java at com google gson gson fromjson gson java at org springframework ide vscode commons languageserver util simpletextdocumentservice lambda computecodeactions simpletextdocumentservice java at java base java util arraylist foreach unknown source at org springframework ide vscode commons languageserver util simpletextdocumentservice computecodeactions simpletextdocumentservice java at org springframework ide vscode commons languageserver util simpletextdocumentservice lambda codeaction simpletextdocumentservice java more caused by java lang illegalstateexception expected begin array but was begin object at path at com google gson internal bind jsontreereader expect jsontreereader java at com google gson internal bind jsontreereader beginarray jsontreereader java at com google gson internal bind collectiontypeadapterfactory adapter read collectiontypeadapterfactory java at com google gson internal bind collectiontypeadapterfactory adapter read collectiontypeadapterfactory java at com google gson gson fromjson gson java more | 1 |
3,954 | 10,344,295,967 | IssuesEvent | 2019-09-04 10:52:21 | open-zaak/open-zaak | https://api.github.com/repos/open-zaak/open-zaak | closed | As stakeholder, I want to have the Authorizations API exposed on OpenZaak | EPIC: Architecture | ... so applications can request their permissions (and theoretically, set permissions).
**Description**
The Authorization API was left out of scope in #3. Which might or might not be the best choice, depending on #3. This US makes sure it gets in.
| 1.0 | As stakeholder, I want to have the Authorizations API exposed on OpenZaak - ... so applications can request their permissions (and theoretically, set permissions).
**Description**
The Authorization API was left out of scope in #3. Which might or might not be the best choice, depending on #3. This US makes sure it gets in.
| architecture | as stakeholder i want to have the authorizations api exposed on openzaak so applications can request their permissions and theoretically set permissions description the authorization api was left out of scope in which might or might not be the best choice depending on this us makes sure it gets in | 1 |
7,907 | 19,916,085,929 | IssuesEvent | 2022-01-25 22:53:47 | MicrosoftDocs/architecture-center | https://api.github.com/repos/MicrosoftDocs/architecture-center | closed | this sentence does not make sense... | doc-bug cxp triaged architecture-center/svc reference-architecture/subsvc Pri2 | [Enter feedback here]
From this article: https://docs.microsoft.com/en-us/azure/architecture/reference-architectures/dmz/nva-ha
This sentence does not make sense / is not understandable -
Since HA Ports for inbound traffic every individual TCP/UDP port needs to be opened in a dedicated load-balancing rule.
You may have meant something like
Since HA ports control (or restrict) inbound traffic, every individual TCP/UDP port needs to be opened in a dedicated load-balancing rule.
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 9bab6a90-43a3-3d83-d680-26683a3d833d
* Version Independent ID: f8eea094-d297-93e6-06ac-7652d059734f
* Content: [Deploy highly available NVAs - Azure Architecture Center](https://docs.microsoft.com/en-us/azure/architecture/reference-architectures/dmz/nva-ha)
* Content Source: [docs/reference-architectures/dmz/nva-ha.yml](https://github.com/microsoftdocs/architecture-center/blob/main/docs/reference-architectures/dmz/nva-ha.yml)
* Service: **architecture-center**
* Sub-service: **reference-architecture**
* GitHub Login: @telmosampaio
* Microsoft Alias: **pnp** | 2.0 | this sentence does not make sense... - [Enter feedback here]
From this article: https://docs.microsoft.com/en-us/azure/architecture/reference-architectures/dmz/nva-ha
This sentence does not make sense / is not understandable -
Since HA Ports for inbound traffic every individual TCP/UDP port needs to be opened in a dedicated load-balancing rule.
You may have meant something like
Since HA ports control (or restrict) inbound traffic, every individual TCP/UDP port needs to be opened in a dedicated load-balancing rule.
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 9bab6a90-43a3-3d83-d680-26683a3d833d
* Version Independent ID: f8eea094-d297-93e6-06ac-7652d059734f
* Content: [Deploy highly available NVAs - Azure Architecture Center](https://docs.microsoft.com/en-us/azure/architecture/reference-architectures/dmz/nva-ha)
* Content Source: [docs/reference-architectures/dmz/nva-ha.yml](https://github.com/microsoftdocs/architecture-center/blob/main/docs/reference-architectures/dmz/nva-ha.yml)
* Service: **architecture-center**
* Sub-service: **reference-architecture**
* GitHub Login: @telmosampaio
* Microsoft Alias: **pnp** | architecture | this sentence does not make sense from this article this sentence does not make sense is not understandable since ha ports for inbound traffic every individual tcp udp port needs to be opened in a dedicated load balancing rule you may have meant something like since ha ports control or restrict inbound traffic every individual tcp udp port needs to be opened in a dedicated load balancing rule document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source service architecture center sub service reference architecture github login telmosampaio microsoft alias pnp | 1 |
3,453 | 9,645,281,781 | IssuesEvent | 2019-05-17 08:15:13 | dotnet/docs | https://api.github.com/repos/dotnet/docs | closed | Non working links | :book: guide - .NET Microservices :books: Area - .NET Guide :card_file_box: Technology - .NET Architecture Source - Docs.ms broken-link doc-bug | codebetter.com has been down for some time.
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 84d83855-4370-90e8-024a-12c5ac8220e9
* Version Independent ID: 01be3db1-4a00-c8f2-aa59-59656f206b93
* Content: [Applying CQRS and CQS approaches in a DDD microservice in eShopOnContainers](https://docs.microsoft.com/en-us/dotnet/standard/microservices-architecture/microservice-ddd-cqrs-patterns/eshoponcontainers-cqrs-ddd-microservice#feedback)
* Content Source: [docs/standard/microservices-architecture/microservice-ddd-cqrs-patterns/eshoponcontainers-cqrs-ddd-microservice.md](https://github.com/dotnet/docs/blob/master/docs/standard/microservices-architecture/microservice-ddd-cqrs-patterns/eshoponcontainers-cqrs-ddd-microservice.md)
* Product: **dotnet**
* Technology: **dotnet-ebooks**
* GitHub Login: @CESARDELATORRE
* Microsoft Alias: **wiwagn** | 1.0 | Non working links - codebetter.com has been down for some time.
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 84d83855-4370-90e8-024a-12c5ac8220e9
* Version Independent ID: 01be3db1-4a00-c8f2-aa59-59656f206b93
* Content: [Applying CQRS and CQS approaches in a DDD microservice in eShopOnContainers](https://docs.microsoft.com/en-us/dotnet/standard/microservices-architecture/microservice-ddd-cqrs-patterns/eshoponcontainers-cqrs-ddd-microservice#feedback)
* Content Source: [docs/standard/microservices-architecture/microservice-ddd-cqrs-patterns/eshoponcontainers-cqrs-ddd-microservice.md](https://github.com/dotnet/docs/blob/master/docs/standard/microservices-architecture/microservice-ddd-cqrs-patterns/eshoponcontainers-cqrs-ddd-microservice.md)
* Product: **dotnet**
* Technology: **dotnet-ebooks**
* GitHub Login: @CESARDELATORRE
* Microsoft Alias: **wiwagn** | architecture | non working links codebetter com has been down for some time document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source product dotnet technology dotnet ebooks github login cesardelatorre microsoft alias wiwagn | 1 |
207,397 | 7,127,893,407 | IssuesEvent | 2018-01-21 04:06:00 | OperationCode/operationcode_frontend | https://api.github.com/repos/OperationCode/operationcode_frontend | closed | Add photos/bio to /team page | Priority: Medium Status: In Progress Type: Feature | # Feature
## Why is this feature being added?
We should make it easy to find board members on our `/team` page. Suggest we break down this page, ex. #contributors, #chapter-leaders, #board, #advisors, etc., and include photos so we can recognize them. Here's an example of what I'm envisioning (via GitHub):

*could be round images instead of square*
- [ ] There should be a `+` sign, or another icon so funders and other interested stakeholders can read up on their bios, ex. `operationcode.org/team/hollomancer` and learn more about Conrad, including GitHub, Twitter handle.
- [ ] Needs to be easy to send the link to x foundation or x corporation, and say, "Meet some of our board members and link to the exact spot where board members are listed.
## Technical Requirements
- [ ] Refactor StaffCard to make BoardCard obsolete, such that these components only render what they're provided.
- [ ] Resize images for uniformity.
- [ ] Make image links relative. | 1.0 | Add photos/bio to /team page - # Feature
## Why is this feature being added?
We should make it easy to find board members on our `/team` page. Suggest we break down this page, ex. #contributors, #chapter-leaders, #board, #advisors, etc., and include photos so we can recognize them. Here's an example of what I'm envisioning (via GitHub):

*could be round images instead of square*
- [ ] There should be a `+` sign, or another icon so funders and other interested stakeholders can read up on their bios, ex. `operationcode.org/team/hollomancer` and learn more about Conrad, including GitHub, Twitter handle.
- [ ] Needs to be easy to send the link to x foundation or x corporation, and say, "Meet some of our board members and link to the exact spot where board members are listed.
## Technical Requirements
- [ ] Refactor StaffCard to make BoardCard obsolete, such that these components only render what they're provided.
- [ ] Resize images for uniformity.
- [ ] Make image links relative. | non_architecture | add photos bio to team page feature why is this feature being added we should make it easy to find board members on our team page suggest we break down this page ex contributors chapter leaders board advisors etc and include photos so we can recognize them here s an example of what i m envisioning via github could be round images instead of square there should be a sign or another icon so funders and other interested stakeholders can read up on their bios ex operationcode org team hollomancer and learn more about conrad including github twitter handle needs to be easy to send the link to x foundation or x corporation and say meet some of our board members and link to the exact spot where board members are listed technical requirements refactor staffcard to make boardcard obsolete such that these components only render what they re provided resize images for uniformity make image links relative | 0 |
59,918 | 14,671,978,283 | IssuesEvent | 2020-12-30 09:31:08 | Raku/old-issue-tracker | https://api.github.com/repos/Raku/old-issue-tracker | closed | MoarVM build fail on termux/Android 6.0.1 | build | Migrated from [rt.perl.org#132785](https://rt-archive.perl.org/perl6/Ticket/Display.html?id=132785) (status was 'new')
Searchable as RT132785$
| 1.0 | MoarVM build fail on termux/Android 6.0.1 - Migrated from [rt.perl.org#132785](https://rt-archive.perl.org/perl6/Ticket/Display.html?id=132785) (status was 'new')
Searchable as RT132785$
| non_architecture | moarvm build fail on termux android migrated from status was new searchable as | 0 |
242,407 | 7,841,901,318 | IssuesEvent | 2018-06-18 21:10:00 | StrangeLoopGames/EcoIssues | https://api.github.com/repos/StrangeLoopGames/EcoIssues | closed | Website not loading additional layers (Linux-Server) | Medium Priority | Hi there,
I'm running my Server on Linux and since 0.6.2.5-beta, the map in the World Status page can't display additional layers (like AirPollutionSpread). Tested with Chrome (v 66) and Edge Browser
How to reproduce: Open the website and wait for map loaded -> Click on Select map layer and select something (happens for all layers) -> Map get's a little bit darker but nothing else changes (but error occurs in Chrome Browser console).
Server versions: Mono linux standalone with 0.6.2.5, 0.7.3.3, 0.7.4.2
Edge Browser is not displaying ocean
Log from Chrome console when switching to AirPollutionSpread layer:
```
ecomap.js:419 Starting to parse AirPollutionSpread
ecomap.js:112 Uncaught TypeError: Cannot read property 'concat' of undefined
at lzwDecode (ecomap.js:112)
at parseImg (ecomap.js:311)
at parseBlock (ecomap.js:331)
```
Default view without selected layer:

View with selected layer:

Default view in Microsoft Edge:

| 1.0 | Website not loading additional layers (Linux-Server) - Hi there,
I'm running my Server on Linux and since 0.6.2.5-beta, the map in the World Status page can't display additional layers (like AirPollutionSpread). Tested with Chrome (v 66) and Edge Browser
How to reproduce: Open the website and wait for map loaded -> Click on Select map layer and select something (happens for all layers) -> Map get's a little bit darker but nothing else changes (but error occurs in Chrome Browser console).
Server versions: Mono linux standalone with 0.6.2.5, 0.7.3.3, 0.7.4.2
Edge Browser is not displaying ocean
Log from Chrome console when switching to AirPollutionSpread layer:
```
ecomap.js:419 Starting to parse AirPollutionSpread
ecomap.js:112 Uncaught TypeError: Cannot read property 'concat' of undefined
at lzwDecode (ecomap.js:112)
at parseImg (ecomap.js:311)
at parseBlock (ecomap.js:331)
```
Default view without selected layer:

View with selected layer:

Default view in Microsoft Edge:

| non_architecture | website not loading additional layers linux server hi there i m running my server on linux and since beta the map in the world status page can t display additional layers like airpollutionspread tested with chrome v and edge browser how to reproduce open the website and wait for map loaded click on select map layer and select something happens for all layers map get s a little bit darker but nothing else changes but error occurs in chrome browser console server versions mono linux standalone with edge browser is not displaying ocean log from chrome console when switching to airpollutionspread layer ecomap js starting to parse airpollutionspread ecomap js uncaught typeerror cannot read property concat of undefined at lzwdecode ecomap js at parseimg ecomap js at parseblock ecomap js default view without selected layer view with selected layer default view in microsoft edge | 0 |
577,916 | 17,139,172,640 | IssuesEvent | 2021-07-13 07:40:29 | googleapis/java-bigtable-hbase | https://api.github.com/repos/googleapis/java-bigtable-hbase | opened | bigtable.hbase.TestBufferedMutator: testAutoFlushOff failed | flakybot: issue priority: p1 type: bug | This test failed!
To configure my behavior, see [the Flaky Bot documentation](https://github.com/googleapis/repo-automation-bots/tree/master/packages/flakybot).
If I'm commenting on this issue too often, add the `flakybot: quiet` label and
I will stop commenting.
---
commit: a891335ce3179c45fade4f3683b7e09d38d0107a
buildURL: [Build Status](https://source.cloud.google.com/results/invocations/64905140-8b04-4270-9ed2-a91e73d97e39), [Sponge](http://sponge2/64905140-8b04-4270-9ed2-a91e73d97e39)
status: failed
<details><summary>Test output</summary><br><pre>org.apache.hadoop.hbase.client.RetriesExhaustedWithDetailsException: Failed 1 action: UnauthenticatedException: 1 time, servers with issues: bigtable.googleapis.com
at com.google.cloud.bigtable.hbase.BigtableBufferedMutator.getExceptions(BigtableBufferedMutator.java:188)
at com.google.cloud.bigtable.hbase.BigtableBufferedMutator.handleExceptions(BigtableBufferedMutator.java:142)
at com.google.cloud.bigtable.hbase.BigtableBufferedMutator.flush(BigtableBufferedMutator.java:93)
at com.google.cloud.bigtable.hbase.TestBufferedMutator.testAutoFlushOff(TestBufferedMutator.java:62)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59)
at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56)
at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306)
at org.junit.runners.BlockJUnit4ClassRunner$1.evaluate(BlockJUnit4ClassRunner.java:100)
at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:366)
at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:103)
at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:63)
at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331)
at org.apache.maven.surefire.junitcore.pc.Scheduler$1.run(Scheduler.java:410)
at org.apache.maven.surefire.junitcore.pc.InvokerStrategy.schedule(InvokerStrategy.java:54)
at org.apache.maven.surefire.junitcore.pc.Scheduler.schedule(Scheduler.java:367)
at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329)
at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66)
at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293)
at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306)
at org.junit.runners.ParentRunner.run(ParentRunner.java:413)
at org.junit.runners.Suite.runChild(Suite.java:128)
at org.junit.runners.Suite.runChild(Suite.java:27)
at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331)
at org.apache.maven.surefire.junitcore.pc.Scheduler$1.run(Scheduler.java:410)
at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
at java.util.concurrent.FutureTask.run(FutureTask.java:266)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:748)
Suppressed: com.google.api.gax.batching.BatchingException: Batching finished with 1 batches failed to apply due to: 1 ApiException(1 INTERNAL) and 0 partial failures.
at com.google.api.gax.batching.BatcherStats.asException(BatcherStats.java:147)
at com.google.api.gax.batching.BatcherImpl.close(BatcherImpl.java:290)
at com.google.cloud.bigtable.hbase.wrappers.veneer.BulkMutationVeneerApi.close(BulkMutationVeneerApi.java:68)
at com.google.cloud.bigtable.hbase.BigtableBufferedMutatorHelper.close(BigtableBufferedMutatorHelper.java:91)
at com.google.cloud.bigtable.hbase.BigtableBufferedMutator.close(BigtableBufferedMutator.java:85)
at com.google.cloud.bigtable.hbase.TestBufferedMutator.testAutoFlushOff(TestBufferedMutator.java:64)
... 31 more
</pre></details> | 1.0 | bigtable.hbase.TestBufferedMutator: testAutoFlushOff failed - This test failed!
To configure my behavior, see [the Flaky Bot documentation](https://github.com/googleapis/repo-automation-bots/tree/master/packages/flakybot).
If I'm commenting on this issue too often, add the `flakybot: quiet` label and
I will stop commenting.
---
commit: a891335ce3179c45fade4f3683b7e09d38d0107a
buildURL: [Build Status](https://source.cloud.google.com/results/invocations/64905140-8b04-4270-9ed2-a91e73d97e39), [Sponge](http://sponge2/64905140-8b04-4270-9ed2-a91e73d97e39)
status: failed
<details><summary>Test output</summary><br><pre>org.apache.hadoop.hbase.client.RetriesExhaustedWithDetailsException: Failed 1 action: UnauthenticatedException: 1 time, servers with issues: bigtable.googleapis.com
at com.google.cloud.bigtable.hbase.BigtableBufferedMutator.getExceptions(BigtableBufferedMutator.java:188)
at com.google.cloud.bigtable.hbase.BigtableBufferedMutator.handleExceptions(BigtableBufferedMutator.java:142)
at com.google.cloud.bigtable.hbase.BigtableBufferedMutator.flush(BigtableBufferedMutator.java:93)
at com.google.cloud.bigtable.hbase.TestBufferedMutator.testAutoFlushOff(TestBufferedMutator.java:62)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59)
at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56)
at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306)
at org.junit.runners.BlockJUnit4ClassRunner$1.evaluate(BlockJUnit4ClassRunner.java:100)
at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:366)
at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:103)
at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:63)
at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331)
at org.apache.maven.surefire.junitcore.pc.Scheduler$1.run(Scheduler.java:410)
at org.apache.maven.surefire.junitcore.pc.InvokerStrategy.schedule(InvokerStrategy.java:54)
at org.apache.maven.surefire.junitcore.pc.Scheduler.schedule(Scheduler.java:367)
at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329)
at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66)
at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293)
at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306)
at org.junit.runners.ParentRunner.run(ParentRunner.java:413)
at org.junit.runners.Suite.runChild(Suite.java:128)
at org.junit.runners.Suite.runChild(Suite.java:27)
at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331)
at org.apache.maven.surefire.junitcore.pc.Scheduler$1.run(Scheduler.java:410)
at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
at java.util.concurrent.FutureTask.run(FutureTask.java:266)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:748)
Suppressed: com.google.api.gax.batching.BatchingException: Batching finished with 1 batches failed to apply due to: 1 ApiException(1 INTERNAL) and 0 partial failures.
at com.google.api.gax.batching.BatcherStats.asException(BatcherStats.java:147)
at com.google.api.gax.batching.BatcherImpl.close(BatcherImpl.java:290)
at com.google.cloud.bigtable.hbase.wrappers.veneer.BulkMutationVeneerApi.close(BulkMutationVeneerApi.java:68)
at com.google.cloud.bigtable.hbase.BigtableBufferedMutatorHelper.close(BigtableBufferedMutatorHelper.java:91)
at com.google.cloud.bigtable.hbase.BigtableBufferedMutator.close(BigtableBufferedMutator.java:85)
at com.google.cloud.bigtable.hbase.TestBufferedMutator.testAutoFlushOff(TestBufferedMutator.java:64)
... 31 more
</pre></details> | non_architecture | bigtable hbase testbufferedmutator testautoflushoff failed this test failed to configure my behavior see if i m commenting on this issue too often add the flakybot quiet label and i will stop commenting commit buildurl status failed test output org apache hadoop hbase client retriesexhaustedwithdetailsexception failed action unauthenticatedexception time servers with issues bigtable googleapis com at com google cloud bigtable hbase bigtablebufferedmutator getexceptions bigtablebufferedmutator java at com google cloud bigtable hbase bigtablebufferedmutator handleexceptions bigtablebufferedmutator java at com google cloud bigtable hbase bigtablebufferedmutator flush bigtablebufferedmutator java at com google cloud bigtable hbase testbufferedmutator testautoflushoff testbufferedmutator java at sun reflect nativemethodaccessorimpl native method at sun reflect nativemethodaccessorimpl invoke nativemethodaccessorimpl java at sun reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java lang reflect method invoke method java at org junit runners model frameworkmethod runreflectivecall frameworkmethod java at org junit internal runners model reflectivecallable run reflectivecallable java at org junit runners model frameworkmethod invokeexplosively frameworkmethod java at org junit internal runners statements invokemethod evaluate invokemethod java at org junit runners parentrunner evaluate parentrunner java at org junit runners evaluate java at org junit runners parentrunner runleaf parentrunner java at org junit runners runchild java at org junit runners runchild java at org junit runners parentrunner run parentrunner java at org apache maven surefire junitcore pc scheduler run scheduler java at org apache maven surefire junitcore pc invokerstrategy schedule invokerstrategy java at org apache maven surefire junitcore pc scheduler schedule scheduler java at org junit runners parentrunner runchildren parentrunner java at org junit runners parentrunner access parentrunner java at org junit runners parentrunner evaluate parentrunner java at org junit runners parentrunner evaluate parentrunner java at org junit runners parentrunner run parentrunner java at org junit runners suite runchild suite java at org junit runners suite runchild suite java at org junit runners parentrunner run parentrunner java at org apache maven surefire junitcore pc scheduler run scheduler java at java util concurrent executors runnableadapter call executors java at java util concurrent futuretask run futuretask java at java util concurrent threadpoolexecutor runworker threadpoolexecutor java at java util concurrent threadpoolexecutor worker run threadpoolexecutor java at java lang thread run thread java suppressed com google api gax batching batchingexception batching finished with batches failed to apply due to apiexception internal and partial failures at com google api gax batching batcherstats asexception batcherstats java at com google api gax batching batcherimpl close batcherimpl java at com google cloud bigtable hbase wrappers veneer bulkmutationveneerapi close bulkmutationveneerapi java at com google cloud bigtable hbase bigtablebufferedmutatorhelper close bigtablebufferedmutatorhelper java at com google cloud bigtable hbase bigtablebufferedmutator close bigtablebufferedmutator java at com google cloud bigtable hbase testbufferedmutator testautoflushoff testbufferedmutator java more | 0 |
257,267 | 27,561,833,137 | IssuesEvent | 2023-03-07 22:49:14 | samqws-marketing/coursera_naptime | https://api.github.com/repos/samqws-marketing/coursera_naptime | closed | CVE-2020-36185 (High) detected in multiple libraries - autoclosed | Mend: dependency security vulnerability | ## CVE-2020-36185 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>jackson-databind-2.9.0.jar</b>, <b>jackson-databind-2.8.11.4.jar</b>, <b>jackson-databind-2.3.3.jar</b></p></summary>
<p>
<details><summary><b>jackson-databind-2.9.0.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to vulnerable library: /home/wss-scanner/.ivy2/cache/com.fasterxml.jackson.core/jackson-databind/bundles/jackson-databind-2.9.0.jar</p>
<p>
Dependency Hierarchy:
- play-ehcache_2.12-2.6.25.jar (Root Library)
- play_2.12-2.6.25.jar
- play-json_2.12-2.6.14.jar
- jackson-datatype-jdk8-2.8.11.jar
- :x: **jackson-databind-2.9.0.jar** (Vulnerable Library)
</details>
<details><summary><b>jackson-databind-2.8.11.4.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to vulnerable library: /home/wss-scanner/.ivy2/cache/com.fasterxml.jackson.core/jackson-databind/bundles/jackson-databind-2.8.11.4.jar</p>
<p>
Dependency Hierarchy:
- play-ehcache_2.12-2.6.25.jar (Root Library)
- play_2.12-2.6.25.jar
- :x: **jackson-databind-2.8.11.4.jar** (Vulnerable Library)
</details>
<details><summary><b>jackson-databind-2.3.3.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Path to vulnerable library: /home/wss-scanner/.ivy2/cache/com.fasterxml.jackson.core/jackson-databind/bundles/jackson-databind-2.3.3.jar</p>
<p>
Dependency Hierarchy:
- sbt-plugin-2.4.4.jar (Root Library)
- sbt-js-engine-1.1.3.jar
- npm_2.10-1.1.1.jar
- webjars-locator-0.26.jar
- :x: **jackson-databind-2.3.3.jar** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/samqws-marketing/coursera_naptime/commit/95750513b615ecf0ea9b7e14fb5f71e577d01a1f">95750513b615ecf0ea9b7e14fb5f71e577d01a1f</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
FasterXML jackson-databind 2.x before 2.9.10.8 mishandles the interaction between serialization gadgets and typing, related to org.apache.tomcat.dbcp.dbcp2.datasources.SharedPoolDataSource.
<p>Publish Date: 2021-01-06
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2020-36185>CVE-2020-36185</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2021-01-06</p>
<p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.9.10.8</p>
</p>
</details>
<p></p>
| True | CVE-2020-36185 (High) detected in multiple libraries - autoclosed - ## CVE-2020-36185 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>jackson-databind-2.9.0.jar</b>, <b>jackson-databind-2.8.11.4.jar</b>, <b>jackson-databind-2.3.3.jar</b></p></summary>
<p>
<details><summary><b>jackson-databind-2.9.0.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to vulnerable library: /home/wss-scanner/.ivy2/cache/com.fasterxml.jackson.core/jackson-databind/bundles/jackson-databind-2.9.0.jar</p>
<p>
Dependency Hierarchy:
- play-ehcache_2.12-2.6.25.jar (Root Library)
- play_2.12-2.6.25.jar
- play-json_2.12-2.6.14.jar
- jackson-datatype-jdk8-2.8.11.jar
- :x: **jackson-databind-2.9.0.jar** (Vulnerable Library)
</details>
<details><summary><b>jackson-databind-2.8.11.4.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to vulnerable library: /home/wss-scanner/.ivy2/cache/com.fasterxml.jackson.core/jackson-databind/bundles/jackson-databind-2.8.11.4.jar</p>
<p>
Dependency Hierarchy:
- play-ehcache_2.12-2.6.25.jar (Root Library)
- play_2.12-2.6.25.jar
- :x: **jackson-databind-2.8.11.4.jar** (Vulnerable Library)
</details>
<details><summary><b>jackson-databind-2.3.3.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Path to vulnerable library: /home/wss-scanner/.ivy2/cache/com.fasterxml.jackson.core/jackson-databind/bundles/jackson-databind-2.3.3.jar</p>
<p>
Dependency Hierarchy:
- sbt-plugin-2.4.4.jar (Root Library)
- sbt-js-engine-1.1.3.jar
- npm_2.10-1.1.1.jar
- webjars-locator-0.26.jar
- :x: **jackson-databind-2.3.3.jar** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/samqws-marketing/coursera_naptime/commit/95750513b615ecf0ea9b7e14fb5f71e577d01a1f">95750513b615ecf0ea9b7e14fb5f71e577d01a1f</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
FasterXML jackson-databind 2.x before 2.9.10.8 mishandles the interaction between serialization gadgets and typing, related to org.apache.tomcat.dbcp.dbcp2.datasources.SharedPoolDataSource.
<p>Publish Date: 2021-01-06
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2020-36185>CVE-2020-36185</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2021-01-06</p>
<p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.9.10.8</p>
</p>
</details>
<p></p>
| non_architecture | cve high detected in multiple libraries autoclosed cve high severity vulnerability vulnerable libraries jackson databind jar jackson databind jar jackson databind jar jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to vulnerable library home wss scanner cache com fasterxml jackson core jackson databind bundles jackson databind jar dependency hierarchy play ehcache jar root library play jar play json jar jackson datatype jar x jackson databind jar vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to vulnerable library home wss scanner cache com fasterxml jackson core jackson databind bundles jackson databind jar dependency hierarchy play ehcache jar root library play jar x jackson databind jar vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api path to vulnerable library home wss scanner cache com fasterxml jackson core jackson databind bundles jackson databind jar dependency hierarchy sbt plugin jar root library sbt js engine jar npm jar webjars locator jar x jackson databind jar vulnerable library found in head commit a href found in base branch master vulnerability details fasterxml jackson databind x before mishandles the interaction between serialization gadgets and typing related to org apache tomcat dbcp datasources sharedpooldatasource publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version release date fix resolution com fasterxml jackson core jackson databind | 0 |
13,152 | 9,888,310,810 | IssuesEvent | 2019-06-25 11:14:46 | wellcometrust/platform | https://api.github.com/repos/wellcometrust/platform | closed | Distinguish between “could not parse bag-info.txt” and “bag-info.txt does not exist” | 📦 Storage service | Follow-up from https://github.com/wellcometrust/storage-service/pull/227 | 1.0 | Distinguish between “could not parse bag-info.txt” and “bag-info.txt does not exist” - Follow-up from https://github.com/wellcometrust/storage-service/pull/227 | non_architecture | distinguish between “could not parse bag info txt” and “bag info txt does not exist” follow up from | 0 |
10,332 | 26,785,479,868 | IssuesEvent | 2023-02-01 02:09:13 | facebook/react-native | https://api.github.com/repos/facebook/react-native | closed | Error "Can't find variable: require" on iOS (RN v.0.68.2) | Stale Needs: Triage :mag: Type: Old Architecture | ### Description
Error _"Can't find variable: require"_ on 0.68.2 on Debug & Release.
Enumeration of presets and babel plugins did not solve the problem. The cache was cleared after each attempt. On the forums of other libraries, the developers dealt with a similar error with a patch in the new version.
The bug was discovered after fixing [another bug](https://github.com/facebook/react-native/issues/33954#issuecomment-1146979111) 0.68.2
Cache was cleared before each launch: react-native start --reset-cache
The babel config was tested from empty to what worked in previous versions and suggested in other forums and repositories ([1](https://github.com/facebook/react-native/issues/21048#issuecomment-426987192), [2](https://github.com/Tecode/react-native-mobx/blob/master/babel.config.js), [3](https://github.com/facebook/react-native/issues/22321)).
### Version
0.68.2
### Output of `npx react-native info`
System:
OS: macOS 11.6
CPU: (6) x64 Intel(R) Core(TM) i5-8500B CPU @ 3.00GHz
Memory: 76.89 MB / 8.00 GB
Shell: 5.8 - /bin/zsh
Binaries:
Node: 16.15.0 - /usr/local/bin/node
Yarn: 1.22.11 - /usr/local/bin/yarn
npm: 8.5.5 - /usr/local/bin/npm
Watchman: 2022.05.30.00 - /usr/local/bin/watchman
Managers:
CocoaPods: 1.11.3 - /usr/local/bin/pod
SDKs:
iOS SDK:
Platforms: DriverKit 21.2, iOS 15.2, macOS 12.1, tvOS 15.2, watchOS 8.3
Android SDK:
API Levels: 28, 29, 30, 31
Build Tools: 29.0.2, 30.0.2, 31.0.0, 32.0.0
System Images: android-30 | Google APIs Intel x86 Atom, android-30 | Google Play Intel x86 Atom, android-31 | Google APIs Intel x86 Atom_64
Android NDK: Not Found
IDEs:
Android Studio: 2020.3 AI-203.7717.56.2031.7678000
Xcode: 13.2.1/13C100 - /usr/bin/xcodebuild
Languages:
Java: 1.8.0_292 - /usr/bin/javac
npmPackages:
@react-native-community/cli: 7.0.3 => 7.0.3
react: ^18.1.0 => 18.1.0
react-native: ^0.68.2 => 0.68.2
react-native-macos: Not Found
npmGlobalPackages:
*react-native*: Not Found
### Steps to reproduce
Any Babel configuration. Tested on iOS 15.2 Simulator.
### Snack, code example, screenshot, or link to a repository

| 1.0 | Error "Can't find variable: require" on iOS (RN v.0.68.2) - ### Description
Error _"Can't find variable: require"_ on 0.68.2 on Debug & Release.
Enumeration of presets and babel plugins did not solve the problem. The cache was cleared after each attempt. On the forums of other libraries, the developers dealt with a similar error with a patch in the new version.
The bug was discovered after fixing [another bug](https://github.com/facebook/react-native/issues/33954#issuecomment-1146979111) 0.68.2
Cache was cleared before each launch: react-native start --reset-cache
The babel config was tested from empty to what worked in previous versions and suggested in other forums and repositories ([1](https://github.com/facebook/react-native/issues/21048#issuecomment-426987192), [2](https://github.com/Tecode/react-native-mobx/blob/master/babel.config.js), [3](https://github.com/facebook/react-native/issues/22321)).
### Version
0.68.2
### Output of `npx react-native info`
System:
OS: macOS 11.6
CPU: (6) x64 Intel(R) Core(TM) i5-8500B CPU @ 3.00GHz
Memory: 76.89 MB / 8.00 GB
Shell: 5.8 - /bin/zsh
Binaries:
Node: 16.15.0 - /usr/local/bin/node
Yarn: 1.22.11 - /usr/local/bin/yarn
npm: 8.5.5 - /usr/local/bin/npm
Watchman: 2022.05.30.00 - /usr/local/bin/watchman
Managers:
CocoaPods: 1.11.3 - /usr/local/bin/pod
SDKs:
iOS SDK:
Platforms: DriverKit 21.2, iOS 15.2, macOS 12.1, tvOS 15.2, watchOS 8.3
Android SDK:
API Levels: 28, 29, 30, 31
Build Tools: 29.0.2, 30.0.2, 31.0.0, 32.0.0
System Images: android-30 | Google APIs Intel x86 Atom, android-30 | Google Play Intel x86 Atom, android-31 | Google APIs Intel x86 Atom_64
Android NDK: Not Found
IDEs:
Android Studio: 2020.3 AI-203.7717.56.2031.7678000
Xcode: 13.2.1/13C100 - /usr/bin/xcodebuild
Languages:
Java: 1.8.0_292 - /usr/bin/javac
npmPackages:
@react-native-community/cli: 7.0.3 => 7.0.3
react: ^18.1.0 => 18.1.0
react-native: ^0.68.2 => 0.68.2
react-native-macos: Not Found
npmGlobalPackages:
*react-native*: Not Found
### Steps to reproduce
Any Babel configuration. Tested on iOS 15.2 Simulator.
### Snack, code example, screenshot, or link to a repository

| architecture | error can t find variable require on ios rn v description error can t find variable require on on debug release enumeration of presets and babel plugins did not solve the problem the cache was cleared after each attempt on the forums of other libraries the developers dealt with a similar error with a patch in the new version the bug was discovered after fixing cache was cleared before each launch react native start reset cache the babel config was tested from empty to what worked in previous versions and suggested in other forums and repositories version output of npx react native info system os macos cpu intel r core tm cpu memory mb gb shell bin zsh binaries node usr local bin node yarn usr local bin yarn npm usr local bin npm watchman usr local bin watchman managers cocoapods usr local bin pod sdks ios sdk platforms driverkit ios macos tvos watchos android sdk api levels build tools system images android google apis intel atom android google play intel atom android google apis intel atom android ndk not found ides android studio ai xcode usr bin xcodebuild languages java usr bin javac npmpackages react native community cli react react native react native macos not found npmglobalpackages react native not found steps to reproduce any babel configuration tested on ios simulator snack code example screenshot or link to a repository | 1 |
6,553 | 14,877,114,902 | IssuesEvent | 2021-01-20 02:23:19 | Azure/azure-sdk | https://api.github.com/repos/Azure/azure-sdk | opened | Board Review: Azure Mixed Reality Authentication client library | architecture board-review | Thank you for submitting this review request. Thorough review of your client library ensures that your APIs are consistent with the guidelines and the consumers of your client library have a consistently good experience when using Azure.
**The Architecture Board reviews [Track 2 libraries](https://azure.github.io/azure-sdk/general_introduction.html) only.** If your library does not meet this requirement, please reach out to [Architecture Board]([email protected]) before creating the issue.
Please reference our [review process guidelines](https://azure.github.io/azure-sdk/policies_reviewprocess.html) to understand what is being asked for in the issue template.
To ensure consistency, all Tier-1 languages (C#, TypeScript, Java, Python) will generally be reviewed together. In expansive libraries, we will pair dynamic languages (Python, TypeScript) together, and strongly typed languages (C#, Java) together in separate meetings.
For Tier-2 languages (C, C++, Go, Android, iOS), the review will be on an as-needed basis.
**Before submitting, ensure you adjust the title of the issue appropriately.**
**Note that the required material must be included before a meeting can be scheduled.**
## Contacts and Timeline
* Responsible service team: [email protected]
* Main contacts: crtreasu, virivera, rgarcia, ariye
* Expected code complete date: 01/29
* Expected release date: 03/01/2021
## About the Service
* Link to documentation introducing/describing the service: https://review.docs.microsoft.com/en-us/azure/object-anchors/overview?branch=release-preview-aou
* Link to the service REST APIs: https://github.com/Azure/azure-rest-api-specs/tree/master/specification/mixedreality/data-plane/Microsoft.MixedReality/preview/2019-02-28-preview
* Link to GitHub issue for previous review sessions, if applicable: https://github.com/Azure/azure-sdk/issues/2005
## About the client library
* Name of the client library: Azure Mixed Reality Authentication
* Languages for this review: JavaScript, Java, Python
## Artifacts required (per language)
Please read through “API Review” section [here](https://azure.github.io/azure-sdk/policies_reviewprocess.html) to understand how these artifacts are generated. **It is critical that these artifacts are present and are in the right format. If not, the language architects cannot review them with the SDK Team’s API review tool.**
### .NET
n/a already completed
* APIView Link:
* Link to Champion Scenarios/Quickstart samples:
### Java
* APIView Link: https://apiview.dev/Assemblies/Review/f3a4bb684ffd4badada05eff7de952d5
* Link to Champion Scenarios/Quickstart samples:
### Python
* APIView Link:
* Link to Champion Scenarios/Quickstart samples:
### TypeScript
* APIView Link: https://apiview.dev/Assemblies/Review/4917626415bc448c8e2534e00c6f3a17
* Link to Champion Scenarios/Quickstart samples:
For all other languages, send a request to the Architecture Board to discuss the best format on individual basis.
## Thank you!
| 1.0 | Board Review: Azure Mixed Reality Authentication client library - Thank you for submitting this review request. Thorough review of your client library ensures that your APIs are consistent with the guidelines and the consumers of your client library have a consistently good experience when using Azure.
**The Architecture Board reviews [Track 2 libraries](https://azure.github.io/azure-sdk/general_introduction.html) only.** If your library does not meet this requirement, please reach out to [Architecture Board]([email protected]) before creating the issue.
Please reference our [review process guidelines](https://azure.github.io/azure-sdk/policies_reviewprocess.html) to understand what is being asked for in the issue template.
To ensure consistency, all Tier-1 languages (C#, TypeScript, Java, Python) will generally be reviewed together. In expansive libraries, we will pair dynamic languages (Python, TypeScript) together, and strongly typed languages (C#, Java) together in separate meetings.
For Tier-2 languages (C, C++, Go, Android, iOS), the review will be on an as-needed basis.
**Before submitting, ensure you adjust the title of the issue appropriately.**
**Note that the required material must be included before a meeting can be scheduled.**
## Contacts and Timeline
* Responsible service team: [email protected]
* Main contacts: crtreasu, virivera, rgarcia, ariye
* Expected code complete date: 01/29
* Expected release date: 03/01/2021
## About the Service
* Link to documentation introducing/describing the service: https://review.docs.microsoft.com/en-us/azure/object-anchors/overview?branch=release-preview-aou
* Link to the service REST APIs: https://github.com/Azure/azure-rest-api-specs/tree/master/specification/mixedreality/data-plane/Microsoft.MixedReality/preview/2019-02-28-preview
* Link to GitHub issue for previous review sessions, if applicable: https://github.com/Azure/azure-sdk/issues/2005
## About the client library
* Name of the client library: Azure Mixed Reality Authentication
* Languages for this review: JavaScript, Java, Python
## Artifacts required (per language)
Please read through “API Review” section [here](https://azure.github.io/azure-sdk/policies_reviewprocess.html) to understand how these artifacts are generated. **It is critical that these artifacts are present and are in the right format. If not, the language architects cannot review them with the SDK Team’s API review tool.**
### .NET
n/a already completed
* APIView Link:
* Link to Champion Scenarios/Quickstart samples:
### Java
* APIView Link: https://apiview.dev/Assemblies/Review/f3a4bb684ffd4badada05eff7de952d5
* Link to Champion Scenarios/Quickstart samples:
### Python
* APIView Link:
* Link to Champion Scenarios/Quickstart samples:
### TypeScript
* APIView Link: https://apiview.dev/Assemblies/Review/4917626415bc448c8e2534e00c6f3a17
* Link to Champion Scenarios/Quickstart samples:
For all other languages, send a request to the Architecture Board to discuss the best format on individual basis.
## Thank you!
| architecture | board review azure mixed reality authentication client library thank you for submitting this review request thorough review of your client library ensures that your apis are consistent with the guidelines and the consumers of your client library have a consistently good experience when using azure the architecture board reviews only if your library does not meet this requirement please reach out to adparch microsoft com before creating the issue please reference our to understand what is being asked for in the issue template to ensure consistency all tier languages c typescript java python will generally be reviewed together in expansive libraries we will pair dynamic languages python typescript together and strongly typed languages c java together in separate meetings for tier languages c c go android ios the review will be on an as needed basis before submitting ensure you adjust the title of the issue appropriately note that the required material must be included before a meeting can be scheduled contacts and timeline responsible service team ou servicesdkdev microsoft com main contacts crtreasu virivera rgarcia ariye expected code complete date expected release date about the service link to documentation introducing describing the service link to the service rest apis link to github issue for previous review sessions if applicable about the client library name of the client library azure mixed reality authentication languages for this review javascript java python artifacts required per language please read through “api review” section to understand how these artifacts are generated it is critical that these artifacts are present and are in the right format if not the language architects cannot review them with the sdk team’s api review tool net n a already completed apiview link link to champion scenarios quickstart samples java apiview link link to champion scenarios quickstart samples python apiview link link to champion scenarios quickstart samples typescript apiview link link to champion scenarios quickstart samples for all other languages send a request to the architecture board to discuss the best format on individual basis thank you | 1 |
4,928 | 11,851,414,736 | IssuesEvent | 2020-03-24 18:05:01 | kubernetes/kubernetes | https://api.github.com/repos/kubernetes/kubernetes | closed | Research for creating a readCoreV1NamespacedPodStatus conformant test | area/conformance sig/architecture sig/testing | # Description
After looking through endpoints which need tests written for in APIsnoop, this appears to be an endpoint which isn't hit.
#
## Fetch dependencies
```shell
go get -v -u k8s.io/apimachinery/pkg/apis/meta/v1
go get -v -u k8s.io/client-go/kubernetes
go get -v -u k8s.io/client-go/tools/clientcmd
go get -v -u github.com/ghodss/yaml
```
## Test draft
```go
package main
import (
"fmt"
"flag"
"time"
"os"
"k8s.io/api/core/v1"
metav1 "k8s.io/apimachinery/pkg/apis/meta/v1"
"k8s.io/client-go/kubernetes"
"k8s.io/client-go/tools/clientcmd"
"github.com/ghodss/yaml"
)
func main() {
// uses the current context in kubeconfig
kubeconfig := flag.String("kubeconfig",
fmt.Sprintf("%v/%v/%v", os.Getenv("HOME"), ".kube", "config"),
"(optional) absolute path to the kubeconfig file")
flag.Parse()
config, err := clientcmd.BuildConfigFromFlags("", *kubeconfig)
if err != nil {
fmt.Println(err)
}
// make our work easier to find in the audit_event queries
config.UserAgent = "live-test-writing"
// creates the clientset
clientset, _ := kubernetes.NewForConfig(config)
// access the API to list pods
_, err = clientset.CoreV1().Pods("default").Create(&v1.Pod{
ObjectMeta: metav1.ObjectMeta{
Name: "nginx",
Labels: map[string]string{
"pod-name": "nginx",
},
},
Spec: v1.PodSpec{
Containers: []v1.Container{{
Image: "nginx",
Name: "nginx",
}},
RestartPolicy: v1.RestartPolicyNever,
},
})
if err != nil {
fmt.Println(err)
return
}
time.Sleep(5 * time.Second)
pod, err := clientset.CoreV1().Pods("default").Get("nginx", metav1.GetOptions{})
if err != nil {
fmt.Println(err)
return
}
podYAML, err := yaml.Marshal(pod)
if err != nil {
fmt.Printf("err: %v\n", err)
return
}
fmt.Println(string(podYAML))
time.Sleep(5 * time.Second)
err = clientset.CoreV1().Pods("default").Delete("nginx", &metav1.DeleteOptions{})
if err != nil {
fmt.Println(err)
return
}
}
```
## Test draft- Python working implementation
```python
pip3 install kubernetes
from __future__ import print_function
import time
import kubernetes.client
from kubernetes.client.rest import ApiException
from pprint import pprint
from kubernetes.client.configuration import Configuration
from kubernetes.config import kube_config
configuration = Configuration()
configuration.host = None
kube_config.load_kube_config(client_configuration=configuration)
# Uncomment below to setup prefix (e.g. Bearer) for API key, if needed
# configuration.api_key_prefix['authorization'] = 'Bearer'
# create an instance of the API class
api_instance = kubernetes.client.CoreV1Api(kubernetes.client.ApiClient(configuration))
name = 'kindnet-c7vtg' # str | name of the Pod
namespace = 'kube-system' # str | object name and auth scope, such as for teams and projects
pretty = 'pretty_example' # str | If 'true', then the output is pretty printed. (optional)
try:
api_response = api_instance.read_namespaced_pod_status(name, namespace, pretty=pretty)
pprint(api_response)
except ApiException as e:
print("Exception when calling CoreV1Api->read_namespaced_pod_status: %s\n" % e)
```
# Verify with APISnoop<a id="sec-4"></a>
## create view for hit endpoints
```sql-mode
CREATE VIEW "public"."endpoints_hit_by_new_test" AS
WITH live_testing_endpoints AS (
SELECT DISTINCT
operation_id,
count(1) as hits
FROM
audit_event
WHERE bucket = 'apisnoop'
AND useragent = 'live-test-writing'
GROUP BY operation_id
), baseline AS (
SELECT DISTINCT
operation_id,
test_hits,
conf_hits
FROM endpoint_coverage where bucket != 'apisnoop'
)
SELECT DISTINCT
lte.operation_id,
b.test_hits as hit_by_ete,
lte.hits as hit_by_new_test
FROM live_testing_endpoints lte
JOIN baseline b ON (b.operation_id = lte.operation_id);
```
## create view for coverage changed
```sql-mode
CREATE OR REPLACE VIEW "public"."projected_change_in_coverage" AS
with baseline as (
SELECT *
FROM
stable_endpoint_stats
WHERE job != 'live'
), test as (
SELECT
count(1) as endpoints_hit
FROM
(
select
operation_id
FROM audit_event
WHERE useragent = 'live-test-writing'
EXCEPT
SELECT
operation_id
FROM
endpoint_coverage
WHERE test_hits > 0
) tested_endpoints
), coverage as (
SELECT
baseline.test_hits as old_coverage,
(baseline.test_hits::int + test.endpoints_hit::int ) as new_coverage
from baseline, test
)
select
'test_coverage' as category,
baseline.total_endpoints,
coverage.old_coverage,
coverage.new_coverage,
(coverage.new_coverage - coverage.old_coverage) as change_in_number
from baseline, coverage
;
```
## find endpoints hit by this test
```sql-mode
select * from endpoints_hit_by_new_test;
```
```
operation_id | hit_by_ete | hit_by_new_test
---------------------------+------------+-----------------
createCoreV1NamespacedPod | 1990 | 2
deleteCoreV1NamespacedPod | 2114 | 2
readCoreV1NamespacedPod | 11421 | 1
(3 rows)
```
## show the change in coverage
```sql-mode
select * from projected_change_in_coverage;
```
```
category | total_endpoints | old_coverage | new_coverage | change_in_number
---------------+-----------------+--------------+--------------+------------------
test_coverage | 430 | 167 | 167 | 0
(1 row)
```
# Final notes
From the endpoints hit report above, it doesn't appear that my draft test hit the target endpoint.
Would it be possible for some help and/or advice on hitting the `/api/v1/namespaces/NAMESPACE/pods/PODNAME/status` endpoint? | 1.0 | Research for creating a readCoreV1NamespacedPodStatus conformant test - # Description
After looking through endpoints which need tests written for in APIsnoop, this appears to be an endpoint which isn't hit.
#
## Fetch dependencies
```shell
go get -v -u k8s.io/apimachinery/pkg/apis/meta/v1
go get -v -u k8s.io/client-go/kubernetes
go get -v -u k8s.io/client-go/tools/clientcmd
go get -v -u github.com/ghodss/yaml
```
## Test draft
```go
package main
import (
"fmt"
"flag"
"time"
"os"
"k8s.io/api/core/v1"
metav1 "k8s.io/apimachinery/pkg/apis/meta/v1"
"k8s.io/client-go/kubernetes"
"k8s.io/client-go/tools/clientcmd"
"github.com/ghodss/yaml"
)
func main() {
// uses the current context in kubeconfig
kubeconfig := flag.String("kubeconfig",
fmt.Sprintf("%v/%v/%v", os.Getenv("HOME"), ".kube", "config"),
"(optional) absolute path to the kubeconfig file")
flag.Parse()
config, err := clientcmd.BuildConfigFromFlags("", *kubeconfig)
if err != nil {
fmt.Println(err)
}
// make our work easier to find in the audit_event queries
config.UserAgent = "live-test-writing"
// creates the clientset
clientset, _ := kubernetes.NewForConfig(config)
// access the API to list pods
_, err = clientset.CoreV1().Pods("default").Create(&v1.Pod{
ObjectMeta: metav1.ObjectMeta{
Name: "nginx",
Labels: map[string]string{
"pod-name": "nginx",
},
},
Spec: v1.PodSpec{
Containers: []v1.Container{{
Image: "nginx",
Name: "nginx",
}},
RestartPolicy: v1.RestartPolicyNever,
},
})
if err != nil {
fmt.Println(err)
return
}
time.Sleep(5 * time.Second)
pod, err := clientset.CoreV1().Pods("default").Get("nginx", metav1.GetOptions{})
if err != nil {
fmt.Println(err)
return
}
podYAML, err := yaml.Marshal(pod)
if err != nil {
fmt.Printf("err: %v\n", err)
return
}
fmt.Println(string(podYAML))
time.Sleep(5 * time.Second)
err = clientset.CoreV1().Pods("default").Delete("nginx", &metav1.DeleteOptions{})
if err != nil {
fmt.Println(err)
return
}
}
```
## Test draft- Python working implementation
```python
pip3 install kubernetes
from __future__ import print_function
import time
import kubernetes.client
from kubernetes.client.rest import ApiException
from pprint import pprint
from kubernetes.client.configuration import Configuration
from kubernetes.config import kube_config
configuration = Configuration()
configuration.host = None
kube_config.load_kube_config(client_configuration=configuration)
# Uncomment below to setup prefix (e.g. Bearer) for API key, if needed
# configuration.api_key_prefix['authorization'] = 'Bearer'
# create an instance of the API class
api_instance = kubernetes.client.CoreV1Api(kubernetes.client.ApiClient(configuration))
name = 'kindnet-c7vtg' # str | name of the Pod
namespace = 'kube-system' # str | object name and auth scope, such as for teams and projects
pretty = 'pretty_example' # str | If 'true', then the output is pretty printed. (optional)
try:
api_response = api_instance.read_namespaced_pod_status(name, namespace, pretty=pretty)
pprint(api_response)
except ApiException as e:
print("Exception when calling CoreV1Api->read_namespaced_pod_status: %s\n" % e)
```
# Verify with APISnoop<a id="sec-4"></a>
## create view for hit endpoints
```sql-mode
CREATE VIEW "public"."endpoints_hit_by_new_test" AS
WITH live_testing_endpoints AS (
SELECT DISTINCT
operation_id,
count(1) as hits
FROM
audit_event
WHERE bucket = 'apisnoop'
AND useragent = 'live-test-writing'
GROUP BY operation_id
), baseline AS (
SELECT DISTINCT
operation_id,
test_hits,
conf_hits
FROM endpoint_coverage where bucket != 'apisnoop'
)
SELECT DISTINCT
lte.operation_id,
b.test_hits as hit_by_ete,
lte.hits as hit_by_new_test
FROM live_testing_endpoints lte
JOIN baseline b ON (b.operation_id = lte.operation_id);
```
## create view for coverage changed
```sql-mode
CREATE OR REPLACE VIEW "public"."projected_change_in_coverage" AS
with baseline as (
SELECT *
FROM
stable_endpoint_stats
WHERE job != 'live'
), test as (
SELECT
count(1) as endpoints_hit
FROM
(
select
operation_id
FROM audit_event
WHERE useragent = 'live-test-writing'
EXCEPT
SELECT
operation_id
FROM
endpoint_coverage
WHERE test_hits > 0
) tested_endpoints
), coverage as (
SELECT
baseline.test_hits as old_coverage,
(baseline.test_hits::int + test.endpoints_hit::int ) as new_coverage
from baseline, test
)
select
'test_coverage' as category,
baseline.total_endpoints,
coverage.old_coverage,
coverage.new_coverage,
(coverage.new_coverage - coverage.old_coverage) as change_in_number
from baseline, coverage
;
```
## find endpoints hit by this test
```sql-mode
select * from endpoints_hit_by_new_test;
```
```
operation_id | hit_by_ete | hit_by_new_test
---------------------------+------------+-----------------
createCoreV1NamespacedPod | 1990 | 2
deleteCoreV1NamespacedPod | 2114 | 2
readCoreV1NamespacedPod | 11421 | 1
(3 rows)
```
## show the change in coverage
```sql-mode
select * from projected_change_in_coverage;
```
```
category | total_endpoints | old_coverage | new_coverage | change_in_number
---------------+-----------------+--------------+--------------+------------------
test_coverage | 430 | 167 | 167 | 0
(1 row)
```
# Final notes
From the endpoints hit report above, it doesn't appear that my draft test hit the target endpoint.
Would it be possible for some help and/or advice on hitting the `/api/v1/namespaces/NAMESPACE/pods/PODNAME/status` endpoint? | architecture | research for creating a conformant test description after looking through endpoints which need tests written for in apisnoop this appears to be an endpoint which isn t hit fetch dependencies shell go get v u io apimachinery pkg apis meta go get v u io client go kubernetes go get v u io client go tools clientcmd go get v u github com ghodss yaml test draft go package main import fmt flag time os io api core io apimachinery pkg apis meta io client go kubernetes io client go tools clientcmd github com ghodss yaml func main uses the current context in kubeconfig kubeconfig flag string kubeconfig fmt sprintf v v v os getenv home kube config optional absolute path to the kubeconfig file flag parse config err clientcmd buildconfigfromflags kubeconfig if err nil fmt println err make our work easier to find in the audit event queries config useragent live test writing creates the clientset clientset kubernetes newforconfig config access the api to list pods err clientset pods default create pod objectmeta objectmeta name nginx labels map string pod name nginx spec podspec containers container image nginx name nginx restartpolicy restartpolicynever if err nil fmt println err return time sleep time second pod err clientset pods default get nginx getoptions if err nil fmt println err return podyaml err yaml marshal pod if err nil fmt printf err v n err return fmt println string podyaml time sleep time second err clientset pods default delete nginx deleteoptions if err nil fmt println err return test draft python working implementation python install kubernetes from future import print function import time import kubernetes client from kubernetes client rest import apiexception from pprint import pprint from kubernetes client configuration import configuration from kubernetes config import kube config configuration configuration configuration host none kube config load kube config client configuration configuration uncomment below to setup prefix e g bearer for api key if needed configuration api key prefix bearer create an instance of the api class api instance kubernetes client kubernetes client apiclient configuration name kindnet str name of the pod namespace kube system str object name and auth scope such as for teams and projects pretty pretty example str if true then the output is pretty printed optional try api response api instance read namespaced pod status name namespace pretty pretty pprint api response except apiexception as e print exception when calling read namespaced pod status s n e verify with apisnoop create view for hit endpoints sql mode create view public endpoints hit by new test as with live testing endpoints as select distinct operation id count as hits from audit event where bucket apisnoop and useragent live test writing group by operation id baseline as select distinct operation id test hits conf hits from endpoint coverage where bucket apisnoop select distinct lte operation id b test hits as hit by ete lte hits as hit by new test from live testing endpoints lte join baseline b on b operation id lte operation id create view for coverage changed sql mode create or replace view public projected change in coverage as with baseline as select from stable endpoint stats where job live test as select count as endpoints hit from select operation id from audit event where useragent live test writing except select operation id from endpoint coverage where test hits tested endpoints coverage as select baseline test hits as old coverage baseline test hits int test endpoints hit int as new coverage from baseline test select test coverage as category baseline total endpoints coverage old coverage coverage new coverage coverage new coverage coverage old coverage as change in number from baseline coverage find endpoints hit by this test sql mode select from endpoints hit by new test operation id hit by ete hit by new test rows show the change in coverage sql mode select from projected change in coverage category total endpoints old coverage new coverage change in number test coverage row final notes from the endpoints hit report above it doesn t appear that my draft test hit the target endpoint would it be possible for some help and or advice on hitting the api namespaces namespace pods podname status endpoint | 1 |
8,702 | 23,287,981,996 | IssuesEvent | 2022-08-05 18:45:42 | Azure/azure-sdk | https://api.github.com/repos/Azure/azure-sdk | closed | Board Review: metrics advisor (Python & .net) | architecture board-review | ## The Basics
* Service team responsible for the client library: Metrics Advisor
* Link to documentation describing the service: https://docs.microsoft.com/en-us/azure/cognitive-services/metrics-advisor/
* Contact email (if service team, provide PM and Dev Lead):
[email protected], [email protected] (dev)
[email protected] (PM)
## About this client library
* Name of the client library: azure-ai-metricsadvisor
* Languages for this review: .net/Python
* Link to the service REST APIs:
https://github.com/bowgong/azure-rest-api-specs/blob/metricsadvisor-preview/specification/cognitiveservices/data-plane/MetricsAdvisor/preview/v1.0/MetricsAdvisor.json
https://westus2.dev.cognitive.microsoft.com/docs/services/MetricsAdvisor/operations/createDataFeed
## Artifacts required (per language)
We use an API review tool ([apiview](https://apiview.azurewebsites.net)) to support .NET and Java API reviews. For Python and TypeScript, use the API extractor tool, then submit the output as a Draft PR to the relevant repository (azure-sdk-for-python or azure-sdk-for-js).
### .NET
* [APIView](https://apiview.dev/Assemblies/Review/8caf3dd1661c45228d8081a536cca3bc)
* Link to samples for champion scenarios: https://github.com/Azure/azure-sdk-for-net/blob/master/sdk/metricsadvisor/Azure.AI.MetricsAdvisor/samples/README.md
### Python
* [APIView](https://apiview.dev/Assemblies/Review/6acc354ec6c5421b82a081a07b481df0)
* Link to samples for champion scenarios: https://github.com/Azure/azure-sdk-for-python/tree/master/sdk/metricsadvisor/azure-ai-metricsadvisor/samples
## Champion Scenarios
A champion scenario is a use case that the consumer of the client library is commonly expected to perform. Champion scenarios are used to ensure the developer experience is exemplary for the common cases. You need to show the entire code sample (including error handling, as an example) for the champion scenarios.
* Champion Scenario 1:
* Describe the champion scenario: create a data feed to ingest data
* Estimate the percentage of developers using the service who would use the champion scenario
* Link to the code samples: [Python](https://gist.github.com/xiangyan99/6c367d45168294b043b2f5685d57f584#create-data-feed) | [.NET](https://gist.github.com/kinelski/c072790394398c37d186df611f6bea44#scenario-1-datafeed-creation)
* Champion Scenario 2:
* Describe the champion scenario: create a configuration to let service know whether a point is anomaly
* Estimate the percentage of developers using the service who would use the champion scenario
* Link to the code samples: [Python](https://gist.github.com/xiangyan99/6c367d45168294b043b2f5685d57f584#create-detection-configuration) | [.NET](https://gist.github.com/kinelski/c072790394398c37d186df611f6bea44#2d-applyingtuning-anomaly-detection)
* Champion Scenario 3:
* Describe the champion scenario: configure the service when to trigger an alert
* Estimate the percentage of developers using the service who would use the champion scenario
* Link to the code samples: [Python](https://gist.github.com/xiangyan99/6c367d45168294b043b2f5685d57f584#config-alert-configuration) | [.NET](https://gist.github.com/kinelski/c072790394398c37d186df611f6bea44#scenario-3-configure-alerts-and-get-incidents-notification-using-a-hook)
* Champion Scenario 4:
* Describe the champion scenario: query anomalies & alerts
* Estimate the percentage of developers using the service who would use the champion scenario
* Link to the code samples: [Python](https://gist.github.com/xiangyan99/6c367d45168294b043b2f5685d57f584#query-anomalies-for-alert-configuration) | [.NET](https://github.com/Azure/azure-sdk-for-net/blob/master/sdk/metricsadvisor/Azure.AI.MetricsAdvisor/README.md#query-detected-anomalies-and-triggered-alerts) | 1.0 | Board Review: metrics advisor (Python & .net) - ## The Basics
* Service team responsible for the client library: Metrics Advisor
* Link to documentation describing the service: https://docs.microsoft.com/en-us/azure/cognitive-services/metrics-advisor/
* Contact email (if service team, provide PM and Dev Lead):
[email protected], [email protected] (dev)
[email protected] (PM)
## About this client library
* Name of the client library: azure-ai-metricsadvisor
* Languages for this review: .net/Python
* Link to the service REST APIs:
https://github.com/bowgong/azure-rest-api-specs/blob/metricsadvisor-preview/specification/cognitiveservices/data-plane/MetricsAdvisor/preview/v1.0/MetricsAdvisor.json
https://westus2.dev.cognitive.microsoft.com/docs/services/MetricsAdvisor/operations/createDataFeed
## Artifacts required (per language)
We use an API review tool ([apiview](https://apiview.azurewebsites.net)) to support .NET and Java API reviews. For Python and TypeScript, use the API extractor tool, then submit the output as a Draft PR to the relevant repository (azure-sdk-for-python or azure-sdk-for-js).
### .NET
* [APIView](https://apiview.dev/Assemblies/Review/8caf3dd1661c45228d8081a536cca3bc)
* Link to samples for champion scenarios: https://github.com/Azure/azure-sdk-for-net/blob/master/sdk/metricsadvisor/Azure.AI.MetricsAdvisor/samples/README.md
### Python
* [APIView](https://apiview.dev/Assemblies/Review/6acc354ec6c5421b82a081a07b481df0)
* Link to samples for champion scenarios: https://github.com/Azure/azure-sdk-for-python/tree/master/sdk/metricsadvisor/azure-ai-metricsadvisor/samples
## Champion Scenarios
A champion scenario is a use case that the consumer of the client library is commonly expected to perform. Champion scenarios are used to ensure the developer experience is exemplary for the common cases. You need to show the entire code sample (including error handling, as an example) for the champion scenarios.
* Champion Scenario 1:
* Describe the champion scenario: create a data feed to ingest data
* Estimate the percentage of developers using the service who would use the champion scenario
* Link to the code samples: [Python](https://gist.github.com/xiangyan99/6c367d45168294b043b2f5685d57f584#create-data-feed) | [.NET](https://gist.github.com/kinelski/c072790394398c37d186df611f6bea44#scenario-1-datafeed-creation)
* Champion Scenario 2:
* Describe the champion scenario: create a configuration to let service know whether a point is anomaly
* Estimate the percentage of developers using the service who would use the champion scenario
* Link to the code samples: [Python](https://gist.github.com/xiangyan99/6c367d45168294b043b2f5685d57f584#create-detection-configuration) | [.NET](https://gist.github.com/kinelski/c072790394398c37d186df611f6bea44#2d-applyingtuning-anomaly-detection)
* Champion Scenario 3:
* Describe the champion scenario: configure the service when to trigger an alert
* Estimate the percentage of developers using the service who would use the champion scenario
* Link to the code samples: [Python](https://gist.github.com/xiangyan99/6c367d45168294b043b2f5685d57f584#config-alert-configuration) | [.NET](https://gist.github.com/kinelski/c072790394398c37d186df611f6bea44#scenario-3-configure-alerts-and-get-incidents-notification-using-a-hook)
* Champion Scenario 4:
* Describe the champion scenario: query anomalies & alerts
* Estimate the percentage of developers using the service who would use the champion scenario
* Link to the code samples: [Python](https://gist.github.com/xiangyan99/6c367d45168294b043b2f5685d57f584#query-anomalies-for-alert-configuration) | [.NET](https://github.com/Azure/azure-sdk-for-net/blob/master/sdk/metricsadvisor/Azure.AI.MetricsAdvisor/README.md#query-detected-anomalies-and-triggered-alerts) | architecture | board review metrics advisor python net the basics service team responsible for the client library metrics advisor link to documentation describing the service contact email if service team provide pm and dev lead bix microsoft com bowgong microsoft com dev quying microsoft com pm about this client library name of the client library azure ai metricsadvisor languages for this review net python link to the service rest apis artifacts required per language we use an api review tool to support net and java api reviews for python and typescript use the api extractor tool then submit the output as a draft pr to the relevant repository azure sdk for python or azure sdk for js net link to samples for champion scenarios python link to samples for champion scenarios champion scenarios a champion scenario is a use case that the consumer of the client library is commonly expected to perform champion scenarios are used to ensure the developer experience is exemplary for the common cases you need to show the entire code sample including error handling as an example for the champion scenarios champion scenario describe the champion scenario create a data feed to ingest data estimate the percentage of developers using the service who would use the champion scenario link to the code samples champion scenario describe the champion scenario create a configuration to let service know whether a point is anomaly estimate the percentage of developers using the service who would use the champion scenario link to the code samples champion scenario describe the champion scenario configure the service when to trigger an alert estimate the percentage of developers using the service who would use the champion scenario link to the code samples champion scenario describe the champion scenario query anomalies alerts estimate the percentage of developers using the service who would use the champion scenario link to the code samples | 1 |