Dataset Viewer
Auto-converted to Parquet
Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
7
112
repo_url
stringlengths
36
141
action
stringclasses
3 values
title
stringlengths
1
900
labels
stringlengths
4
522
body
stringlengths
5
218k
index
stringclasses
6 values
text_combine
stringlengths
96
219k
label
stringclasses
2 values
text
stringlengths
96
102k
binary_label
int64
0
1
318,683
23,734,541,985
IssuesEvent
2022-08-31 06:50:08
codeing999/CLIPs-backend
https://api.github.com/repos/codeing999/CLIPs-backend
closed
git 협업 방법 변경
documentation
기존 브랜치 main : 최종적으로 배포할 브랜치 submain : 각자의 작업 브랜치에서 작업하다가 에러 없이 작업 완료될 때마다 수시로 푸쉬할 브랜치 dev/닉네임 : 각자의 작업 브랜치. 변경된 브랜치 main : 배포할 브랜치 develop : 개발 단계에서 각각의 기능이 완성될 때 마다 합칠 브랜치 feature/기능명, fix/에러 : 각 기능 단위로 브랜치를 새로 만들고 완성 시마다 develop에 합치고 이 브랜치는 삭제. 이 내용은 동의 하시면 docs에도 반영하여 수정하도록 하겠습니다.
1.0
git 협업 방법 변경 - 기존 브랜치 main : 최종적으로 배포할 브랜치 submain : 각자의 작업 브랜치에서 작업하다가 에러 없이 작업 완료될 때마다 수시로 푸쉬할 브랜치 dev/닉네임 : 각자의 작업 브랜치. 변경된 브랜치 main : 배포할 브랜치 develop : 개발 단계에서 각각의 기능이 완성될 때 마다 합칠 브랜치 feature/기능명, fix/에러 : 각 기능 단위로 브랜치를 새로 만들고 완성 시마다 develop에 합치고 이 브랜치는 삭제. 이 내용은 동의 하시면 docs에도 반영하여 수정하도록 하겠습니다.
non_architecture
git 협업 방법 변경 기존 브랜치 main 최종적으로 배포할 브랜치 submain 각자의 작업 브랜치에서 작업하다가 에러 없이 작업 완료될 때마다 수시로 푸쉬할 브랜치 dev 닉네임 각자의 작업 브랜치 변경된 브랜치 main 배포할 브랜치 develop 개발 단계에서 각각의 기능이 완성될 때 마다 합칠 브랜치 feature 기능명 fix 에러 각 기능 단위로 브랜치를 새로 만들고 완성 시마다 develop에 합치고 이 브랜치는 삭제 이 내용은 동의 하시면 docs에도 반영하여 수정하도록 하겠습니다
0
11,270
14,060,158,658
IssuesEvent
2020-11-03 05:16:59
gfx-rs/naga
https://api.github.com/repos/gfx-rs/naga
opened
Typifier -> Classifier
area: processing help wanted kind: feature kind: question
Currently, we have the typifier module that assigns expression types. We need more information induced from the expressions though. One such bit is what is the variability of an expression: global, uniform, or local. We could go at least two ways from here: 1. Introduce an expression visitor, which will be used by the typifier as well as other things, potentially in user space as well. 2. Rename the typifier to "classifier" and make it derive the variability together with the type. The (2) seems more straightforward to me, although I do wonder about a way to let users easily process our IR modules. Maybe it's just too early for that, and we'll need to pick the easier solution.
1.0
Typifier -> Classifier - Currently, we have the typifier module that assigns expression types. We need more information induced from the expressions though. One such bit is what is the variability of an expression: global, uniform, or local. We could go at least two ways from here: 1. Introduce an expression visitor, which will be used by the typifier as well as other things, potentially in user space as well. 2. Rename the typifier to "classifier" and make it derive the variability together with the type. The (2) seems more straightforward to me, although I do wonder about a way to let users easily process our IR modules. Maybe it's just too early for that, and we'll need to pick the easier solution.
non_architecture
typifier classifier currently we have the typifier module that assigns expression types we need more information induced from the expressions though one such bit is what is the variability of an expression global uniform or local we could go at least two ways from here introduce an expression visitor which will be used by the typifier as well as other things potentially in user space as well rename the typifier to classifier and make it derive the variability together with the type the seems more straightforward to me although i do wonder about a way to let users easily process our ir modules maybe it s just too early for that and we ll need to pick the easier solution
0
1,423
5,892,275,610
IssuesEvent
2017-05-17 19:04:08
gctools-outilsgc/gcconnex
https://api.github.com/repos/gctools-outilsgc/gcconnex
closed
design prompt for auto suggesting tags / audience on new content creation
enhancement high-level design Information Architecture - Controlled Vocabulary
design mock up for user prompt suggesting tags / audiences for content.
1.0
design prompt for auto suggesting tags / audience on new content creation - design mock up for user prompt suggesting tags / audiences for content.
architecture
design prompt for auto suggesting tags audience on new content creation design mock up for user prompt suggesting tags audiences for content
1
6,669
15,014,277,621
IssuesEvent
2021-02-01 06:18:02
burespe1/FRAME
https://api.github.com/repos/burespe1/FRAME
closed
Functional View to Physical View
EA Development architecture methodologies automation on hold physical view
following are the steps to be followed to build a physical view. 1) ![Picture14](https://user-images.githubusercontent.com/71774192/99528673-b48bab80-29af-11eb-9c6f-b388fe53c8fe.png) 2) ![Picture15](https://user-images.githubusercontent.com/71774192/99528688-b81f3280-29af-11eb-9094-fb28ca1b3ed6.png) 3) ![Picture16](https://user-images.githubusercontent.com/71774192/99528699-bc4b5000-29af-11eb-807a-4908ee8164b2.png) can we automate this process within EA?
1.0
Functional View to Physical View - following are the steps to be followed to build a physical view. 1) ![Picture14](https://user-images.githubusercontent.com/71774192/99528673-b48bab80-29af-11eb-9c6f-b388fe53c8fe.png) 2) ![Picture15](https://user-images.githubusercontent.com/71774192/99528688-b81f3280-29af-11eb-9094-fb28ca1b3ed6.png) 3) ![Picture16](https://user-images.githubusercontent.com/71774192/99528699-bc4b5000-29af-11eb-807a-4908ee8164b2.png) can we automate this process within EA?
architecture
functional view to physical view following are the steps to be followed to build a physical view can we automate this process within ea
1
7,842
19,649,865,557
IssuesEvent
2022-01-10 04:56:42
Vector35/binaryninja-api
https://api.github.com/repos/Vector35/binaryninja-api
closed
arm64 `fcmp` doesn't work with many condition codes
enhancement architecture ARM64 Effort: Low Impact: Medium
**Version and Platform (required):** - Binary Ninja Version: Version 2.5.3140-dev (Build ID 532595b6) - OS: macOS - OS Version: 11.6 **Bug Description:** [This binary](https://github.com/Vector35/binaryninja-api/files/7790320/fcmp.zip) demonstrates using all 14 ARM condition codes with `fcmp`. Among these: - `eq`, `ne`, `cs`, and `cc` look good. - `ge`, `lt`, `gt`, and `le` are decompiled using `unimplemented`, e.g.: ``` 000000f0 int64_t test_ge(int32_t arg1 @ v0, int32_t arg2 @ v1) 000000f0 int64_t x0 = 0 000000f4 arg1 f- arg2 000000f4 bool v = unimplemented {fcmp s0, s1} 000000f4 bool n = unimplemented {fcmp s0, s1} 000000f8 if (n == v) 00000100 x0 = 1 000000fc return x0 ``` This is unfortunate, since those condition codes are very common. - `ls` and its inverse `hi` are decompiled correctly but suboptimally. For `ls`: ``` 000000d8 int64_t test_ls(float arg1 @ v0, float arg2 @ v1) 000000d8 int64_t x0 = 0 000000dc arg1 - arg2 000000e0 if (arg1 == arg2 || arg1 < arg2) 000000e8 x0 = 1 000000e4 return x0 ``` This could be `arg1 <= arg2`. - `pl` is decompiled incorrectly: ``` 00000078 int64_t test_pl(int32_t arg1 @ v0, int32_t arg2 @ v1) 00000078 int64_t x0 = 0 00000080 if (arg1 f- arg2 s>= 0) 00000088 x0 = 1 00000084 return x0 ``` In reality it should be `not(arg1 < arg2)` (see below table), which is not the same as `arg1 - arg2 >= 0`. For example, if one or both arguments is NaN, `not(arg1 < arg2)` is true, but `arg1 - arg2 >= 0` is false. - `mi` is similarly oddly decompiled as `arg1 f- arg2 s< 0` when it should be `arg1 < arg2`. I can't think of any cases within standard IEEE floating point where these expressions aren't equivalent, but under IEEE floating point with subnormals disabled (common in games), [there are pairs of floats](https://stackoverflow.com/a/54532647) `a`, `b` such that `a < b` but `a - b == 0.0`. Here is the relevant table from the ARM manual: <img width="819" alt="image" src="https://user-images.githubusercontent.com/47517/147696978-bd8616b1-e459-4cbd-bdaf-7db3e19bffda.png">
1.0
arm64 `fcmp` doesn't work with many condition codes - **Version and Platform (required):** - Binary Ninja Version: Version 2.5.3140-dev (Build ID 532595b6) - OS: macOS - OS Version: 11.6 **Bug Description:** [This binary](https://github.com/Vector35/binaryninja-api/files/7790320/fcmp.zip) demonstrates using all 14 ARM condition codes with `fcmp`. Among these: - `eq`, `ne`, `cs`, and `cc` look good. - `ge`, `lt`, `gt`, and `le` are decompiled using `unimplemented`, e.g.: ``` 000000f0 int64_t test_ge(int32_t arg1 @ v0, int32_t arg2 @ v1) 000000f0 int64_t x0 = 0 000000f4 arg1 f- arg2 000000f4 bool v = unimplemented {fcmp s0, s1} 000000f4 bool n = unimplemented {fcmp s0, s1} 000000f8 if (n == v) 00000100 x0 = 1 000000fc return x0 ``` This is unfortunate, since those condition codes are very common. - `ls` and its inverse `hi` are decompiled correctly but suboptimally. For `ls`: ``` 000000d8 int64_t test_ls(float arg1 @ v0, float arg2 @ v1) 000000d8 int64_t x0 = 0 000000dc arg1 - arg2 000000e0 if (arg1 == arg2 || arg1 < arg2) 000000e8 x0 = 1 000000e4 return x0 ``` This could be `arg1 <= arg2`. - `pl` is decompiled incorrectly: ``` 00000078 int64_t test_pl(int32_t arg1 @ v0, int32_t arg2 @ v1) 00000078 int64_t x0 = 0 00000080 if (arg1 f- arg2 s>= 0) 00000088 x0 = 1 00000084 return x0 ``` In reality it should be `not(arg1 < arg2)` (see below table), which is not the same as `arg1 - arg2 >= 0`. For example, if one or both arguments is NaN, `not(arg1 < arg2)` is true, but `arg1 - arg2 >= 0` is false. - `mi` is similarly oddly decompiled as `arg1 f- arg2 s< 0` when it should be `arg1 < arg2`. I can't think of any cases within standard IEEE floating point where these expressions aren't equivalent, but under IEEE floating point with subnormals disabled (common in games), [there are pairs of floats](https://stackoverflow.com/a/54532647) `a`, `b` such that `a < b` but `a - b == 0.0`. Here is the relevant table from the ARM manual: <img width="819" alt="image" src="https://user-images.githubusercontent.com/47517/147696978-bd8616b1-e459-4cbd-bdaf-7db3e19bffda.png">
architecture
fcmp doesn t work with many condition codes version and platform required binary ninja version version dev build id os macos os version bug description demonstrates using all arm condition codes with fcmp among these eq ne cs and cc look good ge lt gt and le are decompiled using unimplemented e g t test ge t t t f bool v unimplemented fcmp bool n unimplemented fcmp if n v return this is unfortunate since those condition codes are very common ls and its inverse hi are decompiled correctly but suboptimally for ls t test ls float float t if return this could be pl is decompiled incorrectly t test pl t t t if f s return in reality it should be not for example if one or both arguments is nan not is false mi is similarly oddly decompiled as f s when it should be i can t think of any cases within standard ieee floating point where these expressions aren t equivalent but under ieee floating point with subnormals disabled common in games a b such that a b but a b here is the relevant table from the arm manual img width alt image src
1
4,840
11,757,762,280
IssuesEvent
2020-03-13 14:16:01
kubernetes/kubernetes
https://api.github.com/repos/kubernetes/kubernetes
closed
Investigate Transitive Deps from docker/libnetwork
area/code-organization area/dependency kind/feature lifecycle/rotten sig/architecture sig/network
**What would you like to be added**: github.com/docker/libnetwork has a lot of transitive deps but we only use its (relatively small) `ipvs` package for kube-proxy. Luckily because it doesn't use go modules yet, updating it doesn't actually update it's transitive deps, but if/when it does, managing it's transitive deps will be a pain because of its overlapping transitive deps to our other dependencies. From doing a quick search of the Kubernetes repo, we only use the `ipvs` package from github.com/docker/libnetwork. Some options going forward would be: * fork that repo with only the `ipvs` package * copy the `ipvs` package to k8s.io/kubernetes * ask the docker maintainers to put the ipvs package into a separate repo * ??? Open to other options I haven't considered yet. **Why is this needed**: Will significantly improve the maintainability of our dependencies once github.com/docker/libnetwork uses go modules.
1.0
Investigate Transitive Deps from docker/libnetwork - **What would you like to be added**: github.com/docker/libnetwork has a lot of transitive deps but we only use its (relatively small) `ipvs` package for kube-proxy. Luckily because it doesn't use go modules yet, updating it doesn't actually update it's transitive deps, but if/when it does, managing it's transitive deps will be a pain because of its overlapping transitive deps to our other dependencies. From doing a quick search of the Kubernetes repo, we only use the `ipvs` package from github.com/docker/libnetwork. Some options going forward would be: * fork that repo with only the `ipvs` package * copy the `ipvs` package to k8s.io/kubernetes * ask the docker maintainers to put the ipvs package into a separate repo * ??? Open to other options I haven't considered yet. **Why is this needed**: Will significantly improve the maintainability of our dependencies once github.com/docker/libnetwork uses go modules.
architecture
investigate transitive deps from docker libnetwork what would you like to be added github com docker libnetwork has a lot of transitive deps but we only use its relatively small ipvs package for kube proxy luckily because it doesn t use go modules yet updating it doesn t actually update it s transitive deps but if when it does managing it s transitive deps will be a pain because of its overlapping transitive deps to our other dependencies from doing a quick search of the kubernetes repo we only use the ipvs package from github com docker libnetwork some options going forward would be fork that repo with only the ipvs package copy the ipvs package to io kubernetes ask the docker maintainers to put the ipvs package into a separate repo open to other options i haven t considered yet why is this needed will significantly improve the maintainability of our dependencies once github com docker libnetwork uses go modules
1
134,670
30,113,591,798
IssuesEvent
2023-06-30 09:40:04
FerretDB/FerretDB
https://api.github.com/repos/FerretDB/FerretDB
opened
Implement `$tsSecond` timestamp expression operator
code/feature not ready area/aggregations
### What should be done? It should be supported in all pipeline stages that support raw expressions and other pipelines that allow the `$expr` operator. * https://www.mongodb.com/docs/manual/reference/operator/aggregation/tsSecond/#mongodb-expression-exp.-tsSecond
1.0
Implement `$tsSecond` timestamp expression operator - ### What should be done? It should be supported in all pipeline stages that support raw expressions and other pipelines that allow the `$expr` operator. * https://www.mongodb.com/docs/manual/reference/operator/aggregation/tsSecond/#mongodb-expression-exp.-tsSecond
non_architecture
implement tssecond timestamp expression operator what should be done it should be supported in all pipeline stages that support raw expressions and other pipelines that allow the expr operator
0
10,997
27,734,774,944
IssuesEvent
2023-03-15 10:28:09
OasisLMF/OasisPlatform
https://api.github.com/repos/OasisLMF/OasisPlatform
opened
Fix Helm customization Readme
bug Documentation scalable architecture
## Issue Description > **carlfischerjba:** Helm appears to be combining the default values files from `OasisPlatform/kubernetes/charts/oasis-models/values.yaml` with the file I specify on the command line. > > Apparently, the `workers` (a mapping) get merged so we have the default `piwind-demo` as well as the New workers I've defined, but the `modelVolumes` (a sequence) are overwritten so we only have the New volumes and not `piwind-model-data-pv`, this leads to the error. > > This means [the methods recommended in the readme](https://github.com/OasisLMF/OasisPlatform/blob/6dd90eb3ced94e48464de158af88417de3b49b9a/kubernetes/charts/README.md?plain=1#L242) don't work. I guess it's not been spotted until now because everyone has kept the PiWind model in place. Trouble starts once you decide you don't need it. Merging values vs maps vs lists is a problem with docker-compose.yml and other types of config files too, including json. It's surprising there's no way to tell Helm to ignore the defaults. Without such an option, the ways to keep everything working are not very satisfactory: > * edit the values files in place > * rename or delete the default values files > * or copy the template as suggested but only add to it, never remove anything > > Another alternative would be for you to rename models/values.yaml to models/values_sample.yaml so it's ignored by Helm and update the instructions in the readme accordingly. > > The same could occur with the platform and monitoring charts but I guess that's less likely because in general a few values are modified and override the defaults instead of adding extra items to mappings or sequences (are those the correct YAML terms?). At a minimum the documentation instructions should be updated to note/fix this problem
1.0
Fix Helm customization Readme - ## Issue Description > **carlfischerjba:** Helm appears to be combining the default values files from `OasisPlatform/kubernetes/charts/oasis-models/values.yaml` with the file I specify on the command line. > > Apparently, the `workers` (a mapping) get merged so we have the default `piwind-demo` as well as the New workers I've defined, but the `modelVolumes` (a sequence) are overwritten so we only have the New volumes and not `piwind-model-data-pv`, this leads to the error. > > This means [the methods recommended in the readme](https://github.com/OasisLMF/OasisPlatform/blob/6dd90eb3ced94e48464de158af88417de3b49b9a/kubernetes/charts/README.md?plain=1#L242) don't work. I guess it's not been spotted until now because everyone has kept the PiWind model in place. Trouble starts once you decide you don't need it. Merging values vs maps vs lists is a problem with docker-compose.yml and other types of config files too, including json. It's surprising there's no way to tell Helm to ignore the defaults. Without such an option, the ways to keep everything working are not very satisfactory: > * edit the values files in place > * rename or delete the default values files > * or copy the template as suggested but only add to it, never remove anything > > Another alternative would be for you to rename models/values.yaml to models/values_sample.yaml so it's ignored by Helm and update the instructions in the readme accordingly. > > The same could occur with the platform and monitoring charts but I guess that's less likely because in general a few values are modified and override the defaults instead of adding extra items to mappings or sequences (are those the correct YAML terms?). At a minimum the documentation instructions should be updated to note/fix this problem
architecture
fix helm customization readme issue description carlfischerjba helm appears to be combining the default values files from oasisplatform kubernetes charts oasis models values yaml with the file i specify on the command line apparently the workers a mapping get merged so we have the default piwind demo as well as the new workers i ve defined but the modelvolumes a sequence are overwritten so we only have the new volumes and not piwind model data pv this leads to the error this means don t work i guess it s not been spotted until now because everyone has kept the piwind model in place trouble starts once you decide you don t need it merging values vs maps vs lists is a problem with docker compose yml and other types of config files too including json it s surprising there s no way to tell helm to ignore the defaults without such an option the ways to keep everything working are not very satisfactory edit the values files in place rename or delete the default values files or copy the template as suggested but only add to it never remove anything another alternative would be for you to rename models values yaml to models values sample yaml so it s ignored by helm and update the instructions in the readme accordingly the same could occur with the platform and monitoring charts but i guess that s less likely because in general a few values are modified and override the defaults instead of adding extra items to mappings or sequences are those the correct yaml terms at a minimum the documentation instructions should be updated to note fix this problem
1
10,129
26,364,651,544
IssuesEvent
2023-01-11 15:43:36
mehab/DTKafkaPOC
https://api.github.com/repos/mehab/DTKafkaPOC
opened
Add architecture diagrams
documentation 📃 architecture 🔮
In order for us (and others) to better understand what we're building here, we should have architecture diagrams. Preferably there should be multiple "resolutions" from a high-level overview to individual services. The topology diagrams we can generate using Kafka Streams will be helpful for the latter.
1.0
Add architecture diagrams - In order for us (and others) to better understand what we're building here, we should have architecture diagrams. Preferably there should be multiple "resolutions" from a high-level overview to individual services. The topology diagrams we can generate using Kafka Streams will be helpful for the latter.
architecture
add architecture diagrams in order for us and others to better understand what we re building here we should have architecture diagrams preferably there should be multiple resolutions from a high level overview to individual services the topology diagrams we can generate using kafka streams will be helpful for the latter
1
445,871
12,837,462,604
IssuesEvent
2020-07-07 15:49:31
code-ready/crc
https://api.github.com/repos/code-ready/crc
closed
Add 'Experimental' messages when `podman-env` command is used.
priority/critical status/stale
We need to add output to the `podman-env` command, as for the time being no changes will happen to this functionality Note: adding an `echo`, so when `podman-env` is `eval`-ed it would still show
1.0
Add 'Experimental' messages when `podman-env` command is used. - We need to add output to the `podman-env` command, as for the time being no changes will happen to this functionality Note: adding an `echo`, so when `podman-env` is `eval`-ed it would still show
non_architecture
add experimental messages when podman env command is used we need to add output to the podman env command as for the time being no changes will happen to this functionality note adding an echo so when podman env is eval ed it would still show
0
4,114
10,584,831,316
IssuesEvent
2019-10-08 16:11:07
fga-eps-mds/2019.2-Over26
https://api.github.com/repos/fga-eps-mds/2019.2-Over26
closed
Elaborar Plano de Qualidade
Architecture Documentation EPS
## Descrição da Mudança * <!--- Forneça um resumo geral da _issue_ --> Criar a primeira versão do plano de qualidade para o projeto. ## Checklist * <!-- Essa checklist propõe a criação de uma boa issue --> <!-- Se a issue é sobre uma história de usuário, seu nome deve ser "USXX - Nome da história--> <!-- Se a issue é sobre um bug, seu nome deve ser "BF - Nome curto do bug"--> <!-- Se a issue é sobre outra tarefa o nome deve ser uma simples descrição da tarefa--> - [x] Esta issue tem um nome significativo. - [x] O nome da issue está no padrão. - [x] Esta issue tem uma descrição de fácil entendimento. - [x] Esta issue tem uma boa definição de critérios de aceitação. - [x] Esta issue tem labels associadas. - [ ] Esta issue está associada à uma milestone. - [ ] Esta issue tem uma pontuação estimada. ## Tarefas * <!-- Adicione aqui as tarefas necessárias para concluir a issue --> - [ ] Criar plano de qualidade ## Critérios de Aceitação * <!-- Liste aqui o conjunto de aspectos mecessários para considerar a atividade como completa--> <!-- Os itens serão adicionados pelo Product Owner --> - [ ] A primeira versão do plano de qualidade deve estar elaborada
1.0
Elaborar Plano de Qualidade - ## Descrição da Mudança * <!--- Forneça um resumo geral da _issue_ --> Criar a primeira versão do plano de qualidade para o projeto. ## Checklist * <!-- Essa checklist propõe a criação de uma boa issue --> <!-- Se a issue é sobre uma história de usuário, seu nome deve ser "USXX - Nome da história--> <!-- Se a issue é sobre um bug, seu nome deve ser "BF - Nome curto do bug"--> <!-- Se a issue é sobre outra tarefa o nome deve ser uma simples descrição da tarefa--> - [x] Esta issue tem um nome significativo. - [x] O nome da issue está no padrão. - [x] Esta issue tem uma descrição de fácil entendimento. - [x] Esta issue tem uma boa definição de critérios de aceitação. - [x] Esta issue tem labels associadas. - [ ] Esta issue está associada à uma milestone. - [ ] Esta issue tem uma pontuação estimada. ## Tarefas * <!-- Adicione aqui as tarefas necessárias para concluir a issue --> - [ ] Criar plano de qualidade ## Critérios de Aceitação * <!-- Liste aqui o conjunto de aspectos mecessários para considerar a atividade como completa--> <!-- Os itens serão adicionados pelo Product Owner --> - [ ] A primeira versão do plano de qualidade deve estar elaborada
architecture
elaborar plano de qualidade descrição da mudança criar a primeira versão do plano de qualidade para o projeto checklist esta issue tem um nome significativo o nome da issue está no padrão esta issue tem uma descrição de fácil entendimento esta issue tem uma boa definição de critérios de aceitação esta issue tem labels associadas esta issue está associada à uma milestone esta issue tem uma pontuação estimada tarefas criar plano de qualidade critérios de aceitação a primeira versão do plano de qualidade deve estar elaborada
1
1,694
6,553,962,218
IssuesEvent
2017-09-06 02:15:13
City-Bureau/documenters-aggregator
https://api.github.com/repos/City-Bureau/documenters-aggregator
opened
What geocoder service should we use?
architecture: spiders priority: high (must have)
See https://github.com/City-Bureau/documenters-aggregator/pull/85#issuecomment-327325384 Leaning towards Mapbox for now and will use to close #85.
1.0
What geocoder service should we use? - See https://github.com/City-Bureau/documenters-aggregator/pull/85#issuecomment-327325384 Leaning towards Mapbox for now and will use to close #85.
architecture
what geocoder service should we use see leaning towards mapbox for now and will use to close
1
2,097
7,276,508,762
IssuesEvent
2018-02-21 16:34:13
AnalyticalGraphicsInc/cesium
https://api.github.com/repos/AnalyticalGraphicsInc/cesium
closed
CesiumMath vs Math naming ambiguity
category - architecture / api category - doc
It's not clear that the `CesiumMath` class is included in the namespace as `Cesium.Math`. This is also inconsistent with other classes that have the Cesium prefix, like `Cesium3DTileset`. If this is not something we want to change in the API, this should be made clear in the documentation. Relevant forum thread: https://groups.google.com/forum/#!topic/cesium-dev/icpMxc_bea8
1.0
CesiumMath vs Math naming ambiguity - It's not clear that the `CesiumMath` class is included in the namespace as `Cesium.Math`. This is also inconsistent with other classes that have the Cesium prefix, like `Cesium3DTileset`. If this is not something we want to change in the API, this should be made clear in the documentation. Relevant forum thread: https://groups.google.com/forum/#!topic/cesium-dev/icpMxc_bea8
architecture
cesiummath vs math naming ambiguity it s not clear that the cesiummath class is included in the namespace as cesium math this is also inconsistent with other classes that have the cesium prefix like if this is not something we want to change in the api this should be made clear in the documentation relevant forum thread
1
179,282
21,557,595,193
IssuesEvent
2022-04-30 17:37:48
NixOS/nixpkgs
https://api.github.com/repos/NixOS/nixpkgs
closed
Vulnerability roundup 113: ffmpeg-5.0.1: 1 advisory [7.5]
1.severity: security
[search](https://search.nix.gsc.io/?q=ffmpeg&i=fosho&repos=NixOS-nixpkgs), [files](https://github.com/NixOS/nixpkgs/search?utf8=%E2%9C%93&q=ffmpeg+in%3Apath&type=Code) * [ ] [CVE-2021-38291](https://nvd.nist.gov/vuln/detail/CVE-2021-38291) CVSSv3=7.5 (nixos-unstable) ## CVE details ### CVE-2021-38291 FFmpeg version (git commit de8e6e67e7523e48bb27ac224a0b446df05e1640) suffers from a an assertion failure at src/libavutil/mathematics.c. ----- Scanned versions: nixos-unstable: ff9efb0724d. Cc @codyopel
True
Vulnerability roundup 113: ffmpeg-5.0.1: 1 advisory [7.5] - [search](https://search.nix.gsc.io/?q=ffmpeg&i=fosho&repos=NixOS-nixpkgs), [files](https://github.com/NixOS/nixpkgs/search?utf8=%E2%9C%93&q=ffmpeg+in%3Apath&type=Code) * [ ] [CVE-2021-38291](https://nvd.nist.gov/vuln/detail/CVE-2021-38291) CVSSv3=7.5 (nixos-unstable) ## CVE details ### CVE-2021-38291 FFmpeg version (git commit de8e6e67e7523e48bb27ac224a0b446df05e1640) suffers from a an assertion failure at src/libavutil/mathematics.c. ----- Scanned versions: nixos-unstable: ff9efb0724d. Cc @codyopel
non_architecture
vulnerability roundup ffmpeg advisory nixos unstable cve details cve ffmpeg version git commit suffers from a an assertion failure at src libavutil mathematics c scanned versions nixos unstable cc codyopel
0
4,072
10,552,476,500
IssuesEvent
2019-10-03 15:14:11
dotnet/docs
https://api.github.com/repos/dotnet/docs
closed
Multuple IHostedService registration
:book: guide - .NET Microservices :books: Area - .NET Architecture Guide Source - Docs.ms
If i try to register two or more services, only one could work properly. For example: ``` services.AddSingleton<IHostedService, ServiceA>(); services.AddSingleton<IHostedService, ServiceB>(); ``` Implementations are simplest as possible: ``` public class ServiceA: IHostedService { public Task StartAsync(CancellationToken cancellationToken) { DoWork(); return Task.CompletedTask; } public Task StopAsync(CancellationToken cancellationToken) { return Task.CompletedTask; } private void DoWork() { while (true) { Console.WriteLine("ServiceA"); Thread.Sleep(2000); } } } ``` and ``` public class ServiceB: IHostedService { public Task StartAsync(CancellationToken cancellationToken) { DoWork(); return Task.CompletedTask; } public Task StopAsync(CancellationToken cancellationToken) { return Task.CompletedTask; } private void DoWork() { while (true) { Console.WriteLine("ServiceB"); Thread.Sleep(1000); } } } ``` In output getting messages only from ServiceA --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: d49a03c0-a844-26eb-48a5-33a612dd3ead * Version Independent ID: 0707865f-9db7-0d71-42a5-bc1a1e89680a * Content: [Implement background tasks in microservices with IHostedService and the BackgroundService class](https://docs.microsoft.com/en-us/dotnet/architecture/microservices/multi-container-microservice-net-applications/background-tasks-with-ihostedservice#feedback) * Content Source: [docs/architecture/microservices/multi-container-microservice-net-applications/background-tasks-with-ihostedservice.md](https://github.com/dotnet/docs/blob/master/docs/architecture/microservices/multi-container-microservice-net-applications/background-tasks-with-ihostedservice.md) * Product: **dotnet** * Technology: **dotnet-ebooks** * GitHub Login: @nishanil * Microsoft Alias: **nanil**
1.0
Multuple IHostedService registration - If i try to register two or more services, only one could work properly. For example: ``` services.AddSingleton<IHostedService, ServiceA>(); services.AddSingleton<IHostedService, ServiceB>(); ``` Implementations are simplest as possible: ``` public class ServiceA: IHostedService { public Task StartAsync(CancellationToken cancellationToken) { DoWork(); return Task.CompletedTask; } public Task StopAsync(CancellationToken cancellationToken) { return Task.CompletedTask; } private void DoWork() { while (true) { Console.WriteLine("ServiceA"); Thread.Sleep(2000); } } } ``` and ``` public class ServiceB: IHostedService { public Task StartAsync(CancellationToken cancellationToken) { DoWork(); return Task.CompletedTask; } public Task StopAsync(CancellationToken cancellationToken) { return Task.CompletedTask; } private void DoWork() { while (true) { Console.WriteLine("ServiceB"); Thread.Sleep(1000); } } } ``` In output getting messages only from ServiceA --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: d49a03c0-a844-26eb-48a5-33a612dd3ead * Version Independent ID: 0707865f-9db7-0d71-42a5-bc1a1e89680a * Content: [Implement background tasks in microservices with IHostedService and the BackgroundService class](https://docs.microsoft.com/en-us/dotnet/architecture/microservices/multi-container-microservice-net-applications/background-tasks-with-ihostedservice#feedback) * Content Source: [docs/architecture/microservices/multi-container-microservice-net-applications/background-tasks-with-ihostedservice.md](https://github.com/dotnet/docs/blob/master/docs/architecture/microservices/multi-container-microservice-net-applications/background-tasks-with-ihostedservice.md) * Product: **dotnet** * Technology: **dotnet-ebooks** * GitHub Login: @nishanil * Microsoft Alias: **nanil**
architecture
multuple ihostedservice registration if i try to register two or more services only one could work properly for example services addsingleton services addsingleton implementations are simplest as possible public class servicea ihostedservice public task startasync cancellationtoken cancellationtoken dowork return task completedtask public task stopasync cancellationtoken cancellationtoken return task completedtask private void dowork while true console writeline servicea thread sleep and public class serviceb ihostedservice public task startasync cancellationtoken cancellationtoken dowork return task completedtask public task stopasync cancellationtoken cancellationtoken return task completedtask private void dowork while true console writeline serviceb thread sleep in output getting messages only from servicea document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source product dotnet technology dotnet ebooks github login nishanil microsoft alias nanil
1
255,598
21,939,757,136
IssuesEvent
2022-05-23 16:49:11
ooni/probe
https://api.github.com/repos/ooni/probe
closed
oonimkall: setRunType: cannot find symbol
bug testing ooni/probe-mobile priority/high platform/android platform/ios ooni/probe-engine
Linking with the oonimkall.aar engine fails with this error: ``` > Task :engine:compileExperimentalReleaseJavaWithJavac FAILED $monorepo/repo/probe-android/engine/src/main/java/org/openobservatory/engine/OONICheckInConfig.java:67: error: cannot find symbol c.setRunType(runType); ^ symbol: method setRunType(String) location: variable c of type CheckInConfig 1 error FAILURE: Build failed with an exception. ``` We need to fix this error in order to release 3.15.0. (cc: @hellais @aanorbel)
1.0
oonimkall: setRunType: cannot find symbol - Linking with the oonimkall.aar engine fails with this error: ``` > Task :engine:compileExperimentalReleaseJavaWithJavac FAILED $monorepo/repo/probe-android/engine/src/main/java/org/openobservatory/engine/OONICheckInConfig.java:67: error: cannot find symbol c.setRunType(runType); ^ symbol: method setRunType(String) location: variable c of type CheckInConfig 1 error FAILURE: Build failed with an exception. ``` We need to fix this error in order to release 3.15.0. (cc: @hellais @aanorbel)
non_architecture
oonimkall setruntype cannot find symbol linking with the oonimkall aar engine fails with this error task engine compileexperimentalreleasejavawithjavac failed monorepo repo probe android engine src main java org openobservatory engine oonicheckinconfig java error cannot find symbol c setruntype runtype symbol method setruntype string location variable c of type checkinconfig error failure build failed with an exception we need to fix this error in order to release cc hellais aanorbel
0
1,559
6,335,238,819
IssuesEvent
2017-07-26 18:24:54
LearnersGuild/echo
https://api.github.com/repos/LearnersGuild/echo
reopened
Move changefeed listeners to web service
architecture chore
Currently, the workers set up listening to db changefeeds and, effectively, queue items for their _own_ work queues. This is problematic because if changes to the database occur that should be handled by a worker while the worker is not running, that change event is lost and the job is never processed. Instead of connecting the changefeed listeners in the worker process, we should do it in the standing web service. This also makes it easier to choose to alternative mechanisms for background task execution (instead of having always-running worker dynos).
1.0
Move changefeed listeners to web service - Currently, the workers set up listening to db changefeeds and, effectively, queue items for their _own_ work queues. This is problematic because if changes to the database occur that should be handled by a worker while the worker is not running, that change event is lost and the job is never processed. Instead of connecting the changefeed listeners in the worker process, we should do it in the standing web service. This also makes it easier to choose to alternative mechanisms for background task execution (instead of having always-running worker dynos).
architecture
move changefeed listeners to web service currently the workers set up listening to db changefeeds and effectively queue items for their own work queues this is problematic because if changes to the database occur that should be handled by a worker while the worker is not running that change event is lost and the job is never processed instead of connecting the changefeed listeners in the worker process we should do it in the standing web service this also makes it easier to choose to alternative mechanisms for background task execution instead of having always running worker dynos
1
796,666
28,123,191,979
IssuesEvent
2023-03-31 15:32:50
thetrevorharmon/thetrevorharmon.com
https://api.github.com/repos/thetrevorharmon/thetrevorharmon.com
closed
Add landing page for mailing list
enhancement low priority
Now that there is a mailing list on the site, It would be good to have a landing page for the signup. Something like `/signup` with a simple form and nice explanation of what the signup gets them.
1.0
Add landing page for mailing list - Now that there is a mailing list on the site, It would be good to have a landing page for the signup. Something like `/signup` with a simple form and nice explanation of what the signup gets them.
non_architecture
add landing page for mailing list now that there is a mailing list on the site it would be good to have a landing page for the signup something like signup with a simple form and nice explanation of what the signup gets them
0
5,105
12,098,281,050
IssuesEvent
2020-04-20 10:02:29
stsrki/Blazorise
https://api.github.com/repos/stsrki/Blazorise
closed
Unit testing of components
Status: Investigate Type: Architecture
Investigate more about the new unit testing made by Steve Sanderson, after the release of Blazor preview 9. Sources: http://blog.stevensanderson.com/2019/08/29/blazor-unit-testing-prototype/ https://github.com/SteveSandersonMS/BlazorUnitTestingPrototype
1.0
Unit testing of components - Investigate more about the new unit testing made by Steve Sanderson, after the release of Blazor preview 9. Sources: http://blog.stevensanderson.com/2019/08/29/blazor-unit-testing-prototype/ https://github.com/SteveSandersonMS/BlazorUnitTestingPrototype
architecture
unit testing of components investigate more about the new unit testing made by steve sanderson after the release of blazor preview sources
1
357,018
10,600,740,255
IssuesEvent
2019-10-10 10:43:48
robotology/whole-body-controllers
https://api.github.com/repos/robotology/whole-body-controllers
opened
Investigate if it makes sense to port matlab-multi-body-sim in wbc
feature priority: normal
I would like to port `matlab-multi-body-sim` in the wbc, but before I need to understand the ratio between effort and benefits
1.0
Investigate if it makes sense to port matlab-multi-body-sim in wbc - I would like to port `matlab-multi-body-sim` in the wbc, but before I need to understand the ratio between effort and benefits
non_architecture
investigate if it makes sense to port matlab multi body sim in wbc i would like to port matlab multi body sim in the wbc but before i need to understand the ratio between effort and benefits
0
6,035
13,541,185,291
IssuesEvent
2020-09-16 15:33:34
MicrosoftDocs/architecture-center
https://api.github.com/repos/MicrosoftDocs/architecture-center
closed
Naming conventions of Icons
Pri2 architecture-center/svc assigned-to-author doc-enhancement triaged
We are doing some work with ARM templates and attempting to use the latest Icon sets. The problem we are seeing is there is no consistence way in which the icons are named which can map to templates or schemas. Please could it be considered . e.g Template schemas refer to Virtual scale sets as 'virtualMachineScaleSets', but the icon is buried in the Compute directory called "10034-icon-service-VM-Scalte-Sets". How are we meant to quick / auto identify between the too. Before the naming convention did link back to the recourses types --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 6624d44c-f9f6-02f8-2c67-0f0637fed62e * Version Independent ID: 9346a8ff-bab5-be2c-9dc1-d546fda1efa9 * Content: [Azure Icons - Azure Architecture Center](https://docs.microsoft.com/en-us/azure/architecture/icons/) * Content Source: [docs/icons/index.md](https://github.com/microsoftdocs/architecture-center/blob/master/docs/icons/index.md) * Service: **architecture-center** * GitHub Login: @doodlemania2 * Microsoft Alias: **pnp**
1.0
Naming conventions of Icons - We are doing some work with ARM templates and attempting to use the latest Icon sets. The problem we are seeing is there is no consistence way in which the icons are named which can map to templates or schemas. Please could it be considered . e.g Template schemas refer to Virtual scale sets as 'virtualMachineScaleSets', but the icon is buried in the Compute directory called "10034-icon-service-VM-Scalte-Sets". How are we meant to quick / auto identify between the too. Before the naming convention did link back to the recourses types --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 6624d44c-f9f6-02f8-2c67-0f0637fed62e * Version Independent ID: 9346a8ff-bab5-be2c-9dc1-d546fda1efa9 * Content: [Azure Icons - Azure Architecture Center](https://docs.microsoft.com/en-us/azure/architecture/icons/) * Content Source: [docs/icons/index.md](https://github.com/microsoftdocs/architecture-center/blob/master/docs/icons/index.md) * Service: **architecture-center** * GitHub Login: @doodlemania2 * Microsoft Alias: **pnp**
architecture
naming conventions of icons we are doing some work with arm templates and attempting to use the latest icon sets the problem we are seeing is there is no consistence way in which the icons are named which can map to templates or schemas please could it be considered e g template schemas refer to virtual scale sets as virtualmachinescalesets but the icon is buried in the compute directory called icon service vm scalte sets how are we meant to quick auto identify between the too before the naming convention did link back to the recourses types document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source service architecture center github login microsoft alias pnp
1
42,411
5,444,055,594
IssuesEvent
2017-03-07 01:15:24
dotnet/roslyn
https://api.github.com/repos/dotnet/roslyn
closed
Alternative approach to introduction of variables into scope.
Area-Language Design Discussion
There have a been a few raised voices that the current and suggested approach to how the new language features introduce variable into scope. He is my considered approach. --- **Default to nearest outer scope** The variable being introduce should be to **nearest outer scope** by default. ``` c# if( !(o is int i) ) { /* i is in scope and definitely not assigned */ throw new ArgumentException("Not an int", nameof(o)); } /* i is in scope and definitely assigned. */ ``` It simplifies one of the initial use cases, and I think the most likely usage. ``` c# int value; if( int.TryParse( text , out value ) ``` into ``` c# if( int.TryParse( text, out int value ) ``` --- It is the more eccentric usage of localizing to the nearest inner scope (of this particular code block). eg the `then` and `else` blocks. That is cause of tension for the community. especially around where it is introduce in "patterns". I propose that in these cases the change from the default (nearest outer) should made explicit and require the coder to specify a change to the **nearest inner scope**. For example a lightweight approach to this is by prefixing `~` on the variable identifier. ``` c# if (o is int ~i) { /* i is scope and definitely assigned */ } /* i is not in scope. /* ``` ``` c# if( !(o is int ~i) ) { /* i is scope finitely not assigned */ } else { /* I is scope and definitely assigned */ } /* i is not in scope. /* ``` --- Then there are the case where the coder would like the variable introduced in one of nearest blocks and not the other. Use case: Guards. ``` c# if( !int.TryParst( text, value ) ) { /* value not in scope. */ } else { /* value is in scope and assigned */ } ``` In pattern I propose we borrow from VB.net and use `IsNot` to indicate the negation of the pattern. ``` c# if (o isnot int ~i) { /* i is scope not assigned a new value*/ /* also is an error as i is being reused for a variable declaration */ } else { /* i is scope and definitely assigned a value.*/ /* also is an error as i is being reused for a variable declaration */ } /* i is in scope. /* ``` This should be easier to read ( ie not missing the easily missed `!` at the start. ), and allow us to know that the usage of variable introduction could be different. --- **When the variable identifier prior exists** What about the case where the variable already exist prior the usage? With the default ( nearest outer scope) the variable is reused. ``` c# int i; ... if( !(o is int i) ) { /* i is in scope and definitely maybe assigned */ throw new ArgumentException("Not an int", nameof(o)); } /* i is in scope and definitely assigned. */ ``` And in the nearest inner scope the variable can not reused and thus produces an compile-time error. Forcing the coder to make a explicit choice, and either make scope outer or use a different variable identifier. ``` c# int i; ... if (o is int ~i) { /* i is scope and definitely assigned */ /* also is and error as i is be reused for a variable declaration */ } /* i is in scope. /* ``` ``` c# int i; ... if (o isnot int ~i) { /* i is scope not assigned a new value*/ /* also is an error as i is being reused for a variable declaration */ } else { /* i is scope and definitely assigned a value.*/ /* also is an error as i is being reused for a variable declaration */ } /* i is in scope. /* ``` if the type of `i` is incompatible with the one in the pattern, it is an error in all cases.
1.0
Alternative approach to introduction of variables into scope. - There have a been a few raised voices that the current and suggested approach to how the new language features introduce variable into scope. He is my considered approach. --- **Default to nearest outer scope** The variable being introduce should be to **nearest outer scope** by default. ``` c# if( !(o is int i) ) { /* i is in scope and definitely not assigned */ throw new ArgumentException("Not an int", nameof(o)); } /* i is in scope and definitely assigned. */ ``` It simplifies one of the initial use cases, and I think the most likely usage. ``` c# int value; if( int.TryParse( text , out value ) ``` into ``` c# if( int.TryParse( text, out int value ) ``` --- It is the more eccentric usage of localizing to the nearest inner scope (of this particular code block). eg the `then` and `else` blocks. That is cause of tension for the community. especially around where it is introduce in "patterns". I propose that in these cases the change from the default (nearest outer) should made explicit and require the coder to specify a change to the **nearest inner scope**. For example a lightweight approach to this is by prefixing `~` on the variable identifier. ``` c# if (o is int ~i) { /* i is scope and definitely assigned */ } /* i is not in scope. /* ``` ``` c# if( !(o is int ~i) ) { /* i is scope finitely not assigned */ } else { /* I is scope and definitely assigned */ } /* i is not in scope. /* ``` --- Then there are the case where the coder would like the variable introduced in one of nearest blocks and not the other. Use case: Guards. ``` c# if( !int.TryParst( text, value ) ) { /* value not in scope. */ } else { /* value is in scope and assigned */ } ``` In pattern I propose we borrow from VB.net and use `IsNot` to indicate the negation of the pattern. ``` c# if (o isnot int ~i) { /* i is scope not assigned a new value*/ /* also is an error as i is being reused for a variable declaration */ } else { /* i is scope and definitely assigned a value.*/ /* also is an error as i is being reused for a variable declaration */ } /* i is in scope. /* ``` This should be easier to read ( ie not missing the easily missed `!` at the start. ), and allow us to know that the usage of variable introduction could be different. --- **When the variable identifier prior exists** What about the case where the variable already exist prior the usage? With the default ( nearest outer scope) the variable is reused. ``` c# int i; ... if( !(o is int i) ) { /* i is in scope and definitely maybe assigned */ throw new ArgumentException("Not an int", nameof(o)); } /* i is in scope and definitely assigned. */ ``` And in the nearest inner scope the variable can not reused and thus produces an compile-time error. Forcing the coder to make a explicit choice, and either make scope outer or use a different variable identifier. ``` c# int i; ... if (o is int ~i) { /* i is scope and definitely assigned */ /* also is and error as i is be reused for a variable declaration */ } /* i is in scope. /* ``` ``` c# int i; ... if (o isnot int ~i) { /* i is scope not assigned a new value*/ /* also is an error as i is being reused for a variable declaration */ } else { /* i is scope and definitely assigned a value.*/ /* also is an error as i is being reused for a variable declaration */ } /* i is in scope. /* ``` if the type of `i` is incompatible with the one in the pattern, it is an error in all cases.
non_architecture
alternative approach to introduction of variables into scope there have a been a few raised voices that the current and suggested approach to how the new language features introduce variable into scope he is my considered approach default to nearest outer scope the variable being introduce should be to nearest outer scope by default c if o is int i i is in scope and definitely not assigned throw new argumentexception not an int nameof o i is in scope and definitely assigned it simplifies one of the initial use cases and i think the most likely usage c int value if int tryparse text out value into c if int tryparse text out int value it is the more eccentric usage of localizing to the nearest inner scope of this particular code block eg the then and else blocks that is cause of tension for the community especially around where it is introduce in patterns i propose that in these cases the change from the default nearest outer should made explicit and require the coder to specify a change to the nearest inner scope for example a lightweight approach to this is by prefixing on the variable identifier c if o is int i i is scope and definitely assigned i is not in scope c if o is int i i is scope finitely not assigned else i is scope and definitely assigned i is not in scope then there are the case where the coder would like the variable introduced in one of nearest blocks and not the other use case guards c if int tryparst text value value not in scope else value is in scope and assigned in pattern i propose we borrow from vb net and use isnot to indicate the negation of the pattern c if o isnot int i i is scope not assigned a new value also is an error as i is being reused for a variable declaration else i is scope and definitely assigned a value also is an error as i is being reused for a variable declaration i is in scope this should be easier to read ie not missing the easily missed at the start and allow us to know that the usage of variable introduction could be different when the variable identifier prior exists what about the case where the variable already exist prior the usage with the default nearest outer scope the variable is reused c int i if o is int i i is in scope and definitely maybe assigned throw new argumentexception not an int nameof o i is in scope and definitely assigned and in the nearest inner scope the variable can not reused and thus produces an compile time error forcing the coder to make a explicit choice and either make scope outer or use a different variable identifier c int i if o is int i i is scope and definitely assigned also is and error as i is be reused for a variable declaration i is in scope c int i if o isnot int i i is scope not assigned a new value also is an error as i is being reused for a variable declaration else i is scope and definitely assigned a value also is an error as i is being reused for a variable declaration i is in scope if the type of i is incompatible with the one in the pattern it is an error in all cases
0
9,595
24,873,438,277
IssuesEvent
2022-10-27 16:59:01
Azure/azure-sdk
https://api.github.com/repos/Azure/azure-sdk
opened
Board Review: <client library name>
architecture board-review
Thank you for submitting this review request. Thorough review of your client library ensures that your APIs are consistent with the guidelines and the consumers of your client library have a consistently good experience when using Azure. **The Architecture Board reviews [Track 2 libraries](https://azure.github.io/azure-sdk/general_introduction.html) only.** If your library does not meet this requirement, please reach out to [Architecture Board]([email protected]) before creating the issue. Please reference our [review process guidelines](https://azure.github.io/azure-sdk/policies_reviewprocess.html) to understand what is being asked for in the issue template. To ensure consistency, all Tier-1 languages (C#, TypeScript, Java, Python) will generally be reviewed together. In expansive libraries, we will pair dynamic languages (Python, TypeScript) together, and strongly typed languages (C#, Java) together in separate meetings. For Tier-2 languages (C, C++, Go, Android, iOS), the review will be on an as-needed basis. **Before submitting, ensure you adjust the title of the issue appropriately.** **Note that the required material must be included before a meeting can be scheduled.** ## Contacts and Timeline * Responsible service team: Liftr Nginx * Main contacts: @SpencerOfwiti @limingu * Expected code complete date: Not Applicable * Expected release date: ## About the Service * Link to documentation introducing/describing the service: https://learn.microsoft.com/en-us/azure/partner-solutions/nginx/ * Link to the service REST APIs: https://github.com/Azure/azure-rest-api-specs/tree/main/specification/nginx/resource-manager/NGINX.NGINXPLUS/stable/2022-08-01 * Link to GitHub issue for previous review sessions, if applicable: ## About the client library * Name of the client library: * Languages for this review: The SDKs are autogenerated from the swagger, this review is only for namespace approval. ## Thank you!
1.0
Board Review: <client library name> - Thank you for submitting this review request. Thorough review of your client library ensures that your APIs are consistent with the guidelines and the consumers of your client library have a consistently good experience when using Azure. **The Architecture Board reviews [Track 2 libraries](https://azure.github.io/azure-sdk/general_introduction.html) only.** If your library does not meet this requirement, please reach out to [Architecture Board]([email protected]) before creating the issue. Please reference our [review process guidelines](https://azure.github.io/azure-sdk/policies_reviewprocess.html) to understand what is being asked for in the issue template. To ensure consistency, all Tier-1 languages (C#, TypeScript, Java, Python) will generally be reviewed together. In expansive libraries, we will pair dynamic languages (Python, TypeScript) together, and strongly typed languages (C#, Java) together in separate meetings. For Tier-2 languages (C, C++, Go, Android, iOS), the review will be on an as-needed basis. **Before submitting, ensure you adjust the title of the issue appropriately.** **Note that the required material must be included before a meeting can be scheduled.** ## Contacts and Timeline * Responsible service team: Liftr Nginx * Main contacts: @SpencerOfwiti @limingu * Expected code complete date: Not Applicable * Expected release date: ## About the Service * Link to documentation introducing/describing the service: https://learn.microsoft.com/en-us/azure/partner-solutions/nginx/ * Link to the service REST APIs: https://github.com/Azure/azure-rest-api-specs/tree/main/specification/nginx/resource-manager/NGINX.NGINXPLUS/stable/2022-08-01 * Link to GitHub issue for previous review sessions, if applicable: ## About the client library * Name of the client library: * Languages for this review: The SDKs are autogenerated from the swagger, this review is only for namespace approval. ## Thank you!
architecture
board review thank you for submitting this review request thorough review of your client library ensures that your apis are consistent with the guidelines and the consumers of your client library have a consistently good experience when using azure the architecture board reviews only if your library does not meet this requirement please reach out to adparch microsoft com before creating the issue please reference our to understand what is being asked for in the issue template to ensure consistency all tier languages c typescript java python will generally be reviewed together in expansive libraries we will pair dynamic languages python typescript together and strongly typed languages c java together in separate meetings for tier languages c c go android ios the review will be on an as needed basis before submitting ensure you adjust the title of the issue appropriately note that the required material must be included before a meeting can be scheduled contacts and timeline responsible service team liftr nginx main contacts spencerofwiti limingu expected code complete date not applicable expected release date about the service link to documentation introducing describing the service link to the service rest apis link to github issue for previous review sessions if applicable about the client library name of the client library languages for this review the sdks are autogenerated from the swagger this review is only for namespace approval thank you
1
213,119
23,966,109,535
IssuesEvent
2022-09-13 01:12:33
DavidSpek/kubeflow
https://api.github.com/repos/DavidSpek/kubeflow
opened
CVE-2022-36083 (Medium) detected in jose-2.0.5.tgz
security vulnerability
## CVE-2022-36083 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jose-2.0.5.tgz</b></p></summary> <p>JSON Web Almost Everything - JWA, JWS, JWE, JWK, JWT, JWKS for Node.js with minimal dependencies</p> <p>Library home page: <a href="https://registry.npmjs.org/jose/-/jose-2.0.5.tgz">https://registry.npmjs.org/jose/-/jose-2.0.5.tgz</a></p> <p>Path to dependency file: /components/crud-web-apps/volumes/frontend/package.json</p> <p>Path to vulnerable library: /components/crud-web-apps/volumes/frontend/node_modules/jose/package.json,/components/crud-web-apps/jupyter/frontend/node_modules/jose/package.json,/components/crud-web-apps/common/frontend/kubeflow-common-lib/node_modules/jose/package.json,/components/crud-web-apps/tensorboards/frontend/node_modules/jose/package.json</p> <p> Dependency Hierarchy: - client-node-0.12.2.tgz (Root Library) - openid-client-4.2.2.tgz - :x: **jose-2.0.5.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/DavidSpek/kubeflow/commit/00cbc9d11a3306fed1e979d79dff6ae36749d4bd">00cbc9d11a3306fed1e979d79dff6ae36749d4bd</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> JOSE is "JSON Web Almost Everything" - JWA, JWS, JWE, JWT, JWK, JWKS with no dependencies using runtime's native crypto in Node.js, Browser, Cloudflare Workers, Electron, and Deno. The PBKDF2-based JWE key management algorithms expect a JOSE Header Parameter named `p2c` PBES2 Count, which determines how many PBKDF2 iterations must be executed in order to derive a CEK wrapping key. The purpose of this parameter is to intentionally slow down the key derivation function in order to make password brute-force and dictionary attacks more expensive. This makes the PBES2 algorithms unsuitable for situations where the JWE is coming from an untrusted source: an adversary can intentionally pick an extremely high PBES2 Count value, that will initiate a CPU-bound computation that may take an unreasonable amount of time to finish. Under certain conditions, it is possible to have the user's environment consume unreasonable amount of CPU time. The impact is limited only to users utilizing the JWE decryption APIs with symmetric secrets to decrypt JWEs from untrusted parties who do not limit the accepted JWE Key Management Algorithms (`alg` Header Parameter) using the `keyManagementAlgorithms` (or `algorithms` in v1.x) decryption option or through other means. The `v1.28.2`, `v2.0.6`, `v3.20.4`, and `v4.9.2` releases limit the maximum PBKDF2 iteration count to `10000` by default. It is possible to adjust this limit with a newly introduced `maxPBES2Count` decryption option. If users are unable to upgrade their required library version, they have two options depending on whether they expect to receive JWEs using any of the three PBKDF2-based JWE key management algorithms. They can use the `keyManagementAlgorithms` decryption option to disable accepting PBKDF2 altogether, or they can inspect the JOSE Header prior to using the decryption API and limit the PBKDF2 iteration count (`p2c` Header Parameter). <p>Publish Date: 2022-09-07 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-36083>CVE-2022-36083</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/panva/jose/security/advisories/GHSA-jv3g-j58f-9mq9">https://github.com/panva/jose/security/advisories/GHSA-jv3g-j58f-9mq9</a></p> <p>Release Date: 2022-09-07</p> <p>Fix Resolution (jose): 2.0.6</p> <p>Direct dependency fix Resolution (@kubernetes/client-node): 0.12.3</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2022-36083 (Medium) detected in jose-2.0.5.tgz - ## CVE-2022-36083 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jose-2.0.5.tgz</b></p></summary> <p>JSON Web Almost Everything - JWA, JWS, JWE, JWK, JWT, JWKS for Node.js with minimal dependencies</p> <p>Library home page: <a href="https://registry.npmjs.org/jose/-/jose-2.0.5.tgz">https://registry.npmjs.org/jose/-/jose-2.0.5.tgz</a></p> <p>Path to dependency file: /components/crud-web-apps/volumes/frontend/package.json</p> <p>Path to vulnerable library: /components/crud-web-apps/volumes/frontend/node_modules/jose/package.json,/components/crud-web-apps/jupyter/frontend/node_modules/jose/package.json,/components/crud-web-apps/common/frontend/kubeflow-common-lib/node_modules/jose/package.json,/components/crud-web-apps/tensorboards/frontend/node_modules/jose/package.json</p> <p> Dependency Hierarchy: - client-node-0.12.2.tgz (Root Library) - openid-client-4.2.2.tgz - :x: **jose-2.0.5.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/DavidSpek/kubeflow/commit/00cbc9d11a3306fed1e979d79dff6ae36749d4bd">00cbc9d11a3306fed1e979d79dff6ae36749d4bd</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> JOSE is "JSON Web Almost Everything" - JWA, JWS, JWE, JWT, JWK, JWKS with no dependencies using runtime's native crypto in Node.js, Browser, Cloudflare Workers, Electron, and Deno. The PBKDF2-based JWE key management algorithms expect a JOSE Header Parameter named `p2c` PBES2 Count, which determines how many PBKDF2 iterations must be executed in order to derive a CEK wrapping key. The purpose of this parameter is to intentionally slow down the key derivation function in order to make password brute-force and dictionary attacks more expensive. This makes the PBES2 algorithms unsuitable for situations where the JWE is coming from an untrusted source: an adversary can intentionally pick an extremely high PBES2 Count value, that will initiate a CPU-bound computation that may take an unreasonable amount of time to finish. Under certain conditions, it is possible to have the user's environment consume unreasonable amount of CPU time. The impact is limited only to users utilizing the JWE decryption APIs with symmetric secrets to decrypt JWEs from untrusted parties who do not limit the accepted JWE Key Management Algorithms (`alg` Header Parameter) using the `keyManagementAlgorithms` (or `algorithms` in v1.x) decryption option or through other means. The `v1.28.2`, `v2.0.6`, `v3.20.4`, and `v4.9.2` releases limit the maximum PBKDF2 iteration count to `10000` by default. It is possible to adjust this limit with a newly introduced `maxPBES2Count` decryption option. If users are unable to upgrade their required library version, they have two options depending on whether they expect to receive JWEs using any of the three PBKDF2-based JWE key management algorithms. They can use the `keyManagementAlgorithms` decryption option to disable accepting PBKDF2 altogether, or they can inspect the JOSE Header prior to using the decryption API and limit the PBKDF2 iteration count (`p2c` Header Parameter). <p>Publish Date: 2022-09-07 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-36083>CVE-2022-36083</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/panva/jose/security/advisories/GHSA-jv3g-j58f-9mq9">https://github.com/panva/jose/security/advisories/GHSA-jv3g-j58f-9mq9</a></p> <p>Release Date: 2022-09-07</p> <p>Fix Resolution (jose): 2.0.6</p> <p>Direct dependency fix Resolution (@kubernetes/client-node): 0.12.3</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_architecture
cve medium detected in jose tgz cve medium severity vulnerability vulnerable library jose tgz json web almost everything jwa jws jwe jwk jwt jwks for node js with minimal dependencies library home page a href path to dependency file components crud web apps volumes frontend package json path to vulnerable library components crud web apps volumes frontend node modules jose package json components crud web apps jupyter frontend node modules jose package json components crud web apps common frontend kubeflow common lib node modules jose package json components crud web apps tensorboards frontend node modules jose package json dependency hierarchy client node tgz root library openid client tgz x jose tgz vulnerable library found in head commit a href found in base branch master vulnerability details jose is json web almost everything jwa jws jwe jwt jwk jwks with no dependencies using runtime s native crypto in node js browser cloudflare workers electron and deno the based jwe key management algorithms expect a jose header parameter named count which determines how many iterations must be executed in order to derive a cek wrapping key the purpose of this parameter is to intentionally slow down the key derivation function in order to make password brute force and dictionary attacks more expensive this makes the algorithms unsuitable for situations where the jwe is coming from an untrusted source an adversary can intentionally pick an extremely high count value that will initiate a cpu bound computation that may take an unreasonable amount of time to finish under certain conditions it is possible to have the user s environment consume unreasonable amount of cpu time the impact is limited only to users utilizing the jwe decryption apis with symmetric secrets to decrypt jwes from untrusted parties who do not limit the accepted jwe key management algorithms alg header parameter using the keymanagementalgorithms or algorithms in x decryption option or through other means the and releases limit the maximum iteration count to by default it is possible to adjust this limit with a newly introduced decryption option if users are unable to upgrade their required library version they have two options depending on whether they expect to receive jwes using any of the three based jwe key management algorithms they can use the keymanagementalgorithms decryption option to disable accepting altogether or they can inspect the jose header prior to using the decryption api and limit the iteration count header parameter publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution jose direct dependency fix resolution kubernetes client node step up your open source security game with mend
0
221,426
24,630,302,413
IssuesEvent
2022-10-17 01:00:10
MendDemo-josh/moby
https://api.github.com/repos/MendDemo-josh/moby
closed
libiberty9.1.0: 1 vulnerabilities (highest severity is: 7.5) - autoclosed
security vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>libiberty9.1.0</b></p></summary> <p> <p>Library home page: <a href=https://ftp.gnu.org/pub/gnu/gcc/gcc-9.1.0/?wsslib=libiberty>https://ftp.gnu.org/pub/gnu/gcc/gcc-9.1.0/?wsslib=libiberty</a></p> </p> </p></p> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> </p> <p></p></details> ## Vulnerabilities | CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in | Remediation Available | | ------------- | ------------- | ----- | ----- | ----- | --- | --- | | [CVE-2022-2879](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-2879) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | libiberty9.1.0 | Direct | go1.18.7,go1.19.2 | &#10060; | ## Details <details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2022-2879</summary> ### Vulnerable Library - <b>libiberty9.1.0</b></p> <p> <p>Library home page: <a href=https://ftp.gnu.org/pub/gnu/gcc/gcc-9.1.0/?wsslib=libiberty>https://ftp.gnu.org/pub/gnu/gcc/gcc-9.1.0/?wsslib=libiberty</a></p> <p>Found in base branch: <b>master</b></p></p> </p></p> ### Vulnerable Source Files (1) <p></p> <p> </p> <p></p> </p> <p></p> ### Vulnerability Details <p> Reader.Read does not set a limit on the maximum size of file headers. A maliciously crafted archive could cause Read to allocate unbounded amounts of memory, potentially causing resource exhaustion or panics. After fix, Reader.Read limits the maximum size of header blocks to 1 MiB. <p>Publish Date: 2022-10-14 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-2879>CVE-2022-2879</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>7.5</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://pkg.go.dev/vuln/GO-2022-1037">https://pkg.go.dev/vuln/GO-2022-1037</a></p> <p>Release Date: 2022-10-14</p> <p>Fix Resolution: go1.18.7,go1.19.2</p> </p> <p></p> </details>
True
libiberty9.1.0: 1 vulnerabilities (highest severity is: 7.5) - autoclosed - <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>libiberty9.1.0</b></p></summary> <p> <p>Library home page: <a href=https://ftp.gnu.org/pub/gnu/gcc/gcc-9.1.0/?wsslib=libiberty>https://ftp.gnu.org/pub/gnu/gcc/gcc-9.1.0/?wsslib=libiberty</a></p> </p> </p></p> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> </p> <p></p></details> ## Vulnerabilities | CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in | Remediation Available | | ------------- | ------------- | ----- | ----- | ----- | --- | --- | | [CVE-2022-2879](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-2879) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | libiberty9.1.0 | Direct | go1.18.7,go1.19.2 | &#10060; | ## Details <details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2022-2879</summary> ### Vulnerable Library - <b>libiberty9.1.0</b></p> <p> <p>Library home page: <a href=https://ftp.gnu.org/pub/gnu/gcc/gcc-9.1.0/?wsslib=libiberty>https://ftp.gnu.org/pub/gnu/gcc/gcc-9.1.0/?wsslib=libiberty</a></p> <p>Found in base branch: <b>master</b></p></p> </p></p> ### Vulnerable Source Files (1) <p></p> <p> </p> <p></p> </p> <p></p> ### Vulnerability Details <p> Reader.Read does not set a limit on the maximum size of file headers. A maliciously crafted archive could cause Read to allocate unbounded amounts of memory, potentially causing resource exhaustion or panics. After fix, Reader.Read limits the maximum size of header blocks to 1 MiB. <p>Publish Date: 2022-10-14 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-2879>CVE-2022-2879</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>7.5</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://pkg.go.dev/vuln/GO-2022-1037">https://pkg.go.dev/vuln/GO-2022-1037</a></p> <p>Release Date: 2022-10-14</p> <p>Fix Resolution: go1.18.7,go1.19.2</p> </p> <p></p> </details>
non_architecture
vulnerabilities highest severity is autoclosed vulnerable library library home page a href vulnerable source files vulnerabilities cve severity cvss dependency type fixed in remediation available high direct details cve vulnerable library library home page a href found in base branch master vulnerable source files vulnerability details reader read does not set a limit on the maximum size of file headers a maliciously crafted archive could cause read to allocate unbounded amounts of memory potentially causing resource exhaustion or panics after fix reader read limits the maximum size of header blocks to mib publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution
0
619
3,862,146,688
IssuesEvent
2016-04-08 00:43:28
rails-oceania/melbourne-ruby
https://api.github.com/repos/rails-oceania/melbourne-ruby
closed
Event Sourcing, part 2
40min architecture code intermediate presentation
I'd like to follow up my RubyConf talk which was meant to make people curious about Event Sourcing, with a talk about some of the practicalities and where to start. For those who missed my RubyConf talk, I'll quickly recap that before launching into particulars. I'd allow for 45 mins for this talk.
1.0
Event Sourcing, part 2 - I'd like to follow up my RubyConf talk which was meant to make people curious about Event Sourcing, with a talk about some of the practicalities and where to start. For those who missed my RubyConf talk, I'll quickly recap that before launching into particulars. I'd allow for 45 mins for this talk.
architecture
event sourcing part i d like to follow up my rubyconf talk which was meant to make people curious about event sourcing with a talk about some of the practicalities and where to start for those who missed my rubyconf talk i ll quickly recap that before launching into particulars i d allow for mins for this talk
1
10,480
27,022,476,040
IssuesEvent
2023-02-11 06:36:04
jsolly/blogthedata
https://api.github.com/repos/jsolly/blogthedata
closed
Separate Porfolio page into its own page, disconnecting it from being a category
Architecture SEO
#### Context The current implementation of the portfolio page is quite hacky. It's technically a 'category' which causes all kinds of wonkiness because I am having to add lots of conditional logic inside categories.html in order to handle special cases on the portfolio page. I also have to have more logic for meta tags as the portfolio page has it's own tags that are unique from the other category pages. #### Ideal behavior The portfolio page is on it's own. It's not a 'category' of posts. #### Things to consider Will have to refactor the how posts are brought onto a page that is not a category Might need to do a database migration to remove the 'category' attribute from portfolio posts. Will need to refactor the category templates to remove the special portfolio logic.
1.0
Separate Porfolio page into its own page, disconnecting it from being a category - #### Context The current implementation of the portfolio page is quite hacky. It's technically a 'category' which causes all kinds of wonkiness because I am having to add lots of conditional logic inside categories.html in order to handle special cases on the portfolio page. I also have to have more logic for meta tags as the portfolio page has it's own tags that are unique from the other category pages. #### Ideal behavior The portfolio page is on it's own. It's not a 'category' of posts. #### Things to consider Will have to refactor the how posts are brought onto a page that is not a category Might need to do a database migration to remove the 'category' attribute from portfolio posts. Will need to refactor the category templates to remove the special portfolio logic.
architecture
separate porfolio page into its own page disconnecting it from being a category context the current implementation of the portfolio page is quite hacky it s technically a category which causes all kinds of wonkiness because i am having to add lots of conditional logic inside categories html in order to handle special cases on the portfolio page i also have to have more logic for meta tags as the portfolio page has it s own tags that are unique from the other category pages ideal behavior the portfolio page is on it s own it s not a category of posts things to consider will have to refactor the how posts are brought onto a page that is not a category might need to do a database migration to remove the category attribute from portfolio posts will need to refactor the category templates to remove the special portfolio logic
1
1,169
5,221,420,375
IssuesEvent
2017-01-27 01:28:11
jung-digital/ringa
https://api.github.com/repos/jung-digital/ringa
opened
Command: add no wait operator [[]]
architecture
Target Code: ``` controller.addListener([ Command1, [[Command2]], Command3 ]); ``` Command 1 should run, Command2 should be started, but then Command3 should be run immediately without waiting for Command2 to finish.
1.0
Command: add no wait operator [[]] - Target Code: ``` controller.addListener([ Command1, [[Command2]], Command3 ]); ``` Command 1 should run, Command2 should be started, but then Command3 should be run immediately without waiting for Command2 to finish.
architecture
command add no wait operator target code controller addlistener command should run should be started but then should be run immediately without waiting for to finish
1
9,819
25,289,125,063
IssuesEvent
2022-11-16 22:07:07
spring-projects/sts4
https://api.github.com/repos/spring-projects/sts4
closed
exception thrown in VSCode when using latest snapshots
type: bug status: needs-investigation theme: internal-architecture for: vscode theme: refactoring
I am using VSCode with the latest pre-releases from: - Language Support for Java - Spring Boot Dashboard - Spring Boot Tools (from the latest VSIX file) I have a project open in my workspace, created from initializr (web + actuator), on Spring Boot 2.6.12. After a little while, I see an error popup showing up, complaining about a problem when asking for `textDocument/codeAction`, and showing this exception in the log output: ``` java.util.concurrent.CompletionException: com.google.gson.JsonSyntaxException: java.lang.IllegalStateException: Expected BEGIN_ARRAY but was BEGIN_OBJECT at path $ at java.base/java.util.concurrent.CompletableFuture.encodeThrowable(Unknown Source) at java.base/java.util.concurrent.CompletableFuture.completeThrowable(Unknown Source) at java.base/java.util.concurrent.CompletableFuture$UniApply.tryFire(Unknown Source) at java.base/java.util.concurrent.CompletableFuture$Completion.run(Unknown Source) at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source) at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source) at java.base/java.lang.Thread.run(Unknown Source) Caused by: com.google.gson.JsonSyntaxException: java.lang.IllegalStateException: Expected BEGIN_ARRAY but was BEGIN_OBJECT at path $ at com.google.gson.Gson.fromJson(Gson.java:1070) at com.google.gson.Gson.fromJson(Gson.java:1129) at org.springframework.ide.vscode.commons.languageserver.util.SimpleTextDocumentService.lambda$computeCodeActions$10(SimpleTextDocumentService.java:445) at java.base/java.util.ArrayList.forEach(Unknown Source) at org.springframework.ide.vscode.commons.languageserver.util.SimpleTextDocumentService.computeCodeActions(SimpleTextDocumentService.java:442) at org.springframework.ide.vscode.commons.languageserver.util.SimpleTextDocumentService.lambda$codeAction$11(SimpleTextDocumentService.java:499) ... 5 more Caused by: java.lang.IllegalStateException: Expected BEGIN_ARRAY but was BEGIN_OBJECT at path $ at com.google.gson.internal.bind.JsonTreeReader.expect(JsonTreeReader.java:163) at com.google.gson.internal.bind.JsonTreeReader.beginArray(JsonTreeReader.java:72) at com.google.gson.internal.bind.CollectionTypeAdapterFactory$Adapter.read(CollectionTypeAdapterFactory.java:80) at com.google.gson.internal.bind.CollectionTypeAdapterFactory$Adapter.read(CollectionTypeAdapterFactory.java:61) at com.google.gson.Gson.fromJson(Gson.java:1058) ... 10 more ```
1.0
exception thrown in VSCode when using latest snapshots - I am using VSCode with the latest pre-releases from: - Language Support for Java - Spring Boot Dashboard - Spring Boot Tools (from the latest VSIX file) I have a project open in my workspace, created from initializr (web + actuator), on Spring Boot 2.6.12. After a little while, I see an error popup showing up, complaining about a problem when asking for `textDocument/codeAction`, and showing this exception in the log output: ``` java.util.concurrent.CompletionException: com.google.gson.JsonSyntaxException: java.lang.IllegalStateException: Expected BEGIN_ARRAY but was BEGIN_OBJECT at path $ at java.base/java.util.concurrent.CompletableFuture.encodeThrowable(Unknown Source) at java.base/java.util.concurrent.CompletableFuture.completeThrowable(Unknown Source) at java.base/java.util.concurrent.CompletableFuture$UniApply.tryFire(Unknown Source) at java.base/java.util.concurrent.CompletableFuture$Completion.run(Unknown Source) at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source) at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source) at java.base/java.lang.Thread.run(Unknown Source) Caused by: com.google.gson.JsonSyntaxException: java.lang.IllegalStateException: Expected BEGIN_ARRAY but was BEGIN_OBJECT at path $ at com.google.gson.Gson.fromJson(Gson.java:1070) at com.google.gson.Gson.fromJson(Gson.java:1129) at org.springframework.ide.vscode.commons.languageserver.util.SimpleTextDocumentService.lambda$computeCodeActions$10(SimpleTextDocumentService.java:445) at java.base/java.util.ArrayList.forEach(Unknown Source) at org.springframework.ide.vscode.commons.languageserver.util.SimpleTextDocumentService.computeCodeActions(SimpleTextDocumentService.java:442) at org.springframework.ide.vscode.commons.languageserver.util.SimpleTextDocumentService.lambda$codeAction$11(SimpleTextDocumentService.java:499) ... 5 more Caused by: java.lang.IllegalStateException: Expected BEGIN_ARRAY but was BEGIN_OBJECT at path $ at com.google.gson.internal.bind.JsonTreeReader.expect(JsonTreeReader.java:163) at com.google.gson.internal.bind.JsonTreeReader.beginArray(JsonTreeReader.java:72) at com.google.gson.internal.bind.CollectionTypeAdapterFactory$Adapter.read(CollectionTypeAdapterFactory.java:80) at com.google.gson.internal.bind.CollectionTypeAdapterFactory$Adapter.read(CollectionTypeAdapterFactory.java:61) at com.google.gson.Gson.fromJson(Gson.java:1058) ... 10 more ```
architecture
exception thrown in vscode when using latest snapshots i am using vscode with the latest pre releases from language support for java spring boot dashboard spring boot tools from the latest vsix file i have a project open in my workspace created from initializr web actuator on spring boot after a little while i see an error popup showing up complaining about a problem when asking for textdocument codeaction and showing this exception in the log output java util concurrent completionexception com google gson jsonsyntaxexception java lang illegalstateexception expected begin array but was begin object at path at java base java util concurrent completablefuture encodethrowable unknown source at java base java util concurrent completablefuture completethrowable unknown source at java base java util concurrent completablefuture uniapply tryfire unknown source at java base java util concurrent completablefuture completion run unknown source at java base java util concurrent threadpoolexecutor runworker unknown source at java base java util concurrent threadpoolexecutor worker run unknown source at java base java lang thread run unknown source caused by com google gson jsonsyntaxexception java lang illegalstateexception expected begin array but was begin object at path at com google gson gson fromjson gson java at com google gson gson fromjson gson java at org springframework ide vscode commons languageserver util simpletextdocumentservice lambda computecodeactions simpletextdocumentservice java at java base java util arraylist foreach unknown source at org springframework ide vscode commons languageserver util simpletextdocumentservice computecodeactions simpletextdocumentservice java at org springframework ide vscode commons languageserver util simpletextdocumentservice lambda codeaction simpletextdocumentservice java more caused by java lang illegalstateexception expected begin array but was begin object at path at com google gson internal bind jsontreereader expect jsontreereader java at com google gson internal bind jsontreereader beginarray jsontreereader java at com google gson internal bind collectiontypeadapterfactory adapter read collectiontypeadapterfactory java at com google gson internal bind collectiontypeadapterfactory adapter read collectiontypeadapterfactory java at com google gson gson fromjson gson java more
1
3,954
10,344,295,967
IssuesEvent
2019-09-04 10:52:21
open-zaak/open-zaak
https://api.github.com/repos/open-zaak/open-zaak
closed
As stakeholder, I want to have the Authorizations API exposed on OpenZaak
EPIC: Architecture
... so applications can request their permissions (and theoretically, set permissions). **Description** The Authorization API was left out of scope in #3. Which might or might not be the best choice, depending on #3. This US makes sure it gets in.
1.0
As stakeholder, I want to have the Authorizations API exposed on OpenZaak - ... so applications can request their permissions (and theoretically, set permissions). **Description** The Authorization API was left out of scope in #3. Which might or might not be the best choice, depending on #3. This US makes sure it gets in.
architecture
as stakeholder i want to have the authorizations api exposed on openzaak so applications can request their permissions and theoretically set permissions description the authorization api was left out of scope in which might or might not be the best choice depending on this us makes sure it gets in
1
7,907
19,916,085,929
IssuesEvent
2022-01-25 22:53:47
MicrosoftDocs/architecture-center
https://api.github.com/repos/MicrosoftDocs/architecture-center
closed
this sentence does not make sense...
doc-bug cxp triaged architecture-center/svc reference-architecture/subsvc Pri2
[Enter feedback here] From this article: https://docs.microsoft.com/en-us/azure/architecture/reference-architectures/dmz/nva-ha This sentence does not make sense / is not understandable - Since HA Ports for inbound traffic every individual TCP/UDP port needs to be opened in a dedicated load-balancing rule. You may have meant something like Since HA ports control (or restrict) inbound traffic, every individual TCP/UDP port needs to be opened in a dedicated load-balancing rule. --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 9bab6a90-43a3-3d83-d680-26683a3d833d * Version Independent ID: f8eea094-d297-93e6-06ac-7652d059734f * Content: [Deploy highly available NVAs - Azure Architecture Center](https://docs.microsoft.com/en-us/azure/architecture/reference-architectures/dmz/nva-ha) * Content Source: [docs/reference-architectures/dmz/nva-ha.yml](https://github.com/microsoftdocs/architecture-center/blob/main/docs/reference-architectures/dmz/nva-ha.yml) * Service: **architecture-center** * Sub-service: **reference-architecture** * GitHub Login: @telmosampaio * Microsoft Alias: **pnp**
2.0
this sentence does not make sense... - [Enter feedback here] From this article: https://docs.microsoft.com/en-us/azure/architecture/reference-architectures/dmz/nva-ha This sentence does not make sense / is not understandable - Since HA Ports for inbound traffic every individual TCP/UDP port needs to be opened in a dedicated load-balancing rule. You may have meant something like Since HA ports control (or restrict) inbound traffic, every individual TCP/UDP port needs to be opened in a dedicated load-balancing rule. --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 9bab6a90-43a3-3d83-d680-26683a3d833d * Version Independent ID: f8eea094-d297-93e6-06ac-7652d059734f * Content: [Deploy highly available NVAs - Azure Architecture Center](https://docs.microsoft.com/en-us/azure/architecture/reference-architectures/dmz/nva-ha) * Content Source: [docs/reference-architectures/dmz/nva-ha.yml](https://github.com/microsoftdocs/architecture-center/blob/main/docs/reference-architectures/dmz/nva-ha.yml) * Service: **architecture-center** * Sub-service: **reference-architecture** * GitHub Login: @telmosampaio * Microsoft Alias: **pnp**
architecture
this sentence does not make sense from this article this sentence does not make sense is not understandable since ha ports for inbound traffic every individual tcp udp port needs to be opened in a dedicated load balancing rule you may have meant something like since ha ports control or restrict inbound traffic every individual tcp udp port needs to be opened in a dedicated load balancing rule document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source service architecture center sub service reference architecture github login telmosampaio microsoft alias pnp
1
3,453
9,645,281,781
IssuesEvent
2019-05-17 08:15:13
dotnet/docs
https://api.github.com/repos/dotnet/docs
closed
Non working links
:book: guide - .NET Microservices :books: Area - .NET Guide :card_file_box: Technology - .NET Architecture Source - Docs.ms broken-link doc-bug
codebetter.com has been down for some time. --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 84d83855-4370-90e8-024a-12c5ac8220e9 * Version Independent ID: 01be3db1-4a00-c8f2-aa59-59656f206b93 * Content: [Applying CQRS and CQS approaches in a DDD microservice in eShopOnContainers](https://docs.microsoft.com/en-us/dotnet/standard/microservices-architecture/microservice-ddd-cqrs-patterns/eshoponcontainers-cqrs-ddd-microservice#feedback) * Content Source: [docs/standard/microservices-architecture/microservice-ddd-cqrs-patterns/eshoponcontainers-cqrs-ddd-microservice.md](https://github.com/dotnet/docs/blob/master/docs/standard/microservices-architecture/microservice-ddd-cqrs-patterns/eshoponcontainers-cqrs-ddd-microservice.md) * Product: **dotnet** * Technology: **dotnet-ebooks** * GitHub Login: @CESARDELATORRE * Microsoft Alias: **wiwagn**
1.0
Non working links - codebetter.com has been down for some time. --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 84d83855-4370-90e8-024a-12c5ac8220e9 * Version Independent ID: 01be3db1-4a00-c8f2-aa59-59656f206b93 * Content: [Applying CQRS and CQS approaches in a DDD microservice in eShopOnContainers](https://docs.microsoft.com/en-us/dotnet/standard/microservices-architecture/microservice-ddd-cqrs-patterns/eshoponcontainers-cqrs-ddd-microservice#feedback) * Content Source: [docs/standard/microservices-architecture/microservice-ddd-cqrs-patterns/eshoponcontainers-cqrs-ddd-microservice.md](https://github.com/dotnet/docs/blob/master/docs/standard/microservices-architecture/microservice-ddd-cqrs-patterns/eshoponcontainers-cqrs-ddd-microservice.md) * Product: **dotnet** * Technology: **dotnet-ebooks** * GitHub Login: @CESARDELATORRE * Microsoft Alias: **wiwagn**
architecture
non working links codebetter com has been down for some time document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source product dotnet technology dotnet ebooks github login cesardelatorre microsoft alias wiwagn
1
207,397
7,127,893,407
IssuesEvent
2018-01-21 04:06:00
OperationCode/operationcode_frontend
https://api.github.com/repos/OperationCode/operationcode_frontend
closed
Add photos/bio to /team page
Priority: Medium Status: In Progress Type: Feature
# Feature ## Why is this feature being added? We should make it easy to find board members on our `/team` page. Suggest we break down this page, ex. #contributors, #chapter-leaders, #board, #advisors, etc., and include photos so we can recognize them. Here's an example of what I'm envisioning (via GitHub): ![images](https://user-images.githubusercontent.com/1848616/28241705-4a3588e8-694e-11e7-9d20-d6eac0b9d94e.png) *could be round images instead of square* - [ ] There should be a `+` sign, or another icon so funders and other interested stakeholders can read up on their bios, ex. `operationcode.org/team/hollomancer` and learn more about Conrad, including GitHub, Twitter handle. - [ ] Needs to be easy to send the link to x foundation or x corporation, and say, "Meet some of our board members and link to the exact spot where board members are listed. ## Technical Requirements - [ ] Refactor StaffCard to make BoardCard obsolete, such that these components only render what they're provided. - [ ] Resize images for uniformity. - [ ] Make image links relative.
1.0
Add photos/bio to /team page - # Feature ## Why is this feature being added? We should make it easy to find board members on our `/team` page. Suggest we break down this page, ex. #contributors, #chapter-leaders, #board, #advisors, etc., and include photos so we can recognize them. Here's an example of what I'm envisioning (via GitHub): ![images](https://user-images.githubusercontent.com/1848616/28241705-4a3588e8-694e-11e7-9d20-d6eac0b9d94e.png) *could be round images instead of square* - [ ] There should be a `+` sign, or another icon so funders and other interested stakeholders can read up on their bios, ex. `operationcode.org/team/hollomancer` and learn more about Conrad, including GitHub, Twitter handle. - [ ] Needs to be easy to send the link to x foundation or x corporation, and say, "Meet some of our board members and link to the exact spot where board members are listed. ## Technical Requirements - [ ] Refactor StaffCard to make BoardCard obsolete, such that these components only render what they're provided. - [ ] Resize images for uniformity. - [ ] Make image links relative.
non_architecture
add photos bio to team page feature why is this feature being added we should make it easy to find board members on our team page suggest we break down this page ex contributors chapter leaders board advisors etc and include photos so we can recognize them here s an example of what i m envisioning via github could be round images instead of square there should be a sign or another icon so funders and other interested stakeholders can read up on their bios ex operationcode org team hollomancer and learn more about conrad including github twitter handle needs to be easy to send the link to x foundation or x corporation and say meet some of our board members and link to the exact spot where board members are listed technical requirements refactor staffcard to make boardcard obsolete such that these components only render what they re provided resize images for uniformity make image links relative
0
59,918
14,671,978,283
IssuesEvent
2020-12-30 09:31:08
Raku/old-issue-tracker
https://api.github.com/repos/Raku/old-issue-tracker
closed
MoarVM build fail on termux/Android 6.0.1
build
Migrated from [rt.perl.org#132785](https://rt-archive.perl.org/perl6/Ticket/Display.html?id=132785) (status was 'new') Searchable as RT132785$
1.0
MoarVM build fail on termux/Android 6.0.1 - Migrated from [rt.perl.org#132785](https://rt-archive.perl.org/perl6/Ticket/Display.html?id=132785) (status was 'new') Searchable as RT132785$
non_architecture
moarvm build fail on termux android migrated from status was new searchable as
0
242,407
7,841,901,318
IssuesEvent
2018-06-18 21:10:00
StrangeLoopGames/EcoIssues
https://api.github.com/repos/StrangeLoopGames/EcoIssues
closed
Website not loading additional layers (Linux-Server)
Medium Priority
Hi there, I'm running my Server on Linux and since 0.6.2.5-beta, the map in the World Status page can't display additional layers (like AirPollutionSpread). Tested with Chrome (v 66) and Edge Browser How to reproduce: Open the website and wait for map loaded -> Click on Select map layer and select something (happens for all layers) -> Map get's a little bit darker but nothing else changes (but error occurs in Chrome Browser console). Server versions: Mono linux standalone with 0.6.2.5, 0.7.3.3, 0.7.4.2 Edge Browser is not displaying ocean Log from Chrome console when switching to AirPollutionSpread layer: ``` ecomap.js:419 Starting to parse AirPollutionSpread ecomap.js:112 Uncaught TypeError: Cannot read property 'concat' of undefined at lzwDecode (ecomap.js:112) at parseImg (ecomap.js:311) at parseBlock (ecomap.js:331) ``` Default view without selected layer: ![2018-05-12 12_07_53-eco-worldmap](https://user-images.githubusercontent.com/1303974/39956322-510c3374-55de-11e8-8d8d-9aa46b9db1fe.png) View with selected layer: ![2018-05-12 12_15_58-eco-layerselected](https://user-images.githubusercontent.com/1303974/39956323-51248348-55de-11e8-855c-66ece023dd3a.png) Default view in Microsoft Edge: ![2018-05-12 12_16_11-microsoft edge](https://user-images.githubusercontent.com/1303974/39956324-513ff146-55de-11e8-9065-833a50563d5f.png)
1.0
Website not loading additional layers (Linux-Server) - Hi there, I'm running my Server on Linux and since 0.6.2.5-beta, the map in the World Status page can't display additional layers (like AirPollutionSpread). Tested with Chrome (v 66) and Edge Browser How to reproduce: Open the website and wait for map loaded -> Click on Select map layer and select something (happens for all layers) -> Map get's a little bit darker but nothing else changes (but error occurs in Chrome Browser console). Server versions: Mono linux standalone with 0.6.2.5, 0.7.3.3, 0.7.4.2 Edge Browser is not displaying ocean Log from Chrome console when switching to AirPollutionSpread layer: ``` ecomap.js:419 Starting to parse AirPollutionSpread ecomap.js:112 Uncaught TypeError: Cannot read property 'concat' of undefined at lzwDecode (ecomap.js:112) at parseImg (ecomap.js:311) at parseBlock (ecomap.js:331) ``` Default view without selected layer: ![2018-05-12 12_07_53-eco-worldmap](https://user-images.githubusercontent.com/1303974/39956322-510c3374-55de-11e8-8d8d-9aa46b9db1fe.png) View with selected layer: ![2018-05-12 12_15_58-eco-layerselected](https://user-images.githubusercontent.com/1303974/39956323-51248348-55de-11e8-855c-66ece023dd3a.png) Default view in Microsoft Edge: ![2018-05-12 12_16_11-microsoft edge](https://user-images.githubusercontent.com/1303974/39956324-513ff146-55de-11e8-9065-833a50563d5f.png)
non_architecture
website not loading additional layers linux server hi there i m running my server on linux and since beta the map in the world status page can t display additional layers like airpollutionspread tested with chrome v and edge browser how to reproduce open the website and wait for map loaded click on select map layer and select something happens for all layers map get s a little bit darker but nothing else changes but error occurs in chrome browser console server versions mono linux standalone with edge browser is not displaying ocean log from chrome console when switching to airpollutionspread layer ecomap js starting to parse airpollutionspread ecomap js uncaught typeerror cannot read property concat of undefined at lzwdecode ecomap js at parseimg ecomap js at parseblock ecomap js default view without selected layer view with selected layer default view in microsoft edge
0
577,916
17,139,172,640
IssuesEvent
2021-07-13 07:40:29
googleapis/java-bigtable-hbase
https://api.github.com/repos/googleapis/java-bigtable-hbase
opened
bigtable.hbase.TestBufferedMutator: testAutoFlushOff failed
flakybot: issue priority: p1 type: bug
This test failed! To configure my behavior, see [the Flaky Bot documentation](https://github.com/googleapis/repo-automation-bots/tree/master/packages/flakybot). If I'm commenting on this issue too often, add the `flakybot: quiet` label and I will stop commenting. --- commit: a891335ce3179c45fade4f3683b7e09d38d0107a buildURL: [Build Status](https://source.cloud.google.com/results/invocations/64905140-8b04-4270-9ed2-a91e73d97e39), [Sponge](http://sponge2/64905140-8b04-4270-9ed2-a91e73d97e39) status: failed <details><summary>Test output</summary><br><pre>org.apache.hadoop.hbase.client.RetriesExhaustedWithDetailsException: Failed 1 action: UnauthenticatedException: 1 time, servers with issues: bigtable.googleapis.com at com.google.cloud.bigtable.hbase.BigtableBufferedMutator.getExceptions(BigtableBufferedMutator.java:188) at com.google.cloud.bigtable.hbase.BigtableBufferedMutator.handleExceptions(BigtableBufferedMutator.java:142) at com.google.cloud.bigtable.hbase.BigtableBufferedMutator.flush(BigtableBufferedMutator.java:93) at com.google.cloud.bigtable.hbase.TestBufferedMutator.testAutoFlushOff(TestBufferedMutator.java:62) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59) at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56) at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17) at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) at org.junit.runners.BlockJUnit4ClassRunner$1.evaluate(BlockJUnit4ClassRunner.java:100) at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:366) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:103) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:63) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.apache.maven.surefire.junitcore.pc.Scheduler$1.run(Scheduler.java:410) at org.apache.maven.surefire.junitcore.pc.InvokerStrategy.schedule(InvokerStrategy.java:54) at org.apache.maven.surefire.junitcore.pc.Scheduler.schedule(Scheduler.java:367) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) at org.junit.runners.ParentRunner.run(ParentRunner.java:413) at org.junit.runners.Suite.runChild(Suite.java:128) at org.junit.runners.Suite.runChild(Suite.java:27) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.apache.maven.surefire.junitcore.pc.Scheduler$1.run(Scheduler.java:410) at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) at java.util.concurrent.FutureTask.run(FutureTask.java:266) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at java.lang.Thread.run(Thread.java:748) Suppressed: com.google.api.gax.batching.BatchingException: Batching finished with 1 batches failed to apply due to: 1 ApiException(1 INTERNAL) and 0 partial failures. at com.google.api.gax.batching.BatcherStats.asException(BatcherStats.java:147) at com.google.api.gax.batching.BatcherImpl.close(BatcherImpl.java:290) at com.google.cloud.bigtable.hbase.wrappers.veneer.BulkMutationVeneerApi.close(BulkMutationVeneerApi.java:68) at com.google.cloud.bigtable.hbase.BigtableBufferedMutatorHelper.close(BigtableBufferedMutatorHelper.java:91) at com.google.cloud.bigtable.hbase.BigtableBufferedMutator.close(BigtableBufferedMutator.java:85) at com.google.cloud.bigtable.hbase.TestBufferedMutator.testAutoFlushOff(TestBufferedMutator.java:64) ... 31 more </pre></details>
1.0
bigtable.hbase.TestBufferedMutator: testAutoFlushOff failed - This test failed! To configure my behavior, see [the Flaky Bot documentation](https://github.com/googleapis/repo-automation-bots/tree/master/packages/flakybot). If I'm commenting on this issue too often, add the `flakybot: quiet` label and I will stop commenting. --- commit: a891335ce3179c45fade4f3683b7e09d38d0107a buildURL: [Build Status](https://source.cloud.google.com/results/invocations/64905140-8b04-4270-9ed2-a91e73d97e39), [Sponge](http://sponge2/64905140-8b04-4270-9ed2-a91e73d97e39) status: failed <details><summary>Test output</summary><br><pre>org.apache.hadoop.hbase.client.RetriesExhaustedWithDetailsException: Failed 1 action: UnauthenticatedException: 1 time, servers with issues: bigtable.googleapis.com at com.google.cloud.bigtable.hbase.BigtableBufferedMutator.getExceptions(BigtableBufferedMutator.java:188) at com.google.cloud.bigtable.hbase.BigtableBufferedMutator.handleExceptions(BigtableBufferedMutator.java:142) at com.google.cloud.bigtable.hbase.BigtableBufferedMutator.flush(BigtableBufferedMutator.java:93) at com.google.cloud.bigtable.hbase.TestBufferedMutator.testAutoFlushOff(TestBufferedMutator.java:62) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59) at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56) at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17) at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) at org.junit.runners.BlockJUnit4ClassRunner$1.evaluate(BlockJUnit4ClassRunner.java:100) at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:366) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:103) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:63) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.apache.maven.surefire.junitcore.pc.Scheduler$1.run(Scheduler.java:410) at org.apache.maven.surefire.junitcore.pc.InvokerStrategy.schedule(InvokerStrategy.java:54) at org.apache.maven.surefire.junitcore.pc.Scheduler.schedule(Scheduler.java:367) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) at org.junit.runners.ParentRunner.run(ParentRunner.java:413) at org.junit.runners.Suite.runChild(Suite.java:128) at org.junit.runners.Suite.runChild(Suite.java:27) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.apache.maven.surefire.junitcore.pc.Scheduler$1.run(Scheduler.java:410) at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) at java.util.concurrent.FutureTask.run(FutureTask.java:266) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at java.lang.Thread.run(Thread.java:748) Suppressed: com.google.api.gax.batching.BatchingException: Batching finished with 1 batches failed to apply due to: 1 ApiException(1 INTERNAL) and 0 partial failures. at com.google.api.gax.batching.BatcherStats.asException(BatcherStats.java:147) at com.google.api.gax.batching.BatcherImpl.close(BatcherImpl.java:290) at com.google.cloud.bigtable.hbase.wrappers.veneer.BulkMutationVeneerApi.close(BulkMutationVeneerApi.java:68) at com.google.cloud.bigtable.hbase.BigtableBufferedMutatorHelper.close(BigtableBufferedMutatorHelper.java:91) at com.google.cloud.bigtable.hbase.BigtableBufferedMutator.close(BigtableBufferedMutator.java:85) at com.google.cloud.bigtable.hbase.TestBufferedMutator.testAutoFlushOff(TestBufferedMutator.java:64) ... 31 more </pre></details>
non_architecture
bigtable hbase testbufferedmutator testautoflushoff failed this test failed to configure my behavior see if i m commenting on this issue too often add the flakybot quiet label and i will stop commenting commit buildurl status failed test output org apache hadoop hbase client retriesexhaustedwithdetailsexception failed action unauthenticatedexception time servers with issues bigtable googleapis com at com google cloud bigtable hbase bigtablebufferedmutator getexceptions bigtablebufferedmutator java at com google cloud bigtable hbase bigtablebufferedmutator handleexceptions bigtablebufferedmutator java at com google cloud bigtable hbase bigtablebufferedmutator flush bigtablebufferedmutator java at com google cloud bigtable hbase testbufferedmutator testautoflushoff testbufferedmutator java at sun reflect nativemethodaccessorimpl native method at sun reflect nativemethodaccessorimpl invoke nativemethodaccessorimpl java at sun reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java lang reflect method invoke method java at org junit runners model frameworkmethod runreflectivecall frameworkmethod java at org junit internal runners model reflectivecallable run reflectivecallable java at org junit runners model frameworkmethod invokeexplosively frameworkmethod java at org junit internal runners statements invokemethod evaluate invokemethod java at org junit runners parentrunner evaluate parentrunner java at org junit runners evaluate java at org junit runners parentrunner runleaf parentrunner java at org junit runners runchild java at org junit runners runchild java at org junit runners parentrunner run parentrunner java at org apache maven surefire junitcore pc scheduler run scheduler java at org apache maven surefire junitcore pc invokerstrategy schedule invokerstrategy java at org apache maven surefire junitcore pc scheduler schedule scheduler java at org junit runners parentrunner runchildren parentrunner java at org junit runners parentrunner access parentrunner java at org junit runners parentrunner evaluate parentrunner java at org junit runners parentrunner evaluate parentrunner java at org junit runners parentrunner run parentrunner java at org junit runners suite runchild suite java at org junit runners suite runchild suite java at org junit runners parentrunner run parentrunner java at org apache maven surefire junitcore pc scheduler run scheduler java at java util concurrent executors runnableadapter call executors java at java util concurrent futuretask run futuretask java at java util concurrent threadpoolexecutor runworker threadpoolexecutor java at java util concurrent threadpoolexecutor worker run threadpoolexecutor java at java lang thread run thread java suppressed com google api gax batching batchingexception batching finished with batches failed to apply due to apiexception internal and partial failures at com google api gax batching batcherstats asexception batcherstats java at com google api gax batching batcherimpl close batcherimpl java at com google cloud bigtable hbase wrappers veneer bulkmutationveneerapi close bulkmutationveneerapi java at com google cloud bigtable hbase bigtablebufferedmutatorhelper close bigtablebufferedmutatorhelper java at com google cloud bigtable hbase bigtablebufferedmutator close bigtablebufferedmutator java at com google cloud bigtable hbase testbufferedmutator testautoflushoff testbufferedmutator java more
0
257,267
27,561,833,137
IssuesEvent
2023-03-07 22:49:14
samqws-marketing/coursera_naptime
https://api.github.com/repos/samqws-marketing/coursera_naptime
closed
CVE-2020-36185 (High) detected in multiple libraries - autoclosed
Mend: dependency security vulnerability
## CVE-2020-36185 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>jackson-databind-2.9.0.jar</b>, <b>jackson-databind-2.8.11.4.jar</b>, <b>jackson-databind-2.3.3.jar</b></p></summary> <p> <details><summary><b>jackson-databind-2.9.0.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to vulnerable library: /home/wss-scanner/.ivy2/cache/com.fasterxml.jackson.core/jackson-databind/bundles/jackson-databind-2.9.0.jar</p> <p> Dependency Hierarchy: - play-ehcache_2.12-2.6.25.jar (Root Library) - play_2.12-2.6.25.jar - play-json_2.12-2.6.14.jar - jackson-datatype-jdk8-2.8.11.jar - :x: **jackson-databind-2.9.0.jar** (Vulnerable Library) </details> <details><summary><b>jackson-databind-2.8.11.4.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to vulnerable library: /home/wss-scanner/.ivy2/cache/com.fasterxml.jackson.core/jackson-databind/bundles/jackson-databind-2.8.11.4.jar</p> <p> Dependency Hierarchy: - play-ehcache_2.12-2.6.25.jar (Root Library) - play_2.12-2.6.25.jar - :x: **jackson-databind-2.8.11.4.jar** (Vulnerable Library) </details> <details><summary><b>jackson-databind-2.3.3.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Path to vulnerable library: /home/wss-scanner/.ivy2/cache/com.fasterxml.jackson.core/jackson-databind/bundles/jackson-databind-2.3.3.jar</p> <p> Dependency Hierarchy: - sbt-plugin-2.4.4.jar (Root Library) - sbt-js-engine-1.1.3.jar - npm_2.10-1.1.1.jar - webjars-locator-0.26.jar - :x: **jackson-databind-2.3.3.jar** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/samqws-marketing/coursera_naptime/commit/95750513b615ecf0ea9b7e14fb5f71e577d01a1f">95750513b615ecf0ea9b7e14fb5f71e577d01a1f</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> FasterXML jackson-databind 2.x before 2.9.10.8 mishandles the interaction between serialization gadgets and typing, related to org.apache.tomcat.dbcp.dbcp2.datasources.SharedPoolDataSource. <p>Publish Date: 2021-01-06 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2020-36185>CVE-2020-36185</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Release Date: 2021-01-06</p> <p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.9.10.8</p> </p> </details> <p></p>
True
CVE-2020-36185 (High) detected in multiple libraries - autoclosed - ## CVE-2020-36185 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>jackson-databind-2.9.0.jar</b>, <b>jackson-databind-2.8.11.4.jar</b>, <b>jackson-databind-2.3.3.jar</b></p></summary> <p> <details><summary><b>jackson-databind-2.9.0.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to vulnerable library: /home/wss-scanner/.ivy2/cache/com.fasterxml.jackson.core/jackson-databind/bundles/jackson-databind-2.9.0.jar</p> <p> Dependency Hierarchy: - play-ehcache_2.12-2.6.25.jar (Root Library) - play_2.12-2.6.25.jar - play-json_2.12-2.6.14.jar - jackson-datatype-jdk8-2.8.11.jar - :x: **jackson-databind-2.9.0.jar** (Vulnerable Library) </details> <details><summary><b>jackson-databind-2.8.11.4.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to vulnerable library: /home/wss-scanner/.ivy2/cache/com.fasterxml.jackson.core/jackson-databind/bundles/jackson-databind-2.8.11.4.jar</p> <p> Dependency Hierarchy: - play-ehcache_2.12-2.6.25.jar (Root Library) - play_2.12-2.6.25.jar - :x: **jackson-databind-2.8.11.4.jar** (Vulnerable Library) </details> <details><summary><b>jackson-databind-2.3.3.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Path to vulnerable library: /home/wss-scanner/.ivy2/cache/com.fasterxml.jackson.core/jackson-databind/bundles/jackson-databind-2.3.3.jar</p> <p> Dependency Hierarchy: - sbt-plugin-2.4.4.jar (Root Library) - sbt-js-engine-1.1.3.jar - npm_2.10-1.1.1.jar - webjars-locator-0.26.jar - :x: **jackson-databind-2.3.3.jar** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/samqws-marketing/coursera_naptime/commit/95750513b615ecf0ea9b7e14fb5f71e577d01a1f">95750513b615ecf0ea9b7e14fb5f71e577d01a1f</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> FasterXML jackson-databind 2.x before 2.9.10.8 mishandles the interaction between serialization gadgets and typing, related to org.apache.tomcat.dbcp.dbcp2.datasources.SharedPoolDataSource. <p>Publish Date: 2021-01-06 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2020-36185>CVE-2020-36185</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Release Date: 2021-01-06</p> <p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.9.10.8</p> </p> </details> <p></p>
non_architecture
cve high detected in multiple libraries autoclosed cve high severity vulnerability vulnerable libraries jackson databind jar jackson databind jar jackson databind jar jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to vulnerable library home wss scanner cache com fasterxml jackson core jackson databind bundles jackson databind jar dependency hierarchy play ehcache jar root library play jar play json jar jackson datatype jar x jackson databind jar vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to vulnerable library home wss scanner cache com fasterxml jackson core jackson databind bundles jackson databind jar dependency hierarchy play ehcache jar root library play jar x jackson databind jar vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api path to vulnerable library home wss scanner cache com fasterxml jackson core jackson databind bundles jackson databind jar dependency hierarchy sbt plugin jar root library sbt js engine jar npm jar webjars locator jar x jackson databind jar vulnerable library found in head commit a href found in base branch master vulnerability details fasterxml jackson databind x before mishandles the interaction between serialization gadgets and typing related to org apache tomcat dbcp datasources sharedpooldatasource publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version release date fix resolution com fasterxml jackson core jackson databind
0
13,152
9,888,310,810
IssuesEvent
2019-06-25 11:14:46
wellcometrust/platform
https://api.github.com/repos/wellcometrust/platform
closed
Distinguish between “could not parse bag-info.txt” and “bag-info.txt does not exist”
📦 Storage service
Follow-up from https://github.com/wellcometrust/storage-service/pull/227
1.0
Distinguish between “could not parse bag-info.txt” and “bag-info.txt does not exist” - Follow-up from https://github.com/wellcometrust/storage-service/pull/227
non_architecture
distinguish between “could not parse bag info txt” and “bag info txt does not exist” follow up from
0
10,332
26,785,479,868
IssuesEvent
2023-02-01 02:09:13
facebook/react-native
https://api.github.com/repos/facebook/react-native
closed
Error "Can't find variable: require" on iOS (RN v.0.68.2)
Stale Needs: Triage :mag: Type: Old Architecture
### Description Error _"Can't find variable: require"_ on 0.68.2 on Debug & Release. Enumeration of presets and babel plugins did not solve the problem. The cache was cleared after each attempt. On the forums of other libraries, the developers dealt with a similar error with a patch in the new version. The bug was discovered after fixing [another bug](https://github.com/facebook/react-native/issues/33954#issuecomment-1146979111) 0.68.2 Cache was cleared before each launch: react-native start --reset-cache The babel config was tested from empty to what worked in previous versions and suggested in other forums and repositories ([1](https://github.com/facebook/react-native/issues/21048#issuecomment-426987192), [2](https://github.com/Tecode/react-native-mobx/blob/master/babel.config.js), [3](https://github.com/facebook/react-native/issues/22321)). ### Version 0.68.2 ### Output of `npx react-native info` System: OS: macOS 11.6 CPU: (6) x64 Intel(R) Core(TM) i5-8500B CPU @ 3.00GHz Memory: 76.89 MB / 8.00 GB Shell: 5.8 - /bin/zsh Binaries: Node: 16.15.0 - /usr/local/bin/node Yarn: 1.22.11 - /usr/local/bin/yarn npm: 8.5.5 - /usr/local/bin/npm Watchman: 2022.05.30.00 - /usr/local/bin/watchman Managers: CocoaPods: 1.11.3 - /usr/local/bin/pod SDKs: iOS SDK: Platforms: DriverKit 21.2, iOS 15.2, macOS 12.1, tvOS 15.2, watchOS 8.3 Android SDK: API Levels: 28, 29, 30, 31 Build Tools: 29.0.2, 30.0.2, 31.0.0, 32.0.0 System Images: android-30 | Google APIs Intel x86 Atom, android-30 | Google Play Intel x86 Atom, android-31 | Google APIs Intel x86 Atom_64 Android NDK: Not Found IDEs: Android Studio: 2020.3 AI-203.7717.56.2031.7678000 Xcode: 13.2.1/13C100 - /usr/bin/xcodebuild Languages: Java: 1.8.0_292 - /usr/bin/javac npmPackages: @react-native-community/cli: 7.0.3 => 7.0.3 react: ^18.1.0 => 18.1.0 react-native: ^0.68.2 => 0.68.2 react-native-macos: Not Found npmGlobalPackages: *react-native*: Not Found ### Steps to reproduce Any Babel configuration. Tested on iOS 15.2 Simulator. ### Snack, code example, screenshot, or link to a repository ![simulator_screenshot_5265B248-4225-43AB-8D7D-2DF37358D657](https://user-images.githubusercontent.com/71312494/172103012-55091305-04b7-4fb0-99b9-563caca90a0f.png)
1.0
Error "Can't find variable: require" on iOS (RN v.0.68.2) - ### Description Error _"Can't find variable: require"_ on 0.68.2 on Debug & Release. Enumeration of presets and babel plugins did not solve the problem. The cache was cleared after each attempt. On the forums of other libraries, the developers dealt with a similar error with a patch in the new version. The bug was discovered after fixing [another bug](https://github.com/facebook/react-native/issues/33954#issuecomment-1146979111) 0.68.2 Cache was cleared before each launch: react-native start --reset-cache The babel config was tested from empty to what worked in previous versions and suggested in other forums and repositories ([1](https://github.com/facebook/react-native/issues/21048#issuecomment-426987192), [2](https://github.com/Tecode/react-native-mobx/blob/master/babel.config.js), [3](https://github.com/facebook/react-native/issues/22321)). ### Version 0.68.2 ### Output of `npx react-native info` System: OS: macOS 11.6 CPU: (6) x64 Intel(R) Core(TM) i5-8500B CPU @ 3.00GHz Memory: 76.89 MB / 8.00 GB Shell: 5.8 - /bin/zsh Binaries: Node: 16.15.0 - /usr/local/bin/node Yarn: 1.22.11 - /usr/local/bin/yarn npm: 8.5.5 - /usr/local/bin/npm Watchman: 2022.05.30.00 - /usr/local/bin/watchman Managers: CocoaPods: 1.11.3 - /usr/local/bin/pod SDKs: iOS SDK: Platforms: DriverKit 21.2, iOS 15.2, macOS 12.1, tvOS 15.2, watchOS 8.3 Android SDK: API Levels: 28, 29, 30, 31 Build Tools: 29.0.2, 30.0.2, 31.0.0, 32.0.0 System Images: android-30 | Google APIs Intel x86 Atom, android-30 | Google Play Intel x86 Atom, android-31 | Google APIs Intel x86 Atom_64 Android NDK: Not Found IDEs: Android Studio: 2020.3 AI-203.7717.56.2031.7678000 Xcode: 13.2.1/13C100 - /usr/bin/xcodebuild Languages: Java: 1.8.0_292 - /usr/bin/javac npmPackages: @react-native-community/cli: 7.0.3 => 7.0.3 react: ^18.1.0 => 18.1.0 react-native: ^0.68.2 => 0.68.2 react-native-macos: Not Found npmGlobalPackages: *react-native*: Not Found ### Steps to reproduce Any Babel configuration. Tested on iOS 15.2 Simulator. ### Snack, code example, screenshot, or link to a repository ![simulator_screenshot_5265B248-4225-43AB-8D7D-2DF37358D657](https://user-images.githubusercontent.com/71312494/172103012-55091305-04b7-4fb0-99b9-563caca90a0f.png)
architecture
error can t find variable require on ios rn v description error can t find variable require on on debug release enumeration of presets and babel plugins did not solve the problem the cache was cleared after each attempt on the forums of other libraries the developers dealt with a similar error with a patch in the new version the bug was discovered after fixing cache was cleared before each launch react native start reset cache the babel config was tested from empty to what worked in previous versions and suggested in other forums and repositories version output of npx react native info system os macos cpu intel r core tm cpu memory mb gb shell bin zsh binaries node usr local bin node yarn usr local bin yarn npm usr local bin npm watchman usr local bin watchman managers cocoapods usr local bin pod sdks ios sdk platforms driverkit ios macos tvos watchos android sdk api levels build tools system images android google apis intel atom android google play intel atom android google apis intel atom android ndk not found ides android studio ai xcode usr bin xcodebuild languages java usr bin javac npmpackages react native community cli react react native react native macos not found npmglobalpackages react native not found steps to reproduce any babel configuration tested on ios simulator snack code example screenshot or link to a repository
1
6,553
14,877,114,902
IssuesEvent
2021-01-20 02:23:19
Azure/azure-sdk
https://api.github.com/repos/Azure/azure-sdk
opened
Board Review: Azure Mixed Reality Authentication client library
architecture board-review
Thank you for submitting this review request. Thorough review of your client library ensures that your APIs are consistent with the guidelines and the consumers of your client library have a consistently good experience when using Azure. **The Architecture Board reviews [Track 2 libraries](https://azure.github.io/azure-sdk/general_introduction.html) only.** If your library does not meet this requirement, please reach out to [Architecture Board]([email protected]) before creating the issue. Please reference our [review process guidelines](https://azure.github.io/azure-sdk/policies_reviewprocess.html) to understand what is being asked for in the issue template. To ensure consistency, all Tier-1 languages (C#, TypeScript, Java, Python) will generally be reviewed together. In expansive libraries, we will pair dynamic languages (Python, TypeScript) together, and strongly typed languages (C#, Java) together in separate meetings. For Tier-2 languages (C, C++, Go, Android, iOS), the review will be on an as-needed basis. **Before submitting, ensure you adjust the title of the issue appropriately.** **Note that the required material must be included before a meeting can be scheduled.** ## Contacts and Timeline * Responsible service team: [email protected] * Main contacts: crtreasu, virivera, rgarcia, ariye * Expected code complete date: 01/29 * Expected release date: 03/01/2021 ## About the Service * Link to documentation introducing/describing the service: https://review.docs.microsoft.com/en-us/azure/object-anchors/overview?branch=release-preview-aou * Link to the service REST APIs: https://github.com/Azure/azure-rest-api-specs/tree/master/specification/mixedreality/data-plane/Microsoft.MixedReality/preview/2019-02-28-preview * Link to GitHub issue for previous review sessions, if applicable: https://github.com/Azure/azure-sdk/issues/2005 ## About the client library * Name of the client library: Azure Mixed Reality Authentication * Languages for this review: JavaScript, Java, Python ## Artifacts required (per language) Please read through “API Review” section [here](https://azure.github.io/azure-sdk/policies_reviewprocess.html) to understand how these artifacts are generated. **It is critical that these artifacts are present and are in the right format. If not, the language architects cannot review them with the SDK Team’s API review tool.** ### .NET n/a already completed * APIView Link: * Link to Champion Scenarios/Quickstart samples: ### Java * APIView Link: https://apiview.dev/Assemblies/Review/f3a4bb684ffd4badada05eff7de952d5 * Link to Champion Scenarios/Quickstart samples: ### Python * APIView Link: * Link to Champion Scenarios/Quickstart samples: ### TypeScript * APIView Link: https://apiview.dev/Assemblies/Review/4917626415bc448c8e2534e00c6f3a17 * Link to Champion Scenarios/Quickstart samples: For all other languages, send a request to the Architecture Board to discuss the best format on individual basis. ## Thank you!
1.0
Board Review: Azure Mixed Reality Authentication client library - Thank you for submitting this review request. Thorough review of your client library ensures that your APIs are consistent with the guidelines and the consumers of your client library have a consistently good experience when using Azure. **The Architecture Board reviews [Track 2 libraries](https://azure.github.io/azure-sdk/general_introduction.html) only.** If your library does not meet this requirement, please reach out to [Architecture Board]([email protected]) before creating the issue. Please reference our [review process guidelines](https://azure.github.io/azure-sdk/policies_reviewprocess.html) to understand what is being asked for in the issue template. To ensure consistency, all Tier-1 languages (C#, TypeScript, Java, Python) will generally be reviewed together. In expansive libraries, we will pair dynamic languages (Python, TypeScript) together, and strongly typed languages (C#, Java) together in separate meetings. For Tier-2 languages (C, C++, Go, Android, iOS), the review will be on an as-needed basis. **Before submitting, ensure you adjust the title of the issue appropriately.** **Note that the required material must be included before a meeting can be scheduled.** ## Contacts and Timeline * Responsible service team: [email protected] * Main contacts: crtreasu, virivera, rgarcia, ariye * Expected code complete date: 01/29 * Expected release date: 03/01/2021 ## About the Service * Link to documentation introducing/describing the service: https://review.docs.microsoft.com/en-us/azure/object-anchors/overview?branch=release-preview-aou * Link to the service REST APIs: https://github.com/Azure/azure-rest-api-specs/tree/master/specification/mixedreality/data-plane/Microsoft.MixedReality/preview/2019-02-28-preview * Link to GitHub issue for previous review sessions, if applicable: https://github.com/Azure/azure-sdk/issues/2005 ## About the client library * Name of the client library: Azure Mixed Reality Authentication * Languages for this review: JavaScript, Java, Python ## Artifacts required (per language) Please read through “API Review” section [here](https://azure.github.io/azure-sdk/policies_reviewprocess.html) to understand how these artifacts are generated. **It is critical that these artifacts are present and are in the right format. If not, the language architects cannot review them with the SDK Team’s API review tool.** ### .NET n/a already completed * APIView Link: * Link to Champion Scenarios/Quickstart samples: ### Java * APIView Link: https://apiview.dev/Assemblies/Review/f3a4bb684ffd4badada05eff7de952d5 * Link to Champion Scenarios/Quickstart samples: ### Python * APIView Link: * Link to Champion Scenarios/Quickstart samples: ### TypeScript * APIView Link: https://apiview.dev/Assemblies/Review/4917626415bc448c8e2534e00c6f3a17 * Link to Champion Scenarios/Quickstart samples: For all other languages, send a request to the Architecture Board to discuss the best format on individual basis. ## Thank you!
architecture
board review azure mixed reality authentication client library thank you for submitting this review request thorough review of your client library ensures that your apis are consistent with the guidelines and the consumers of your client library have a consistently good experience when using azure the architecture board reviews only if your library does not meet this requirement please reach out to adparch microsoft com before creating the issue please reference our to understand what is being asked for in the issue template to ensure consistency all tier languages c typescript java python will generally be reviewed together in expansive libraries we will pair dynamic languages python typescript together and strongly typed languages c java together in separate meetings for tier languages c c go android ios the review will be on an as needed basis before submitting ensure you adjust the title of the issue appropriately note that the required material must be included before a meeting can be scheduled contacts and timeline responsible service team ou servicesdkdev microsoft com main contacts crtreasu virivera rgarcia ariye expected code complete date expected release date about the service link to documentation introducing describing the service link to the service rest apis link to github issue for previous review sessions if applicable about the client library name of the client library azure mixed reality authentication languages for this review javascript java python artifacts required per language please read through “api review” section to understand how these artifacts are generated it is critical that these artifacts are present and are in the right format if not the language architects cannot review them with the sdk team’s api review tool net n a already completed apiview link link to champion scenarios quickstart samples java apiview link link to champion scenarios quickstart samples python apiview link link to champion scenarios quickstart samples typescript apiview link link to champion scenarios quickstart samples for all other languages send a request to the architecture board to discuss the best format on individual basis thank you
1
4,928
11,851,414,736
IssuesEvent
2020-03-24 18:05:01
kubernetes/kubernetes
https://api.github.com/repos/kubernetes/kubernetes
closed
Research for creating a readCoreV1NamespacedPodStatus conformant test
area/conformance sig/architecture sig/testing
# Description After looking through endpoints which need tests written for in APIsnoop, this appears to be an endpoint which isn't hit. # ## Fetch dependencies ```shell go get -v -u k8s.io/apimachinery/pkg/apis/meta/v1 go get -v -u k8s.io/client-go/kubernetes go get -v -u k8s.io/client-go/tools/clientcmd go get -v -u github.com/ghodss/yaml ``` ## Test draft ```go package main import ( "fmt" "flag" "time" "os" "k8s.io/api/core/v1" metav1 "k8s.io/apimachinery/pkg/apis/meta/v1" "k8s.io/client-go/kubernetes" "k8s.io/client-go/tools/clientcmd" "github.com/ghodss/yaml" ) func main() { // uses the current context in kubeconfig kubeconfig := flag.String("kubeconfig", fmt.Sprintf("%v/%v/%v", os.Getenv("HOME"), ".kube", "config"), "(optional) absolute path to the kubeconfig file") flag.Parse() config, err := clientcmd.BuildConfigFromFlags("", *kubeconfig) if err != nil { fmt.Println(err) } // make our work easier to find in the audit_event queries config.UserAgent = "live-test-writing" // creates the clientset clientset, _ := kubernetes.NewForConfig(config) // access the API to list pods _, err = clientset.CoreV1().Pods("default").Create(&v1.Pod{ ObjectMeta: metav1.ObjectMeta{ Name: "nginx", Labels: map[string]string{ "pod-name": "nginx", }, }, Spec: v1.PodSpec{ Containers: []v1.Container{{ Image: "nginx", Name: "nginx", }}, RestartPolicy: v1.RestartPolicyNever, }, }) if err != nil { fmt.Println(err) return } time.Sleep(5 * time.Second) pod, err := clientset.CoreV1().Pods("default").Get("nginx", metav1.GetOptions{}) if err != nil { fmt.Println(err) return } podYAML, err := yaml.Marshal(pod) if err != nil { fmt.Printf("err: %v\n", err) return } fmt.Println(string(podYAML)) time.Sleep(5 * time.Second) err = clientset.CoreV1().Pods("default").Delete("nginx", &metav1.DeleteOptions{}) if err != nil { fmt.Println(err) return } } ``` ## Test draft- Python working implementation ```python pip3 install kubernetes from __future__ import print_function import time import kubernetes.client from kubernetes.client.rest import ApiException from pprint import pprint from kubernetes.client.configuration import Configuration from kubernetes.config import kube_config configuration = Configuration() configuration.host = None kube_config.load_kube_config(client_configuration=configuration) # Uncomment below to setup prefix (e.g. Bearer) for API key, if needed # configuration.api_key_prefix['authorization'] = 'Bearer' # create an instance of the API class api_instance = kubernetes.client.CoreV1Api(kubernetes.client.ApiClient(configuration)) name = 'kindnet-c7vtg' # str | name of the Pod namespace = 'kube-system' # str | object name and auth scope, such as for teams and projects pretty = 'pretty_example' # str | If 'true', then the output is pretty printed. (optional) try: api_response = api_instance.read_namespaced_pod_status(name, namespace, pretty=pretty) pprint(api_response) except ApiException as e: print("Exception when calling CoreV1Api->read_namespaced_pod_status: %s\n" % e) ``` # Verify with APISnoop<a id="sec-4"></a> ## create view for hit endpoints ```sql-mode CREATE VIEW "public"."endpoints_hit_by_new_test" AS WITH live_testing_endpoints AS ( SELECT DISTINCT operation_id, count(1) as hits FROM audit_event WHERE bucket = 'apisnoop' AND useragent = 'live-test-writing' GROUP BY operation_id ), baseline AS ( SELECT DISTINCT operation_id, test_hits, conf_hits FROM endpoint_coverage where bucket != 'apisnoop' ) SELECT DISTINCT lte.operation_id, b.test_hits as hit_by_ete, lte.hits as hit_by_new_test FROM live_testing_endpoints lte JOIN baseline b ON (b.operation_id = lte.operation_id); ``` ## create view for coverage changed ```sql-mode CREATE OR REPLACE VIEW "public"."projected_change_in_coverage" AS with baseline as ( SELECT * FROM stable_endpoint_stats WHERE job != 'live' ), test as ( SELECT count(1) as endpoints_hit FROM ( select operation_id FROM audit_event WHERE useragent = 'live-test-writing' EXCEPT SELECT operation_id FROM endpoint_coverage WHERE test_hits > 0 ) tested_endpoints ), coverage as ( SELECT baseline.test_hits as old_coverage, (baseline.test_hits::int + test.endpoints_hit::int ) as new_coverage from baseline, test ) select 'test_coverage' as category, baseline.total_endpoints, coverage.old_coverage, coverage.new_coverage, (coverage.new_coverage - coverage.old_coverage) as change_in_number from baseline, coverage ; ``` ## find endpoints hit by this test ```sql-mode select * from endpoints_hit_by_new_test; ``` ``` operation_id | hit_by_ete | hit_by_new_test ---------------------------+------------+----------------- createCoreV1NamespacedPod | 1990 | 2 deleteCoreV1NamespacedPod | 2114 | 2 readCoreV1NamespacedPod | 11421 | 1 (3 rows) ``` ## show the change in coverage ```sql-mode select * from projected_change_in_coverage; ``` ``` category | total_endpoints | old_coverage | new_coverage | change_in_number ---------------+-----------------+--------------+--------------+------------------ test_coverage | 430 | 167 | 167 | 0 (1 row) ``` # Final notes From the endpoints hit report above, it doesn't appear that my draft test hit the target endpoint. Would it be possible for some help and/or advice on hitting the `/api/v1/namespaces/NAMESPACE/pods/PODNAME/status` endpoint?
1.0
Research for creating a readCoreV1NamespacedPodStatus conformant test - # Description After looking through endpoints which need tests written for in APIsnoop, this appears to be an endpoint which isn't hit. # ## Fetch dependencies ```shell go get -v -u k8s.io/apimachinery/pkg/apis/meta/v1 go get -v -u k8s.io/client-go/kubernetes go get -v -u k8s.io/client-go/tools/clientcmd go get -v -u github.com/ghodss/yaml ``` ## Test draft ```go package main import ( "fmt" "flag" "time" "os" "k8s.io/api/core/v1" metav1 "k8s.io/apimachinery/pkg/apis/meta/v1" "k8s.io/client-go/kubernetes" "k8s.io/client-go/tools/clientcmd" "github.com/ghodss/yaml" ) func main() { // uses the current context in kubeconfig kubeconfig := flag.String("kubeconfig", fmt.Sprintf("%v/%v/%v", os.Getenv("HOME"), ".kube", "config"), "(optional) absolute path to the kubeconfig file") flag.Parse() config, err := clientcmd.BuildConfigFromFlags("", *kubeconfig) if err != nil { fmt.Println(err) } // make our work easier to find in the audit_event queries config.UserAgent = "live-test-writing" // creates the clientset clientset, _ := kubernetes.NewForConfig(config) // access the API to list pods _, err = clientset.CoreV1().Pods("default").Create(&v1.Pod{ ObjectMeta: metav1.ObjectMeta{ Name: "nginx", Labels: map[string]string{ "pod-name": "nginx", }, }, Spec: v1.PodSpec{ Containers: []v1.Container{{ Image: "nginx", Name: "nginx", }}, RestartPolicy: v1.RestartPolicyNever, }, }) if err != nil { fmt.Println(err) return } time.Sleep(5 * time.Second) pod, err := clientset.CoreV1().Pods("default").Get("nginx", metav1.GetOptions{}) if err != nil { fmt.Println(err) return } podYAML, err := yaml.Marshal(pod) if err != nil { fmt.Printf("err: %v\n", err) return } fmt.Println(string(podYAML)) time.Sleep(5 * time.Second) err = clientset.CoreV1().Pods("default").Delete("nginx", &metav1.DeleteOptions{}) if err != nil { fmt.Println(err) return } } ``` ## Test draft- Python working implementation ```python pip3 install kubernetes from __future__ import print_function import time import kubernetes.client from kubernetes.client.rest import ApiException from pprint import pprint from kubernetes.client.configuration import Configuration from kubernetes.config import kube_config configuration = Configuration() configuration.host = None kube_config.load_kube_config(client_configuration=configuration) # Uncomment below to setup prefix (e.g. Bearer) for API key, if needed # configuration.api_key_prefix['authorization'] = 'Bearer' # create an instance of the API class api_instance = kubernetes.client.CoreV1Api(kubernetes.client.ApiClient(configuration)) name = 'kindnet-c7vtg' # str | name of the Pod namespace = 'kube-system' # str | object name and auth scope, such as for teams and projects pretty = 'pretty_example' # str | If 'true', then the output is pretty printed. (optional) try: api_response = api_instance.read_namespaced_pod_status(name, namespace, pretty=pretty) pprint(api_response) except ApiException as e: print("Exception when calling CoreV1Api->read_namespaced_pod_status: %s\n" % e) ``` # Verify with APISnoop<a id="sec-4"></a> ## create view for hit endpoints ```sql-mode CREATE VIEW "public"."endpoints_hit_by_new_test" AS WITH live_testing_endpoints AS ( SELECT DISTINCT operation_id, count(1) as hits FROM audit_event WHERE bucket = 'apisnoop' AND useragent = 'live-test-writing' GROUP BY operation_id ), baseline AS ( SELECT DISTINCT operation_id, test_hits, conf_hits FROM endpoint_coverage where bucket != 'apisnoop' ) SELECT DISTINCT lte.operation_id, b.test_hits as hit_by_ete, lte.hits as hit_by_new_test FROM live_testing_endpoints lte JOIN baseline b ON (b.operation_id = lte.operation_id); ``` ## create view for coverage changed ```sql-mode CREATE OR REPLACE VIEW "public"."projected_change_in_coverage" AS with baseline as ( SELECT * FROM stable_endpoint_stats WHERE job != 'live' ), test as ( SELECT count(1) as endpoints_hit FROM ( select operation_id FROM audit_event WHERE useragent = 'live-test-writing' EXCEPT SELECT operation_id FROM endpoint_coverage WHERE test_hits > 0 ) tested_endpoints ), coverage as ( SELECT baseline.test_hits as old_coverage, (baseline.test_hits::int + test.endpoints_hit::int ) as new_coverage from baseline, test ) select 'test_coverage' as category, baseline.total_endpoints, coverage.old_coverage, coverage.new_coverage, (coverage.new_coverage - coverage.old_coverage) as change_in_number from baseline, coverage ; ``` ## find endpoints hit by this test ```sql-mode select * from endpoints_hit_by_new_test; ``` ``` operation_id | hit_by_ete | hit_by_new_test ---------------------------+------------+----------------- createCoreV1NamespacedPod | 1990 | 2 deleteCoreV1NamespacedPod | 2114 | 2 readCoreV1NamespacedPod | 11421 | 1 (3 rows) ``` ## show the change in coverage ```sql-mode select * from projected_change_in_coverage; ``` ``` category | total_endpoints | old_coverage | new_coverage | change_in_number ---------------+-----------------+--------------+--------------+------------------ test_coverage | 430 | 167 | 167 | 0 (1 row) ``` # Final notes From the endpoints hit report above, it doesn't appear that my draft test hit the target endpoint. Would it be possible for some help and/or advice on hitting the `/api/v1/namespaces/NAMESPACE/pods/PODNAME/status` endpoint?
architecture
research for creating a conformant test description after looking through endpoints which need tests written for in apisnoop this appears to be an endpoint which isn t hit fetch dependencies shell go get v u io apimachinery pkg apis meta go get v u io client go kubernetes go get v u io client go tools clientcmd go get v u github com ghodss yaml test draft go package main import fmt flag time os io api core io apimachinery pkg apis meta io client go kubernetes io client go tools clientcmd github com ghodss yaml func main uses the current context in kubeconfig kubeconfig flag string kubeconfig fmt sprintf v v v os getenv home kube config optional absolute path to the kubeconfig file flag parse config err clientcmd buildconfigfromflags kubeconfig if err nil fmt println err make our work easier to find in the audit event queries config useragent live test writing creates the clientset clientset kubernetes newforconfig config access the api to list pods err clientset pods default create pod objectmeta objectmeta name nginx labels map string pod name nginx spec podspec containers container image nginx name nginx restartpolicy restartpolicynever if err nil fmt println err return time sleep time second pod err clientset pods default get nginx getoptions if err nil fmt println err return podyaml err yaml marshal pod if err nil fmt printf err v n err return fmt println string podyaml time sleep time second err clientset pods default delete nginx deleteoptions if err nil fmt println err return test draft python working implementation python install kubernetes from future import print function import time import kubernetes client from kubernetes client rest import apiexception from pprint import pprint from kubernetes client configuration import configuration from kubernetes config import kube config configuration configuration configuration host none kube config load kube config client configuration configuration uncomment below to setup prefix e g bearer for api key if needed configuration api key prefix bearer create an instance of the api class api instance kubernetes client kubernetes client apiclient configuration name kindnet str name of the pod namespace kube system str object name and auth scope such as for teams and projects pretty pretty example str if true then the output is pretty printed optional try api response api instance read namespaced pod status name namespace pretty pretty pprint api response except apiexception as e print exception when calling read namespaced pod status s n e verify with apisnoop create view for hit endpoints sql mode create view public endpoints hit by new test as with live testing endpoints as select distinct operation id count as hits from audit event where bucket apisnoop and useragent live test writing group by operation id baseline as select distinct operation id test hits conf hits from endpoint coverage where bucket apisnoop select distinct lte operation id b test hits as hit by ete lte hits as hit by new test from live testing endpoints lte join baseline b on b operation id lte operation id create view for coverage changed sql mode create or replace view public projected change in coverage as with baseline as select from stable endpoint stats where job live test as select count as endpoints hit from select operation id from audit event where useragent live test writing except select operation id from endpoint coverage where test hits tested endpoints coverage as select baseline test hits as old coverage baseline test hits int test endpoints hit int as new coverage from baseline test select test coverage as category baseline total endpoints coverage old coverage coverage new coverage coverage new coverage coverage old coverage as change in number from baseline coverage find endpoints hit by this test sql mode select from endpoints hit by new test operation id hit by ete hit by new test rows show the change in coverage sql mode select from projected change in coverage category total endpoints old coverage new coverage change in number test coverage row final notes from the endpoints hit report above it doesn t appear that my draft test hit the target endpoint would it be possible for some help and or advice on hitting the api namespaces namespace pods podname status endpoint
1
8,702
23,287,981,996
IssuesEvent
2022-08-05 18:45:42
Azure/azure-sdk
https://api.github.com/repos/Azure/azure-sdk
closed
Board Review: metrics advisor (Python & .net)
architecture board-review
## The Basics * Service team responsible for the client library: Metrics Advisor * Link to documentation describing the service: https://docs.microsoft.com/en-us/azure/cognitive-services/metrics-advisor/ * Contact email (if service team, provide PM and Dev Lead): [email protected], [email protected] (dev) [email protected] (PM) ## About this client library * Name of the client library: azure-ai-metricsadvisor * Languages for this review: .net/Python * Link to the service REST APIs: https://github.com/bowgong/azure-rest-api-specs/blob/metricsadvisor-preview/specification/cognitiveservices/data-plane/MetricsAdvisor/preview/v1.0/MetricsAdvisor.json https://westus2.dev.cognitive.microsoft.com/docs/services/MetricsAdvisor/operations/createDataFeed ## Artifacts required (per language) We use an API review tool ([apiview](https://apiview.azurewebsites.net)) to support .NET and Java API reviews. For Python and TypeScript, use the API extractor tool, then submit the output as a Draft PR to the relevant repository (azure-sdk-for-python or azure-sdk-for-js). ### .NET * [APIView](https://apiview.dev/Assemblies/Review/8caf3dd1661c45228d8081a536cca3bc) * Link to samples for champion scenarios: https://github.com/Azure/azure-sdk-for-net/blob/master/sdk/metricsadvisor/Azure.AI.MetricsAdvisor/samples/README.md ### Python * [APIView](https://apiview.dev/Assemblies/Review/6acc354ec6c5421b82a081a07b481df0) * Link to samples for champion scenarios: https://github.com/Azure/azure-sdk-for-python/tree/master/sdk/metricsadvisor/azure-ai-metricsadvisor/samples ## Champion Scenarios A champion scenario is a use case that the consumer of the client library is commonly expected to perform. Champion scenarios are used to ensure the developer experience is exemplary for the common cases. You need to show the entire code sample (including error handling, as an example) for the champion scenarios. * Champion Scenario 1: * Describe the champion scenario: create a data feed to ingest data * Estimate the percentage of developers using the service who would use the champion scenario * Link to the code samples: [Python](https://gist.github.com/xiangyan99/6c367d45168294b043b2f5685d57f584#create-data-feed) | [.NET](https://gist.github.com/kinelski/c072790394398c37d186df611f6bea44#scenario-1-datafeed-creation) * Champion Scenario 2: * Describe the champion scenario: create a configuration to let service know whether a point is anomaly * Estimate the percentage of developers using the service who would use the champion scenario * Link to the code samples: [Python](https://gist.github.com/xiangyan99/6c367d45168294b043b2f5685d57f584#create-detection-configuration) | [.NET](https://gist.github.com/kinelski/c072790394398c37d186df611f6bea44#2d-applyingtuning-anomaly-detection) * Champion Scenario 3: * Describe the champion scenario: configure the service when to trigger an alert * Estimate the percentage of developers using the service who would use the champion scenario * Link to the code samples: [Python](https://gist.github.com/xiangyan99/6c367d45168294b043b2f5685d57f584#config-alert-configuration) | [.NET](https://gist.github.com/kinelski/c072790394398c37d186df611f6bea44#scenario-3-configure-alerts-and-get-incidents-notification-using-a-hook) * Champion Scenario 4: * Describe the champion scenario: query anomalies & alerts * Estimate the percentage of developers using the service who would use the champion scenario * Link to the code samples: [Python](https://gist.github.com/xiangyan99/6c367d45168294b043b2f5685d57f584#query-anomalies-for-alert-configuration) | [.NET](https://github.com/Azure/azure-sdk-for-net/blob/master/sdk/metricsadvisor/Azure.AI.MetricsAdvisor/README.md#query-detected-anomalies-and-triggered-alerts)
1.0
Board Review: metrics advisor (Python & .net) - ## The Basics * Service team responsible for the client library: Metrics Advisor * Link to documentation describing the service: https://docs.microsoft.com/en-us/azure/cognitive-services/metrics-advisor/ * Contact email (if service team, provide PM and Dev Lead): [email protected], [email protected] (dev) [email protected] (PM) ## About this client library * Name of the client library: azure-ai-metricsadvisor * Languages for this review: .net/Python * Link to the service REST APIs: https://github.com/bowgong/azure-rest-api-specs/blob/metricsadvisor-preview/specification/cognitiveservices/data-plane/MetricsAdvisor/preview/v1.0/MetricsAdvisor.json https://westus2.dev.cognitive.microsoft.com/docs/services/MetricsAdvisor/operations/createDataFeed ## Artifacts required (per language) We use an API review tool ([apiview](https://apiview.azurewebsites.net)) to support .NET and Java API reviews. For Python and TypeScript, use the API extractor tool, then submit the output as a Draft PR to the relevant repository (azure-sdk-for-python or azure-sdk-for-js). ### .NET * [APIView](https://apiview.dev/Assemblies/Review/8caf3dd1661c45228d8081a536cca3bc) * Link to samples for champion scenarios: https://github.com/Azure/azure-sdk-for-net/blob/master/sdk/metricsadvisor/Azure.AI.MetricsAdvisor/samples/README.md ### Python * [APIView](https://apiview.dev/Assemblies/Review/6acc354ec6c5421b82a081a07b481df0) * Link to samples for champion scenarios: https://github.com/Azure/azure-sdk-for-python/tree/master/sdk/metricsadvisor/azure-ai-metricsadvisor/samples ## Champion Scenarios A champion scenario is a use case that the consumer of the client library is commonly expected to perform. Champion scenarios are used to ensure the developer experience is exemplary for the common cases. You need to show the entire code sample (including error handling, as an example) for the champion scenarios. * Champion Scenario 1: * Describe the champion scenario: create a data feed to ingest data * Estimate the percentage of developers using the service who would use the champion scenario * Link to the code samples: [Python](https://gist.github.com/xiangyan99/6c367d45168294b043b2f5685d57f584#create-data-feed) | [.NET](https://gist.github.com/kinelski/c072790394398c37d186df611f6bea44#scenario-1-datafeed-creation) * Champion Scenario 2: * Describe the champion scenario: create a configuration to let service know whether a point is anomaly * Estimate the percentage of developers using the service who would use the champion scenario * Link to the code samples: [Python](https://gist.github.com/xiangyan99/6c367d45168294b043b2f5685d57f584#create-detection-configuration) | [.NET](https://gist.github.com/kinelski/c072790394398c37d186df611f6bea44#2d-applyingtuning-anomaly-detection) * Champion Scenario 3: * Describe the champion scenario: configure the service when to trigger an alert * Estimate the percentage of developers using the service who would use the champion scenario * Link to the code samples: [Python](https://gist.github.com/xiangyan99/6c367d45168294b043b2f5685d57f584#config-alert-configuration) | [.NET](https://gist.github.com/kinelski/c072790394398c37d186df611f6bea44#scenario-3-configure-alerts-and-get-incidents-notification-using-a-hook) * Champion Scenario 4: * Describe the champion scenario: query anomalies & alerts * Estimate the percentage of developers using the service who would use the champion scenario * Link to the code samples: [Python](https://gist.github.com/xiangyan99/6c367d45168294b043b2f5685d57f584#query-anomalies-for-alert-configuration) | [.NET](https://github.com/Azure/azure-sdk-for-net/blob/master/sdk/metricsadvisor/Azure.AI.MetricsAdvisor/README.md#query-detected-anomalies-and-triggered-alerts)
architecture
board review metrics advisor python net the basics service team responsible for the client library metrics advisor link to documentation describing the service contact email if service team provide pm and dev lead bix microsoft com bowgong microsoft com dev quying microsoft com pm about this client library name of the client library azure ai metricsadvisor languages for this review net python link to the service rest apis artifacts required per language we use an api review tool to support net and java api reviews for python and typescript use the api extractor tool then submit the output as a draft pr to the relevant repository azure sdk for python or azure sdk for js net link to samples for champion scenarios python link to samples for champion scenarios champion scenarios a champion scenario is a use case that the consumer of the client library is commonly expected to perform champion scenarios are used to ensure the developer experience is exemplary for the common cases you need to show the entire code sample including error handling as an example for the champion scenarios champion scenario describe the champion scenario create a data feed to ingest data estimate the percentage of developers using the service who would use the champion scenario link to the code samples champion scenario describe the champion scenario create a configuration to let service know whether a point is anomaly estimate the percentage of developers using the service who would use the champion scenario link to the code samples champion scenario describe the champion scenario configure the service when to trigger an alert estimate the percentage of developers using the service who would use the champion scenario link to the code samples champion scenario describe the champion scenario query anomalies alerts estimate the percentage of developers using the service who would use the champion scenario link to the code samples
1
End of preview. Expand in Data Studio

Dataset Card for "binary-10IQR-architecture"

More Information needed

Downloads last month
24

Collection including karths/binary-10IQR-architecture