Technical Debt and it Types Datasets
Collection
24 items
•
Updated
•
1
Unnamed: 0
int64 0
1.97k
| id
float64 2.49B
32.1B
| type
stringclasses 1
value | created_at
stringdate 2015-01-02 17:29:17
2023-09-25 12:49:49
| repo
stringclasses 857
values | repo_url
stringclasses 857
values | action
stringclasses 3
values | title
stringlengths 3
238
| labels
stringlengths 9
347
| body
stringlengths 4
84.9k
| index
float64 1
7
| text_combine
stringlengths 30
85k
| label
stringclasses 13
values | text
stringlengths 13
80.4k
|
---|---|---|---|---|---|---|---|---|---|---|---|---|---|
0 | 2,562,691,887 | IssuesEvent | 2015-02-06 05:04:33 | NordikSoft/minesweeper | https://api.github.com/repos/NordikSoft/minesweeper | closed | add tests for the command line flag parser | architecture debt task | Make the command line parser separated from the rest of the application so that it can be published and tested. Write test cases for the utility to cover the following cases:
* all six of the arguments are called and captured
* calling with unknown values
* calling with garbled data
* calling with impossible values (negative, more mines than spots, zero, ...)
* any other tests that seem worthwhile
Modify the functionality so that the values are checked on run, before passing them to the board curator. | 1 | add tests for the command line flag parser - Make the command line parser separated from the rest of the application so that it can be published and tested. Write test cases for the utility to cover the following cases:
* all six of the arguments are called and captured
* calling with unknown values
* calling with garbled data
* calling with impossible values (negative, more mines than spots, zero, ...)
* any other tests that seem worthwhile
Modify the functionality so that the values are checked on run, before passing them to the board curator. | architecture | add tests for the command line flag parser make the command line parser separated from the rest of the application so that it can be published and tested write test cases for the utility to cover the following cases all six of the arguments are called and captured calling with unknown values calling with garbled data calling with impossible values negative more mines than spots zero any other tests that seem worthwhile modify the functionality so that the values are checked on run before passing them to the board curator |
1 | 2,565,130,473 | IssuesEvent | 2015-02-07 02:15:41 | NordikSoft/minesweeper | https://api.github.com/repos/NordikSoft/minesweeper | closed | move constant values to header files | architecture debt task | Move the constant values into header files to reduce the need for recompilation of source files and ease re-use of values. | 1 | move constant values to header files - Move the constant values into header files to reduce the need for recompilation of source files and ease re-use of values. | architecture | move constant values to header files move the constant values into header files to reduce the need for recompilation of source files and ease re use of values |
2 | 3,090,544,695 | IssuesEvent | 2015-08-26 07:31:47 | ravaj-group/core-issues | https://api.github.com/repos/ravaj-group/core-issues | opened | [AdminBundle] Improve general performance | architecture technical debt | There are many things we can improve for a better general performance of the panel, such as:
- [ ] Cache assets (javascripts, translations, etc) in users browser and load them only once. We can show a progress bar on first load so user understands why it's taking so long to load.
- [ ] Investigate directives and scopes using profiling tools to find performance bottlenecks. | 1 | [AdminBundle] Improve general performance - There are many things we can improve for a better general performance of the panel, such as:
- [ ] Cache assets (javascripts, translations, etc) in users browser and load them only once. We can show a progress bar on first load so user understands why it's taking so long to load.
- [ ] Investigate directives and scopes using profiling tools to find performance bottlenecks. | architecture | improve general performance there are many things we can improve for a better general performance of the panel such as cache assets javascripts translations etc in users browser and load them only once we can show a progress bar on first load so user understands why it s taking so long to load investigate directives and scopes using profiling tools to find performance bottlenecks |
3 | 4,269,642,997 | IssuesEvent | 2016-07-13 01:41:27 | jakimber/TDiary | https://api.github.com/repos/jakimber/TDiary | closed | Shotgun Surgery | architecture question tech debt | Added "Rating" to a Diary Item and it required too many file changes to implement fully. There has to be a better way! | 1 | Shotgun Surgery - Added "Rating" to a Diary Item and it required too many file changes to implement fully. There has to be a better way! | architecture | shotgun surgery added rating to a diary item and it required too many file changes to implement fully there has to be a better way |
5 | 7,840,527,338 | IssuesEvent | 2018-06-18 16:37:48 | AlwaysInMind/aim-web-app | https://api.github.com/repos/AlwaysInMind/aim-web-app | closed | Make AI description fetch optional | architecture technical debt | The Azure cognitive API used to get the photo descriptions is a paid service and always called. If the Caption option is off then the description is not required so should not be fetched.
- [ ] API request to server needs an option
- [ ] API is set at module load so need to re-architecture to use dynamic - like IDs | 1 | Make AI description fetch optional - The Azure cognitive API used to get the photo descriptions is a paid service and always called. If the Caption option is off then the description is not required so should not be fetched.
- [ ] API request to server needs an option
- [ ] API is set at module load so need to re-architecture to use dynamic - like IDs | architecture | make ai description fetch optional the azure cognitive api used to get the photo descriptions is a paid service and always called if the caption option is off then the description is not required so should not be fetched api request to server needs an option api is set at module load so need to re architecture to use dynamic like ids |
8 | 8,576,385,544 | IssuesEvent | 2018-11-12 20:13:05 | fga-eps-mds/2018.2-GamesBI | https://api.github.com/repos/fga-eps-mds/2018.2-GamesBI | closed | Retirar app importdata do CrossData | 0 - Architecture 1 - API 2 - Bug 2 - Technical Viability 3 - Debt 4 - Evolution 6 - Python/Flask | ## Descrição
<!--- Describe the reason of the issue, what is the problem you want to solve, the bug you want to fix, etc -->
Após a atualização da arquitetura do projeto, o app 'importdata' não teve mais utilidade, portanto deve ser retirado.
## Tarefas
- [x] Transferências das models para o app API
- [x] Transferências dos serializers para o app API
- [x] Retirar todas as menções ao importdata no projeto
- [x] Apagar o app importdata
- [x] Gerar novas fixtures
## Critérios de aceitação
<!--- describe what needs to be done so this issue can be closed -->
O projeto deve continuar funcionando normalmente
## Como isso vai beneficiar o projeto?
Irá trazer maior organização para o projeto
| 1 | Retirar app importdata do CrossData - ## Descrição
<!--- Describe the reason of the issue, what is the problem you want to solve, the bug you want to fix, etc -->
Após a atualização da arquitetura do projeto, o app 'importdata' não teve mais utilidade, portanto deve ser retirado.
## Tarefas
- [x] Transferências das models para o app API
- [x] Transferências dos serializers para o app API
- [x] Retirar todas as menções ao importdata no projeto
- [x] Apagar o app importdata
- [x] Gerar novas fixtures
## Critérios de aceitação
<!--- describe what needs to be done so this issue can be closed -->
O projeto deve continuar funcionando normalmente
## Como isso vai beneficiar o projeto?
Irá trazer maior organização para o projeto
| architecture | retirar app importdata do crossdata descrição após a atualização da arquitetura do projeto o app importdata não teve mais utilidade portanto deve ser retirado tarefas transferências das models para o app api transferências dos serializers para o app api retirar todas as menções ao importdata no projeto apagar o app importdata gerar novas fixtures critérios de aceitação o projeto deve continuar funcionando normalmente como isso vai beneficiar o projeto irá trazer maior organização para o projeto |
11 | 9,186,831,666 | IssuesEvent | 2019-03-06 00:17:48 | vmware/vic | https://api.github.com/repos/vmware/vic | closed | Structured errors | Epic kind/architecture kind/debt kind/enhancement priority/p4 resolution/will-not-fix | https://github.com/vmware/vic/pull/1708/files#diff-20cac0652b997c8f1ee1d0fb89688480R706 is an example what happens when we don't define errors as types.
We should implement a set of types like net package (https://golang.org/pkg/net/) and start using type assertions
#3636 should be defining these errors as an integral part of the interface work for the portlayer. | 1 | Structured errors - https://github.com/vmware/vic/pull/1708/files#diff-20cac0652b997c8f1ee1d0fb89688480R706 is an example what happens when we don't define errors as types.
We should implement a set of types like net package (https://golang.org/pkg/net/) and start using type assertions
#3636 should be defining these errors as an integral part of the interface work for the portlayer. | architecture | structured errors is an example what happens when we don t define errors as types we should implement a set of types like net package and start using type assertions should be defining these errors as an integral part of the interface work for the portlayer |
12 | 9,383,358,987 | IssuesEvent | 2019-04-05 03:01:32 | fga-eps-mds/2019.1-Wendy | https://api.github.com/repos/fga-eps-mds/2019.1-Wendy | closed | Criar roadmap de cada frente | Architecture Owner Devops EPS Product Owner Tech Lead Technical Debt | Nessa issue será realizado:
- [x] DevOps deve criar o seu roadmap para a primeira release
- [x] Architecture Owner deve criar o seu roadmap para a primeira release
- [x] Product Owner deve criar o seu roadmap para a primeira release
- [x] Tech Lead deve criar o seu roadmap para a primeira release
Critérios de aceitação:
- [x] Estar documentado na wiki
- [x] Cada papel com sua cor de acordo com a paleta de cores do projeto
- [x] Todas as sprints da R1 preenchidas
| 1 | Criar roadmap de cada frente - Nessa issue será realizado:
- [x] DevOps deve criar o seu roadmap para a primeira release
- [x] Architecture Owner deve criar o seu roadmap para a primeira release
- [x] Product Owner deve criar o seu roadmap para a primeira release
- [x] Tech Lead deve criar o seu roadmap para a primeira release
Critérios de aceitação:
- [x] Estar documentado na wiki
- [x] Cada papel com sua cor de acordo com a paleta de cores do projeto
- [x] Todas as sprints da R1 preenchidas
| architecture | criar roadmap de cada frente nessa issue será realizado devops deve criar o seu roadmap para a primeira release architecture owner deve criar o seu roadmap para a primeira release product owner deve criar o seu roadmap para a primeira release tech lead deve criar o seu roadmap para a primeira release critérios de aceitação estar documentado na wiki cada papel com sua cor de acordo com a paleta de cores do projeto todas as sprints da preenchidas |
13 | 9,478,932,162 | IssuesEvent | 2019-04-20 02:59:46 | fga-eps-mds/2019.1-Gaia | https://api.github.com/repos/fga-eps-mds/2019.1-Gaia | closed | Definir como será a integração entre os microsserviços | Architecture Owner EPS Project Backlog Technical Debt | Definir como será a integração entre os microsserviços
Nessa issue será realizado:
- [x] Explicar como a integração entre microsserviços será feita
- [x] Justificar a escolha de integração
Critérios de aceitação:
- [x] Estar documentado na wiki
| 1 | Definir como será a integração entre os microsserviços - Definir como será a integração entre os microsserviços
Nessa issue será realizado:
- [x] Explicar como a integração entre microsserviços será feita
- [x] Justificar a escolha de integração
Critérios de aceitação:
- [x] Estar documentado na wiki
| architecture | definir como será a integração entre os microsserviços definir como será a integração entre os microsserviços nessa issue será realizado explicar como a integração entre microsserviços será feita justificar a escolha de integração critérios de aceitação estar documentado na wiki |
14 | 9,975,770,741 | IssuesEvent | 2019-07-09 13:46:51 | woocommerce/woocommerce-android | https://api.github.com/repos/woocommerce/woocommerce-android | closed | Convert to using Android Navigation Components | AndroidX Architecture Tech Debt | I met with a google engineer during Google I/O to talk about some of the issues we've been having with managing fragment states in this woo app. The grand takeaways from my session were:
1. Everyone has issues with managing the fragment lifecycle, even developers at Google.
2. This is precisely why the android team created the jetpack navigation components.
So the recommendation was to switch our architecture over to using these new components now that they are in production. | 1 | Convert to using Android Navigation Components - I met with a google engineer during Google I/O to talk about some of the issues we've been having with managing fragment states in this woo app. The grand takeaways from my session were:
1. Everyone has issues with managing the fragment lifecycle, even developers at Google.
2. This is precisely why the android team created the jetpack navigation components.
So the recommendation was to switch our architecture over to using these new components now that they are in production. | architecture | convert to using android navigation components i met with a google engineer during google i o to talk about some of the issues we ve been having with managing fragment states in this woo app the grand takeaways from my session were everyone has issues with managing the fragment lifecycle even developers at google this is precisely why the android team created the jetpack navigation components so the recommendation was to switch our architecture over to using these new components now that they are in production |
15 | 10,878,807,026 | IssuesEvent | 2019-11-16 20:16:39 | fga-eps-mds/2019.2-Acacia | https://api.github.com/repos/fga-eps-mds/2019.2-Acacia | closed | Configurar ambiente de produção | EPS architecture devops technical debt technical story | **Descrição**
Eu, como ***devops***, gostaria de ***configurar o ambiente de produção*** para ***disponibilizar de forma estável o produto***.
**Depende de:**
#79
**Critérios de Aceitação**
- Deverá ter o campo ...;
- Deverá calcular/apresentar ...;
- Funcionalidade testada;
**Tarefas**
- [ ] Não é um épico. [Leia](https://sitecampus.com.br/user-story-epico-e-tema-qual-diferenca/);
- [ ] É testável;
- [ ] É estimável pelo time de desenvolvimento;
- [ ] Traz valor ao negócio;
**Observações**
- A issue deve ser pontuada;
- A issue deve ser delegada a alguém;
- A issue deve ter labels;
| 1 | Configurar ambiente de produção - **Descrição**
Eu, como ***devops***, gostaria de ***configurar o ambiente de produção*** para ***disponibilizar de forma estável o produto***.
**Depende de:**
#79
**Critérios de Aceitação**
- Deverá ter o campo ...;
- Deverá calcular/apresentar ...;
- Funcionalidade testada;
**Tarefas**
- [ ] Não é um épico. [Leia](https://sitecampus.com.br/user-story-epico-e-tema-qual-diferenca/);
- [ ] É testável;
- [ ] É estimável pelo time de desenvolvimento;
- [ ] Traz valor ao negócio;
**Observações**
- A issue deve ser pontuada;
- A issue deve ser delegada a alguém;
- A issue deve ter labels;
| architecture | configurar ambiente de produção descrição eu como devops gostaria de configurar o ambiente de produção para disponibilizar de forma estável o produto depende de critérios de aceitação deverá ter o campo deverá calcular apresentar funcionalidade testada tarefas não é um épico é testável é estimável pelo time de desenvolvimento traz valor ao negócio observações a issue deve ser pontuada a issue deve ser delegada a alguém a issue deve ter labels |
16 | 10,932,557,749 | IssuesEvent | 2019-11-23 18:35:49 | fga-eps-mds/2019.2-Acacia | https://api.github.com/repos/fga-eps-mds/2019.2-Acacia | closed | Configurar entrega contínua | EPS architecture devops technical debt technical story | **Descrição**
Eu, como ***devops***, gostaria de ***configurar a entrega contínua*** para ***automatizar a entrega do produto***.
**Critérios de Aceitação**
- Pushs na branch master ativando o workflow de deploy do ambiente de produção
**Tarefas**
- [x] Configurar workflow do github actions
**Observações**
- Essa issue depende da configuração do ambiente de produção para ser validada por completo
| 1 | Configurar entrega contínua - **Descrição**
Eu, como ***devops***, gostaria de ***configurar a entrega contínua*** para ***automatizar a entrega do produto***.
**Critérios de Aceitação**
- Pushs na branch master ativando o workflow de deploy do ambiente de produção
**Tarefas**
- [x] Configurar workflow do github actions
**Observações**
- Essa issue depende da configuração do ambiente de produção para ser validada por completo
| architecture | configurar entrega contínua descrição eu como devops gostaria de configurar a entrega contínua para automatizar a entrega do produto critérios de aceitação pushs na branch master ativando o workflow de deploy do ambiente de produção tarefas configurar workflow do github actions observações essa issue depende da configuração do ambiente de produção para ser validada por completo |
18 | 12,060,151,218 | IssuesEvent | 2020-04-15 20:38:12 | COVID-19-electronic-health-system/Corona-tracker | https://api.github.com/repos/COVID-19-electronic-health-system/Corona-tracker | closed | Create QA client in CoronaTracker AWS | architecture tech debt v2 | ## Summary
- [ ] Create a new S3 bucket containing the most up-to-date copy of the site
- [ ] Share this with the community, preferably pinning it in large Discord groups and potentially adding it to the README (reach out to @whoabuddy for this)
## Motivation
While longterm we may look to add more environments with separate test backends and all, but for the time being, a simple separate client where we will deploy in order to test functionality before deploy to should will suffice.
This also fully removes my personal AWS from the picture.
## Describe alternatives you've considered
Waiting until we have a separate QA/Staging backend in order to implement this, but we really should have somewhere to test the client as that's currently the most heavily-under-development aspect of CoronaTracker | 1 | Create QA client in CoronaTracker AWS - ## Summary
- [ ] Create a new S3 bucket containing the most up-to-date copy of the site
- [ ] Share this with the community, preferably pinning it in large Discord groups and potentially adding it to the README (reach out to @whoabuddy for this)
## Motivation
While longterm we may look to add more environments with separate test backends and all, but for the time being, a simple separate client where we will deploy in order to test functionality before deploy to should will suffice.
This also fully removes my personal AWS from the picture.
## Describe alternatives you've considered
Waiting until we have a separate QA/Staging backend in order to implement this, but we really should have somewhere to test the client as that's currently the most heavily-under-development aspect of CoronaTracker | architecture | create qa client in coronatracker aws summary create a new bucket containing the most up to date copy of the site share this with the community preferably pinning it in large discord groups and potentially adding it to the readme reach out to whoabuddy for this motivation while longterm we may look to add more environments with separate test backends and all but for the time being a simple separate client where we will deploy in order to test functionality before deploy to should will suffice this also fully removes my personal aws from the picture describe alternatives you ve considered waiting until we have a separate qa staging backend in order to implement this but we really should have somewhere to test the client as that s currently the most heavily under development aspect of coronatracker |
19 | 12,237,227,972 | IssuesEvent | 2020-05-04 17:39:59 | eurofurence/ef-app_ios | https://api.github.com/repos/eurofurence/ef-app_ios | closed | Swap out ApplicationDirector for content routers | rearchitecture technical debt | The ApplicationDirector is pretty huge/gross as it has to handle actions explicitly from modules alongside deep linking. Moving to a router-based system will make things a lot more flexible/sane | 1 | Swap out ApplicationDirector for content routers - The ApplicationDirector is pretty huge/gross as it has to handle actions explicitly from modules alongside deep linking. Moving to a router-based system will make things a lot more flexible/sane | architecture | swap out applicationdirector for content routers the applicationdirector is pretty huge gross as it has to handle actions explicitly from modules alongside deep linking moving to a router based system will make things a lot more flexible sane |
20 | 12,301,448,147 | IssuesEvent | 2020-05-11 15:25:37 | kids-first/kf-portal-ui | https://api.github.com/repos/kids-first/kf-portal-ui | closed | Persona: Migrate storage from MongoDB to DocumentDB | architecture backend tech debt to groom | As devops admins, we want a system that is as easy as possible to deploy.
Our devops asked us to move away from MongoDB as much as possible.
We can then discuss moving to DocumentDB, or another backend.
## Acceptance criteria
- Migrate Persona so that its storage backend is DocumentDB, not MongoDB.
| 1 | Persona: Migrate storage from MongoDB to DocumentDB - As devops admins, we want a system that is as easy as possible to deploy.
Our devops asked us to move away from MongoDB as much as possible.
We can then discuss moving to DocumentDB, or another backend.
## Acceptance criteria
- Migrate Persona so that its storage backend is DocumentDB, not MongoDB.
| architecture | persona migrate storage from mongodb to documentdb as devops admins we want a system that is as easy as possible to deploy our devops asked us to move away from mongodb as much as possible we can then discuss moving to documentdb or another backend acceptance criteria migrate persona so that its storage backend is documentdb not mongodb |
21 | 13,246,793,308 | IssuesEvent | 2020-08-19 16:12:56 | dusk-network/dusk-blindbid | https://api.github.com/repos/dusk-network/dusk-blindbid | closed | Migrate the lib circuit deps to plonk_gadgets | area:architecture type:tech-debt | Since we ported to the repo https://github.com/dusk-network/plonk_gadgets all of the general-purpose gadgets we currently have. It would be nice to migrate the library to use them as a dependency. So we reduce the code and the responsabilities of this repo to just hold the logic for the blindbid ops. | 1 | Migrate the lib circuit deps to plonk_gadgets - Since we ported to the repo https://github.com/dusk-network/plonk_gadgets all of the general-purpose gadgets we currently have. It would be nice to migrate the library to use them as a dependency. So we reduce the code and the responsabilities of this repo to just hold the logic for the blindbid ops. | architecture | migrate the lib circuit deps to plonk gadgets since we ported to the repo all of the general purpose gadgets we currently have it would be nice to migrate the library to use them as a dependency so we reduce the code and the responsabilities of this repo to just hold the logic for the blindbid ops |
23 | 13,657,383,859 | IssuesEvent | 2020-09-28 05:36:33 | infinyon/fluvio | https://api.github.com/repos/infinyon/fluvio | closed | PartitionStatus mapping error | Kubernetes doc/architecture technical debt | platform: `ubuntu-18`
steps to reproduce:
```make smoke-test-tls```
log:
```
fluvio_stream_dispatcher::dispatcher::k8_ws_service: invalid type: string "Failure", expected struct PartitionStatus at line 7 column 21
Sep 23 17:13:48.020 ERROR flv_tls_proxy: error copying: Broken pipe (os error 32)
Sep 23 17:13:48.098 ERROR k8_client::native::client: error decoding raw stream :
"message": "Operation cannot be fulfilled on partitions.fluvio.infinyon.com \"topic2-0\": the object has been modified; please apply your changes to the latest version and try again",
"reason": "Conflict",
"details": {
"name": "topic2-0",
"group": "fluvio.infinyon.com",
"kind": "partitions"
},
"code": 409
}
```
issue: Error is not properly mapped into Status.
| 1 | PartitionStatus mapping error - platform: `ubuntu-18`
steps to reproduce:
```make smoke-test-tls```
log:
```
fluvio_stream_dispatcher::dispatcher::k8_ws_service: invalid type: string "Failure", expected struct PartitionStatus at line 7 column 21
Sep 23 17:13:48.020 ERROR flv_tls_proxy: error copying: Broken pipe (os error 32)
Sep 23 17:13:48.098 ERROR k8_client::native::client: error decoding raw stream :
"message": "Operation cannot be fulfilled on partitions.fluvio.infinyon.com \"topic2-0\": the object has been modified; please apply your changes to the latest version and try again",
"reason": "Conflict",
"details": {
"name": "topic2-0",
"group": "fluvio.infinyon.com",
"kind": "partitions"
},
"code": 409
}
```
issue: Error is not properly mapped into Status.
| architecture | partitionstatus mapping error platform ubuntu steps to reproduce make smoke test tls log fluvio stream dispatcher dispatcher ws service invalid type string failure expected struct partitionstatus at line column sep error flv tls proxy error copying broken pipe os error sep error client native client error decoding raw stream message operation cannot be fulfilled on partitions fluvio infinyon com the object has been modified please apply your changes to the latest version and try again reason conflict details name group fluvio infinyon com kind partitions code issue error is not properly mapped into status |
24 | 13,683,302,912 | IssuesEvent | 2020-09-30 01:24:35 | woocommerce/woocommerce-ios | https://api.github.com/repos/woocommerce/woocommerce-ios | closed | StorageType+Woo: siteID is part of the pK! | Architecture Tech Debt | ### Details:
`StorageType+Woo` implements a set of methods used by the OrderStore. We need to extend those methods, so that the **siteID** is also considered as part of the Order's primary key.
| 1 | StorageType+Woo: siteID is part of the pK! - ### Details:
`StorageType+Woo` implements a set of methods used by the OrderStore. We need to extend those methods, so that the **siteID** is also considered as part of the Order's primary key.
| architecture | storagetype woo siteid is part of the pk details storagetype woo implements a set of methods used by the orderstore we need to extend those methods so that the siteid is also considered as part of the order s primary key |
25 | 13,876,693,211 | IssuesEvent | 2020-10-17 00:19:22 | infinyon/fluvio | https://api.github.com/repos/infinyon/fluvio | closed | SPU is receiving unnecessary replica update event from SC | Priority - Critical SC SPU bug doc/architecture technical debt | steps reproduce:
```
flvt --produce-iteration 10
```
In the SPU Log, we see log with:
```
Oct 10 01:39:33.494 DEBUG sc_request_loop:update_replica_request: fluvio_spu::controllers::sc::dispatcher: received replica update from sc: UpdateReplicaRequest {
epoch: 0,
changes: [],
all: [
Replica {
id: ReplicaKey {
topic: "topic0",
partition: 0,
},
leader: 5001,
replicas: [
5001,
],
},
],
}
```
and in short-time later
```
Oct 10 01:39:33.511 DEBUG sc_request_loop:update_replica_request: fluvio_spu::controllers::sc::dispatcher: received replica update from sc: UpdateReplicaRequest {
epoch: 0,
changes: [],
all: [
Replica {
id: ReplicaKey {
topic: "topic0",
partition: 0,
},
leader: 5001,
replicas: [
5001,
],
},
],
}
``` | 1 | SPU is receiving unnecessary replica update event from SC - steps reproduce:
```
flvt --produce-iteration 10
```
In the SPU Log, we see log with:
```
Oct 10 01:39:33.494 DEBUG sc_request_loop:update_replica_request: fluvio_spu::controllers::sc::dispatcher: received replica update from sc: UpdateReplicaRequest {
epoch: 0,
changes: [],
all: [
Replica {
id: ReplicaKey {
topic: "topic0",
partition: 0,
},
leader: 5001,
replicas: [
5001,
],
},
],
}
```
and in short-time later
```
Oct 10 01:39:33.511 DEBUG sc_request_loop:update_replica_request: fluvio_spu::controllers::sc::dispatcher: received replica update from sc: UpdateReplicaRequest {
epoch: 0,
changes: [],
all: [
Replica {
id: ReplicaKey {
topic: "topic0",
partition: 0,
},
leader: 5001,
replicas: [
5001,
],
},
],
}
``` | architecture | spu is receiving unnecessary replica update event from sc steps reproduce flvt produce iteration in the spu log we see log with oct debug sc request loop update replica request fluvio spu controllers sc dispatcher received replica update from sc updatereplicarequest epoch changes all replica id replicakey topic partition leader replicas and in short time later oct debug sc request loop update replica request fluvio spu controllers sc dispatcher received replica update from sc updatereplicarequest epoch changes all replica id replicakey topic partition leader replicas |
26 | 13,890,138,035 | IssuesEvent | 2020-10-19 08:52:55 | dusk-network/dusk-blockchain | https://api.github.com/repos/dusk-network/dusk-blockchain | closed | Review the mutex's ReadLock to prevent recursive RLock deadlock | area:architecture status:blocker type:bug type:tech-debt | Since RLock should not be used recursively, we need to review all files and make sure we do not incur in this weird situation. A way would be to avoid RLocking in unexported functions | 1 | Review the mutex's ReadLock to prevent recursive RLock deadlock - Since RLock should not be used recursively, we need to review all files and make sure we do not incur in this weird situation. A way would be to avoid RLocking in unexported functions | architecture | review the mutex s readlock to prevent recursive rlock deadlock since rlock should not be used recursively we need to review all files and make sure we do not incur in this weird situation a way would be to avoid rlocking in unexported functions |
27 | 14,234,716,834 | IssuesEvent | 2020-11-18 13:57:36 | dusk-network/dusk-blockchain | https://api.github.com/repos/dusk-network/dusk-blockchain | closed | Investigate and adjust any RPCBus messages which are currently encoded, but can be passed directly | area:architecture need:investigation type:tech-debt | So far, I have identified:
- LastCertificate
- GetRoundResults
- GetCandidate
- GetMempoolTxs | 1 | Investigate and adjust any RPCBus messages which are currently encoded, but can be passed directly - So far, I have identified:
- LastCertificate
- GetRoundResults
- GetCandidate
- GetMempoolTxs | architecture | investigate and adjust any rpcbus messages which are currently encoded but can be passed directly so far i have identified lastcertificate getroundresults getcandidate getmempooltxs |
28 | 14,258,467,990 | IssuesEvent | 2020-11-20 06:21:06 | infinyon/fluvio | https://api.github.com/repos/infinyon/fluvio | opened | tracking: Command Extension Mechanism | CLI Installation Usability doc/architecture enhancement extensions technical debt | A new "Extension" mechanism will replace the current monolithic CLI.
Existing CLI commands will move into the following extensions:
* Consumer: Topic, Partitions, Consume, Produce
* Cluster: SPU, Install, Uninstall, check
* Engines: SPU Engine, SC Engine
CLI will have commands related to extension management and profile:
* List Extension
* Install/Update/Uninstall Extensions
* Profile
* Self Update
CLI will no longer have explicit dependencies to any of the extensions.
An extension can be implemented in any language. The extensions will be be stored in the `extensions` folder in the fluvio configuration folder. CLI will query each extension so it can be show commands to the user.
| 1 | tracking: Command Extension Mechanism - A new "Extension" mechanism will replace the current monolithic CLI.
Existing CLI commands will move into the following extensions:
* Consumer: Topic, Partitions, Consume, Produce
* Cluster: SPU, Install, Uninstall, check
* Engines: SPU Engine, SC Engine
CLI will have commands related to extension management and profile:
* List Extension
* Install/Update/Uninstall Extensions
* Profile
* Self Update
CLI will no longer have explicit dependencies to any of the extensions.
An extension can be implemented in any language. The extensions will be be stored in the `extensions` folder in the fluvio configuration folder. CLI will query each extension so it can be show commands to the user.
| architecture | tracking command extension mechanism a new extension mechanism will replace the current monolithic cli existing cli commands will move into the following extensions consumer topic partitions consume produce cluster spu install uninstall check engines spu engine sc engine cli will have commands related to extension management and profile list extension install update uninstall extensions profile self update cli will no longer have explicit dependencies to any of the extensions an extension can be implemented in any language the extensions will be be stored in the extensions folder in the fluvio configuration folder cli will query each extension so it can be show commands to the user |
30 | 14,762,040,121 | IssuesEvent | 2021-01-09 01:30:56 | woocommerce/woocommerce-ios | https://api.github.com/repos/woocommerce/woocommerce-ios | closed | Review: ViewModels | Architecture Tech Debt [Type] Enhancement | ### Details:
Analyze and (if possible) replace the ViewModels with extensions / cell configuration methods. | 1 | Review: ViewModels - ### Details:
Analyze and (if possible) replace the ViewModels with extensions / cell configuration methods. | architecture | review viewmodels details analyze and if possible replace the viewmodels with extensions cell configuration methods |
32 | 15,226,667,288 | IssuesEvent | 2021-02-18 09:11:58 | dusk-network/rusk | https://api.github.com/repos/dusk-network/rusk | closed | Implement a Circuit selector based on VerifierKeys for The Host | area: genesis-contracts area:architecture area:cryptography team:Core type:refactor type:tech-debt | On the current design that is being done in #51 there is no support for handling `VerifierKey`s as arguments and match over them to use the appropiate `Circuit` struct to verify the `Proof`s on an easy way.
The goal is to match inside the `verify_proof` host function implemented inside `RuskExternals` over the hash of the `VerifierKey` (`H(VerifierKey)`) which is also the name of the file where the circuit-related data is stored.
We should also bear in mind that we will need to build some kind of map/connection between circuit structures and it's hash namefiles in order to be able to match from the hash of a `VerifierKey` into the Circuit that we want to use. | 1 | Implement a Circuit selector based on VerifierKeys for The Host - On the current design that is being done in #51 there is no support for handling `VerifierKey`s as arguments and match over them to use the appropiate `Circuit` struct to verify the `Proof`s on an easy way.
The goal is to match inside the `verify_proof` host function implemented inside `RuskExternals` over the hash of the `VerifierKey` (`H(VerifierKey)`) which is also the name of the file where the circuit-related data is stored.
We should also bear in mind that we will need to build some kind of map/connection between circuit structures and it's hash namefiles in order to be able to match from the hash of a `VerifierKey` into the Circuit that we want to use. | architecture | implement a circuit selector based on verifierkeys for the host on the current design that is being done in there is no support for handling verifierkey s as arguments and match over them to use the appropiate circuit struct to verify the proof s on an easy way the goal is to match inside the verify proof host function implemented inside ruskexternals over the hash of the verifierkey h verifierkey which is also the name of the file where the circuit related data is stored we should also bear in mind that we will need to build some kind of map connection between circuit structures and it s hash namefiles in order to be able to match from the hash of a verifierkey into the circuit that we want to use |
33 | 15,708,778,763 | IssuesEvent | 2021-03-26 21:08:43 | woocommerce/woocommerce-ios | https://api.github.com/repos/woocommerce/woocommerce-ios | closed | Networking: Invalid Token Handler | Architecture Tech Debt type: enhancement type: task | Update the the Networking layer so that any Authentication Error is properly parsed, and relayed thru a common channel.
| 1 | Networking: Invalid Token Handler - Update the the Networking layer so that any Authentication Error is properly parsed, and relayed thru a common channel.
| architecture | networking invalid token handler update the the networking layer so that any authentication error is properly parsed and relayed thru a common channel |
34 | 16,175,902,189 | IssuesEvent | 2021-05-03 06:42:23 | woocommerce/woocommerce-ios | https://api.github.com/repos/woocommerce/woocommerce-ios | closed | Allow Testing of Localized Strings | Architecture Tech Debt type: task | There are times when you would like to test a localized string. For example, you may want to make sure that the quantity and price are the values attached to the final string of this:
```swift
enum Localization {
static func subtitle(quantity: String, price: String) -> String {
let format = NSLocalizedString("%1$@ x %2$@", comment: "")
return String.localizedStringWithFormat(format, quantity, price)
}
}
```
We can technically do this now. But if the tests are running on a non-English simulator, then the test would fail because the `NSLocalizedString()` call would return the localized value.
## Possible Solutions
We'd probably have to create our own localization functions. These functions can probably allow overrides so that the tests can _expect_ a specific language to be used.
We'd probably also need to modify `localize.py` so it will not just look for `NSLocalizedString()` calls.
### Stretch Goal
It'd probably be better if we can also organize our localized strings into a single file. Kind of like how [SwiftGen does it](https://github.com/SwiftGen/SwiftGen#strings). | 1 | Allow Testing of Localized Strings - There are times when you would like to test a localized string. For example, you may want to make sure that the quantity and price are the values attached to the final string of this:
```swift
enum Localization {
static func subtitle(quantity: String, price: String) -> String {
let format = NSLocalizedString("%1$@ x %2$@", comment: "")
return String.localizedStringWithFormat(format, quantity, price)
}
}
```
We can technically do this now. But if the tests are running on a non-English simulator, then the test would fail because the `NSLocalizedString()` call would return the localized value.
## Possible Solutions
We'd probably have to create our own localization functions. These functions can probably allow overrides so that the tests can _expect_ a specific language to be used.
We'd probably also need to modify `localize.py` so it will not just look for `NSLocalizedString()` calls.
### Stretch Goal
It'd probably be better if we can also organize our localized strings into a single file. Kind of like how [SwiftGen does it](https://github.com/SwiftGen/SwiftGen#strings). | architecture | allow testing of localized strings there are times when you would like to test a localized string for example you may want to make sure that the quantity and price are the values attached to the final string of this swift enum localization static func subtitle quantity string price string string let format nslocalizedstring x comment return string localizedstringwithformat format quantity price we can technically do this now but if the tests are running on a non english simulator then the test would fail because the nslocalizedstring call would return the localized value possible solutions we d probably have to create our own localization functions these functions can probably allow overrides so that the tests can expect a specific language to be used we d probably also need to modify localize py so it will not just look for nslocalizedstring calls stretch goal it d probably be better if we can also organize our localized strings into a single file kind of like how |
35 | 17,116,841,368 | IssuesEvent | 2021-07-11 14:32:30 | spacemeshos/go-spacemesh | https://api.github.com/repos/spacemeshos/go-spacemesh | opened | Replace dynamic XDR encoding/ decoding with typed encoding using .x files | After MN architecture technical debt | ## Description
Currently we use dynamic encoding / decoding of our structs we should move to use statically defined .x files that describes the structs primitive types for better performance.
| 1 | Replace dynamic XDR encoding/ decoding with typed encoding using .x files - ## Description
Currently we use dynamic encoding / decoding of our structs we should move to use statically defined .x files that describes the structs primitive types for better performance.
| architecture | replace dynamic xdr encoding decoding with typed encoding using x files description currently we use dynamic encoding decoding of our structs we should move to use statically defined x files that describes the structs primitive types for better performance |
36 | 17,261,365,839 | IssuesEvent | 2021-07-22 08:07:52 | dusk-network/rusk-vm | https://api.github.com/repos/dusk-network/rusk-vm | opened | Reduce size of contract arguments | area:architecture team:Core type:enhancement type:tech-debt | At the moment the struct size of `Call` is _1312_, and the encoded size is _2526_.
As encoded sizes should in general be smaller than the in-memory representation, we need to understand why it is the opposite here.
One potential optimization has to do with avoiding one single `enum` for all possible call arguments. This is a waste of space, since each instance of the enum will necessarily occupy the space of the largest variant.
| 1 | Reduce size of contract arguments - At the moment the struct size of `Call` is _1312_, and the encoded size is _2526_.
As encoded sizes should in general be smaller than the in-memory representation, we need to understand why it is the opposite here.
One potential optimization has to do with avoiding one single `enum` for all possible call arguments. This is a waste of space, since each instance of the enum will necessarily occupy the space of the largest variant.
| architecture | reduce size of contract arguments at the moment the struct size of call is and the encoded size is as encoded sizes should in general be smaller than the in memory representation we need to understand why it is the opposite here one potential optimization has to do with avoiding one single enum for all possible call arguments this is a waste of space since each instance of the enum will necessarily occupy the space of the largest variant |
37 | 17,261,389,259 | IssuesEvent | 2021-07-22 08:09:44 | dusk-network/rusk-vm | https://api.github.com/repos/dusk-network/rusk-vm | opened | Reduce levels of function nesting | area:architecture team:Core type:tech-debt | At the moment, the Transfer Contract contains the `Tree` structure, which contains the `PoseidonTree`, which in turn contains the `NStack`. One hypothesis is that this could lead to nested function calls that could be avoided.
This issue is worthy of a deeper research | 1 | Reduce levels of function nesting - At the moment, the Transfer Contract contains the `Tree` structure, which contains the `PoseidonTree`, which in turn contains the `NStack`. One hypothesis is that this could lead to nested function calls that could be avoided.
This issue is worthy of a deeper research | architecture | reduce levels of function nesting at the moment the transfer contract contains the tree structure which contains the poseidontree which in turn contains the nstack one hypothesis is that this could lead to nested function calls that could be avoided this issue is worthy of a deeper research |
39 | 18,164,188,539 | IssuesEvent | 2021-09-27 13:04:58 | centerofci/mathesar | https://api.github.com/repos/centerofci/mathesar | reopened | Refactor frontend stores for table-data | type: enhancement affects: architecture affects: technical debt work: frontend status: ready restricted: maintainers | ## Purpose of this refactor:
1. Our tableData store handles colums, records and meta information needed for parameters and display purposes. It also includes the type definitions for all of them. This makes it cluttered, and there is an overlap of different concerns.
A better approach would be to split it into separate stores, and have a parent store which maintains the rest.
2. Currently, logic of certain requirements such passing parameters to table record request, refetching table data when params change etc., are done on the component. This can be side stepped and done entirely by observing store values. This gives a cleaner view/data model.
3. We need to simplify certain display specific stores such as column position map, to improve readability.
4. General improvements, separate out cells from header and row as different components.
5. Update svelte, vite and related packages. | 1 | Refactor frontend stores for table-data - ## Purpose of this refactor:
1. Our tableData store handles colums, records and meta information needed for parameters and display purposes. It also includes the type definitions for all of them. This makes it cluttered, and there is an overlap of different concerns.
A better approach would be to split it into separate stores, and have a parent store which maintains the rest.
2. Currently, logic of certain requirements such passing parameters to table record request, refetching table data when params change etc., are done on the component. This can be side stepped and done entirely by observing store values. This gives a cleaner view/data model.
3. We need to simplify certain display specific stores such as column position map, to improve readability.
4. General improvements, separate out cells from header and row as different components.
5. Update svelte, vite and related packages. | architecture | refactor frontend stores for table data purpose of this refactor our tabledata store handles colums records and meta information needed for parameters and display purposes it also includes the type definitions for all of them this makes it cluttered and there is an overlap of different concerns a better approach would be to split it into separate stores and have a parent store which maintains the rest currently logic of certain requirements such passing parameters to table record request refetching table data when params change etc are done on the component this can be side stepped and done entirely by observing store values this gives a cleaner view data model we need to simplify certain display specific stores such as column position map to improve readability general improvements separate out cells from header and row as different components update svelte vite and related packages |
42 | 19,100,061,270 | IssuesEvent | 2021-11-29 21:18:53 | department-of-veterans-affairs/va.gov-team | https://api.github.com/repos/department-of-veterans-affairs/va.gov-team | opened | Redirect Request - for VAMC lgbt web pages to new VAMC lgbtq+ web pages. | ia tech-debt VAMC-Upgrade platform-architecture-working-group | ### Story
As a Veteran I need to be redirected from old LGBT web page to new LGBTQ+ web page, so I assure I have the latest information and health services.
### Type of request
- [ ] We are retiring or taking down a page and need to redirect the URL (complete redirect section)
- [X] We are changing the URL of an existing page (complete redirect section)
- [ ] We need a custom vanity URL (complete vanity URL section)
### Implementation date
Immediately. These pages were found to have incorrect URLs on the updated webpages, the new Web pages will take affect ASAP, to mitigate any broken links for Veterans, please process redirects as soon as possible. Thank you.
### Redirects
Current URL | Redirect Destination or New URL
https://www.va.gov/miami-health-care/health-services/lgbt-veteran-care/ | https://www.va.gov/miami-health-care/health-services/lgbtq-veteran-care/
https://www.va.gov/houston-health-care/health-services/lgbt-veteran-care/ | https://www.va.gov/houston-health-care/health-services/lgbtq-veteran-care/
https://www.va.gov/montana-health-care/health-services/lgbt-veteran-care/ | https://www.va.gov/montana-health-care/health-services/lgbtq-veteran-care/
https://www.va.gov/nebraska-western-iowa-health-care/health-services/lgbt-veteran-care/ | https://www.va.gov/nebraska-western-iowa-health-care/health-services/lgbtq-veteran-care/
https://www.va.gov/new-jersey-health-care/health-services/lgbt-veteran-care/ | https://www.va.gov/new-jersey-health-care/health-services/lgbtq-veteran-care/
https://www.va.gov/north-florida-health-care/health-services/lgbt-veteran-care/ | https://www.va.gov/north-florida-health-care/health-services/lgbtq-veteran-care/
https://www.va.gov/northern-california-health-care/health-services/lgbt-veteran-care/ | https://www.va.gov/northern-california-health-care/health-services/lgbtq-veteran-care/
https://www.va.gov/oklahoma-city-health-care/health-services/lgbt-veteran-care/ | https://www.va.gov/oklahoma-city-health-care/health-services/lgbtq-veteran-care/
https://www.va.gov/orlando-health-care/health-services/lgbt-veteran-care/ | https://www.va.gov/orlando-health-care/health-services/lgbtq-veteran-care/
https://www.va.gov/shreveport-health-care/health-services/lgbt-veteran-care/ | https://www.va.gov/shreveport-health-care/health-services/lgbtq-veteran-care/
https://www.va.gov/pacific-islands-health-care/health-services/lgbt-veteran-care/ | https://www.va.gov/pacific-islands-health-care/health-services/lgbtq-veteran-care/
https://www.va.gov/palo-alto-health-care/health-services/lgbt-veteran-care/ | https://www.va.gov/palo-alto-health-care/health-services/lgbtq-veteran-care/
https://www.va.gov/pittsburgh-health-care/health-services/lgbt-veteran-care/ | https://www.va.gov/pittsburgh-health-care/health-services/lgbtq-veteran-care/
https://www.va.gov/salem-health-care/health-services/lgbt-veteran-care/ | https://www.va.gov/salem-health-care/health-services/lgbtq-veteran-care/
https://www.va.gov/salisbury-health-care/health-services/lgbt-veteran-care/ | https://www.va.gov/salisbury-health-care/health-services/lgbtq-veteran-care/
https://www.va.gov/salt-lake-city-health-care/health-services/lgbt-veteran-care/ | https://www.va.gov/salt-lake-city-health-care/health-services/lgbtq-veteran-care/
https://www.va.gov/san-francisco-health-care/health-services/lgbt-veteran-care/ | https://www.va.gov/san-francisco-health-care/health-services/lgbtq-veteran-care/
https://www.va.gov/sioux-falls-health-care/health-services/lgbt-veteran-care/ | https://www.va.gov/sioux-falls-health-care/health-services/lgbtq-veteran-care/
https://www.va.gov/southeast-louisiana-health-care/health-services/lgbt-veteran-care/ | https://www.va.gov/southeast-louisiana-health-care/health-services/lgbtq-veteran-care/
https://www.va.gov/southern-nevada-health-care/health-services/lgbt-veteran-care/ | https://www.va.gov/southern-nevada-health-care/health-services/lgbtq-veteran-care/
https://www.va.gov/syracuse-health-care/health-services/lgbt-veteran-care/ | https://www.va.gov/syracuse-health-care/health-services/lgbtq-veteran-care/
https://www.va.gov/tuscaloosa-health-care/health-services/lgbt-veteran-care/ | https://www.va.gov/tuscaloosa-health-care/health-services/lgbtq-veteran-care/
https://www.va.gov/fayetteville-arkansas-health-care/health-services/lgbt-veteran-care/ | https://www.va.gov/fayetteville-arkansas-health-care/health-services/lgbtq-veteran-care/
https://www.va.gov/west-palm-beach-health-care/health-services/lgbt-veteran-care/ | https://www.va.gov/west-palm-beach-health-care/health-services/lgbtq-veteran-care/
### Vanity URLs
N/A
**Link to campaign landing page request issue:**
### Process, Roles and Responsibilities
- [x] Requesting team: Above information is provided
- [x] Requesting team: All appropriate VA stakeholders are notified as appropriate
- [ ] IA: Request is vetted and documented and implementation plan is clear
- [ ] IA: Request is assigned to appropriate team for implementation
- [ ] Implementation team: Work is complete
- [ ] Implementation team: Validated in production
- [ ] Requesting team: Validates everything is correct in production and closes ticket
| 1 | Redirect Request - for VAMC lgbt web pages to new VAMC lgbtq+ web pages. - ### Story
As a Veteran I need to be redirected from old LGBT web page to new LGBTQ+ web page, so I assure I have the latest information and health services.
### Type of request
- [ ] We are retiring or taking down a page and need to redirect the URL (complete redirect section)
- [X] We are changing the URL of an existing page (complete redirect section)
- [ ] We need a custom vanity URL (complete vanity URL section)
### Implementation date
Immediately. These pages were found to have incorrect URLs on the updated webpages, the new Web pages will take affect ASAP, to mitigate any broken links for Veterans, please process redirects as soon as possible. Thank you.
### Redirects
Current URL | Redirect Destination or New URL
https://www.va.gov/miami-health-care/health-services/lgbt-veteran-care/ | https://www.va.gov/miami-health-care/health-services/lgbtq-veteran-care/
https://www.va.gov/houston-health-care/health-services/lgbt-veteran-care/ | https://www.va.gov/houston-health-care/health-services/lgbtq-veteran-care/
https://www.va.gov/montana-health-care/health-services/lgbt-veteran-care/ | https://www.va.gov/montana-health-care/health-services/lgbtq-veteran-care/
https://www.va.gov/nebraska-western-iowa-health-care/health-services/lgbt-veteran-care/ | https://www.va.gov/nebraska-western-iowa-health-care/health-services/lgbtq-veteran-care/
https://www.va.gov/new-jersey-health-care/health-services/lgbt-veteran-care/ | https://www.va.gov/new-jersey-health-care/health-services/lgbtq-veteran-care/
https://www.va.gov/north-florida-health-care/health-services/lgbt-veteran-care/ | https://www.va.gov/north-florida-health-care/health-services/lgbtq-veteran-care/
https://www.va.gov/northern-california-health-care/health-services/lgbt-veteran-care/ | https://www.va.gov/northern-california-health-care/health-services/lgbtq-veteran-care/
https://www.va.gov/oklahoma-city-health-care/health-services/lgbt-veteran-care/ | https://www.va.gov/oklahoma-city-health-care/health-services/lgbtq-veteran-care/
https://www.va.gov/orlando-health-care/health-services/lgbt-veteran-care/ | https://www.va.gov/orlando-health-care/health-services/lgbtq-veteran-care/
https://www.va.gov/shreveport-health-care/health-services/lgbt-veteran-care/ | https://www.va.gov/shreveport-health-care/health-services/lgbtq-veteran-care/
https://www.va.gov/pacific-islands-health-care/health-services/lgbt-veteran-care/ | https://www.va.gov/pacific-islands-health-care/health-services/lgbtq-veteran-care/
https://www.va.gov/palo-alto-health-care/health-services/lgbt-veteran-care/ | https://www.va.gov/palo-alto-health-care/health-services/lgbtq-veteran-care/
https://www.va.gov/pittsburgh-health-care/health-services/lgbt-veteran-care/ | https://www.va.gov/pittsburgh-health-care/health-services/lgbtq-veteran-care/
https://www.va.gov/salem-health-care/health-services/lgbt-veteran-care/ | https://www.va.gov/salem-health-care/health-services/lgbtq-veteran-care/
https://www.va.gov/salisbury-health-care/health-services/lgbt-veteran-care/ | https://www.va.gov/salisbury-health-care/health-services/lgbtq-veteran-care/
https://www.va.gov/salt-lake-city-health-care/health-services/lgbt-veteran-care/ | https://www.va.gov/salt-lake-city-health-care/health-services/lgbtq-veteran-care/
https://www.va.gov/san-francisco-health-care/health-services/lgbt-veteran-care/ | https://www.va.gov/san-francisco-health-care/health-services/lgbtq-veteran-care/
https://www.va.gov/sioux-falls-health-care/health-services/lgbt-veteran-care/ | https://www.va.gov/sioux-falls-health-care/health-services/lgbtq-veteran-care/
https://www.va.gov/southeast-louisiana-health-care/health-services/lgbt-veteran-care/ | https://www.va.gov/southeast-louisiana-health-care/health-services/lgbtq-veteran-care/
https://www.va.gov/southern-nevada-health-care/health-services/lgbt-veteran-care/ | https://www.va.gov/southern-nevada-health-care/health-services/lgbtq-veteran-care/
https://www.va.gov/syracuse-health-care/health-services/lgbt-veteran-care/ | https://www.va.gov/syracuse-health-care/health-services/lgbtq-veteran-care/
https://www.va.gov/tuscaloosa-health-care/health-services/lgbt-veteran-care/ | https://www.va.gov/tuscaloosa-health-care/health-services/lgbtq-veteran-care/
https://www.va.gov/fayetteville-arkansas-health-care/health-services/lgbt-veteran-care/ | https://www.va.gov/fayetteville-arkansas-health-care/health-services/lgbtq-veteran-care/
https://www.va.gov/west-palm-beach-health-care/health-services/lgbt-veteran-care/ | https://www.va.gov/west-palm-beach-health-care/health-services/lgbtq-veteran-care/
### Vanity URLs
N/A
**Link to campaign landing page request issue:**
### Process, Roles and Responsibilities
- [x] Requesting team: Above information is provided
- [x] Requesting team: All appropriate VA stakeholders are notified as appropriate
- [ ] IA: Request is vetted and documented and implementation plan is clear
- [ ] IA: Request is assigned to appropriate team for implementation
- [ ] Implementation team: Work is complete
- [ ] Implementation team: Validated in production
- [ ] Requesting team: Validates everything is correct in production and closes ticket
| architecture | redirect request for vamc lgbt web pages to new vamc lgbtq web pages story as a veteran i need to be redirected from old lgbt web page to new lgbtq web page so i assure i have the latest information and health services type of request we are retiring or taking down a page and need to redirect the url complete redirect section we are changing the url of an existing page complete redirect section we need a custom vanity url complete vanity url section implementation date immediately these pages were found to have incorrect urls on the updated webpages the new web pages will take affect asap to mitigate any broken links for veterans please process redirects as soon as possible thank you redirects current url redirect destination or new url vanity urls n a link to campaign landing page request issue process roles and responsibilities requesting team above information is provided requesting team all appropriate va stakeholders are notified as appropriate ia request is vetted and documented and implementation plan is clear ia request is assigned to appropriate team for implementation implementation team work is complete implementation team validated in production requesting team validates everything is correct in production and closes ticket |
43 | 19,287,426,961 | IssuesEvent | 2021-12-11 07:05:14 | spacemeshos/go-spacemesh | https://api.github.com/repos/spacemeshos/go-spacemesh | closed | Async/on-demand tortoise | Tortoise Protocol technical debt architecture Before MN | Right now, tortoise receives data in several ways:
- when Hare finishes processing a layer, its output is sent to `tortoise.HandleIncomingLayer` via `mesh.ValidateLayer`
- when a layer is received via sync, the same thing happens
- when a late block arrives via sync or gossip, it's sent individually into `tortoise.HandleLateBlocks`
All of these are currently handled synchronously: e.g., the syncer is waiting for tortoise to finish running and processing the new layer before it continues syncing.
The tortoise should be totally asynchronous. It should be a separate, autonomous "background process" in its own goroutine (like the hare broker). All incoming data should be buffered on channels, and the caller should never wait for it to finish. This would have a few advantages:
- easier to process incoming blocks and layers in batches (especially useful for late blocks, see #2412). This is the main motivation for this change. The way in which tortoise processes data should not be tightly coupled to the messages it receives about new blocks and layers.
- architecturally, it makes more sense: there's no reason hare, mesh, or syncer should block on tortoise. It makes the API simpler.
- makes it a bit simpler to rerun the verifying tortoise from scratch periodically, on its own schedule (without worrying about blocking data providers), or to trigger tortoise once in a while when an accounting of the voting weight of incoming blocks requires it
No block data should be passed directly into tortoise (this was done already in #2400). Tortoise should just receive notifications that new data is waiting to be processed. The API should basically just be:
- new incoming layer
- late block received | 1 | Async/on-demand tortoise - Right now, tortoise receives data in several ways:
- when Hare finishes processing a layer, its output is sent to `tortoise.HandleIncomingLayer` via `mesh.ValidateLayer`
- when a layer is received via sync, the same thing happens
- when a late block arrives via sync or gossip, it's sent individually into `tortoise.HandleLateBlocks`
All of these are currently handled synchronously: e.g., the syncer is waiting for tortoise to finish running and processing the new layer before it continues syncing.
The tortoise should be totally asynchronous. It should be a separate, autonomous "background process" in its own goroutine (like the hare broker). All incoming data should be buffered on channels, and the caller should never wait for it to finish. This would have a few advantages:
- easier to process incoming blocks and layers in batches (especially useful for late blocks, see #2412). This is the main motivation for this change. The way in which tortoise processes data should not be tightly coupled to the messages it receives about new blocks and layers.
- architecturally, it makes more sense: there's no reason hare, mesh, or syncer should block on tortoise. It makes the API simpler.
- makes it a bit simpler to rerun the verifying tortoise from scratch periodically, on its own schedule (without worrying about blocking data providers), or to trigger tortoise once in a while when an accounting of the voting weight of incoming blocks requires it
No block data should be passed directly into tortoise (this was done already in #2400). Tortoise should just receive notifications that new data is waiting to be processed. The API should basically just be:
- new incoming layer
- late block received | architecture | async on demand tortoise right now tortoise receives data in several ways when hare finishes processing a layer its output is sent to tortoise handleincominglayer via mesh validatelayer when a layer is received via sync the same thing happens when a late block arrives via sync or gossip it s sent individually into tortoise handlelateblocks all of these are currently handled synchronously e g the syncer is waiting for tortoise to finish running and processing the new layer before it continues syncing the tortoise should be totally asynchronous it should be a separate autonomous background process in its own goroutine like the hare broker all incoming data should be buffered on channels and the caller should never wait for it to finish this would have a few advantages easier to process incoming blocks and layers in batches especially useful for late blocks see this is the main motivation for this change the way in which tortoise processes data should not be tightly coupled to the messages it receives about new blocks and layers architecturally it makes more sense there s no reason hare mesh or syncer should block on tortoise it makes the api simpler makes it a bit simpler to rerun the verifying tortoise from scratch periodically on its own schedule without worrying about blocking data providers or to trigger tortoise once in a while when an accounting of the voting weight of incoming blocks requires it no block data should be passed directly into tortoise this was done already in tortoise should just receive notifications that new data is waiting to be processed the api should basically just be new incoming layer late block received |
44 | 19,292,339,503 | IssuesEvent | 2021-12-12 01:40:39 | centerofci/mathesar | https://api.github.com/repos/centerofci/mathesar | closed | Define common error structure | type: bug affects: architecture affects: technical debt work: backend status: review | ## Description
Currently, there is no common error structure on the backend.
* Most requests with bad input return 500, while they should return 400.
* The current 400 errors do not follow a common structure. In some observed cases:
* they return a JSON array of strings
* they don't return anything
* they return a JSON object with a property named `detail`
## Expected behaviour
* There needs to be a well defined error structure that is common for all requests.
* Errors with bad input should only return status code 400.
| 1 | Define common error structure - ## Description
Currently, there is no common error structure on the backend.
* Most requests with bad input return 500, while they should return 400.
* The current 400 errors do not follow a common structure. In some observed cases:
* they return a JSON array of strings
* they don't return anything
* they return a JSON object with a property named `detail`
## Expected behaviour
* There needs to be a well defined error structure that is common for all requests.
* Errors with bad input should only return status code 400.
| architecture | define common error structure description currently there is no common error structure on the backend most requests with bad input return while they should return the current errors do not follow a common structure in some observed cases they return a json array of strings they don t return anything they return a json object with a property named detail expected behaviour there needs to be a well defined error structure that is common for all requests errors with bad input should only return status code |
45 | 20,367,093,769 | IssuesEvent | 2022-02-21 07:23:25 | centerofci/mathesar | https://api.github.com/repos/centerofci/mathesar | opened | Assess database type infrastructure | type: enhancement affects: dx affects: architecture affects: technical debt work: backend work: database status: draft type: meta | ## Problem
<!-- Please provide a clear and concise description of the problem that this feature request is designed to solve.-->
As we near the end of the initial round of implementing database types, we should reassess our current setup and look for improvements. Especially since some GSoC projects involve adding types or adding features to types.
## Proposed solution
<!-- A clear and concise description of your proposed solution or feature. -->
We should make it as easy and obvious as possible to
- Add a new custom type
- Add support for native PostgreSQL types
- Add support for common 3rd-party types (e.g., PostGIS)
- add support for `type_options` for a type.
## Additional context
<!-- Add any other context or screenshots about the feature request here.-->
This is a draft while we figure out what improvements might actually be possible. | 1 | Assess database type infrastructure - ## Problem
<!-- Please provide a clear and concise description of the problem that this feature request is designed to solve.-->
As we near the end of the initial round of implementing database types, we should reassess our current setup and look for improvements. Especially since some GSoC projects involve adding types or adding features to types.
## Proposed solution
<!-- A clear and concise description of your proposed solution or feature. -->
We should make it as easy and obvious as possible to
- Add a new custom type
- Add support for native PostgreSQL types
- Add support for common 3rd-party types (e.g., PostGIS)
- add support for `type_options` for a type.
## Additional context
<!-- Add any other context or screenshots about the feature request here.-->
This is a draft while we figure out what improvements might actually be possible. | architecture | assess database type infrastructure problem as we near the end of the initial round of implementing database types we should reassess our current setup and look for improvements especially since some gsoc projects involve adding types or adding features to types proposed solution we should make it as easy and obvious as possible to add a new custom type add support for native postgresql types add support for common party types e g postgis add support for type options for a type additional context this is a draft while we figure out what improvements might actually be possible |
46 | 20,753,135,542 | IssuesEvent | 2022-03-15 09:39:51 | woocommerce/woocommerce-android | https://api.github.com/repos/woocommerce/woocommerce-android | closed | Remove unused Order list FluxC code | type: task feature: order list category: architecture type: technical debt | Once #1559 is in production for a bit and deemed stable, remove the old unused order list code in FluxC. | 1 | Remove unused Order list FluxC code - Once #1559 is in production for a bit and deemed stable, remove the old unused order list code in FluxC. | architecture | remove unused order list fluxc code once is in production for a bit and deemed stable remove the old unused order list code in fluxc |
47 | 20,991,626,000 | IssuesEvent | 2022-03-29 09:48:28 | woocommerce/woocommerce-android | https://api.github.com/repos/woocommerce/woocommerce-android | opened | Refactor SitePicker screens to MVVM | category: architecture type: technical debt | This master issue lists the changes we need to migrate the site picker screen to use MVVM.
## FluxC
- Migrate fetch supported woo version methods into a suspendable function.
- Migrate fetch site settings methods into a suspendable function.
- Migrate fetch site product settings methods into a suspendable function.
## Woo
- Update existing methods in Woo that fetch site settings & product settings remove usages of [deprecated event bus architecture](https://github.com/wordpress-mobile/WordPress-FluxC-Android/wiki/%5BDeprecated%5D-Architecture).
- Since the `SIGN OUT` action is the only action in the site picker that is utilising event bus, I thought of moving it to a separate class for now. It looks like the `AccountStore` is completely in Java and migrating the sign out action to suspendable is out of scope for this task.
- Create a `SitePickerRepository` to communicate with FluxC to fetch site related details from the API.
- Create a `ViewModel` that handles business logic for the Site picker screens.
- Update the fragment classes and remove legacy code. | 1 | Refactor SitePicker screens to MVVM - This master issue lists the changes we need to migrate the site picker screen to use MVVM.
## FluxC
- Migrate fetch supported woo version methods into a suspendable function.
- Migrate fetch site settings methods into a suspendable function.
- Migrate fetch site product settings methods into a suspendable function.
## Woo
- Update existing methods in Woo that fetch site settings & product settings remove usages of [deprecated event bus architecture](https://github.com/wordpress-mobile/WordPress-FluxC-Android/wiki/%5BDeprecated%5D-Architecture).
- Since the `SIGN OUT` action is the only action in the site picker that is utilising event bus, I thought of moving it to a separate class for now. It looks like the `AccountStore` is completely in Java and migrating the sign out action to suspendable is out of scope for this task.
- Create a `SitePickerRepository` to communicate with FluxC to fetch site related details from the API.
- Create a `ViewModel` that handles business logic for the Site picker screens.
- Update the fragment classes and remove legacy code. | architecture | refactor sitepicker screens to mvvm this master issue lists the changes we need to migrate the site picker screen to use mvvm fluxc migrate fetch supported woo version methods into a suspendable function migrate fetch site settings methods into a suspendable function migrate fetch site product settings methods into a suspendable function woo update existing methods in woo that fetch site settings product settings remove usages of since the sign out action is the only action in the site picker that is utilising event bus i thought of moving it to a separate class for now it looks like the accountstore is completely in java and migrating the sign out action to suspendable is out of scope for this task create a sitepickerrepository to communicate with fluxc to fetch site related details from the api create a viewmodel that handles business logic for the site picker screens update the fragment classes and remove legacy code |
48 | 22,267,621,896 | IssuesEvent | 2022-06-10 09:02:14 | woocommerce/woocommerce-ios | https://api.github.com/repos/woocommerce/woocommerce-ios | closed | Direct access to App Settings | category: architecture type: technical debt | Currently, every time we want to read/write from app settings we have to dispatch an `AppSettingsAction` and go through `AppSettingsStore`. The Flux-like model we have in Yosemite has worked well for dealing with data that is stored in Core Data, but the abstraction is making other data access harder.
In the case of app settings, having to though Yosemite means that, for each new setting, we have to:
1. Create a setter action to save a new value
2. Create a getter action with a completion block because actions can't return data
3. Because of that ☝🏽, reading an app setting becomes an async operation, when the reality is that we load settings synchronously
This becomes specially manifest when we are dealing with Experimental Features, which we store as app settings. For instance, [a recent PR](https://github.com/woocommerce/woocommerce-ios/pull/6954) to remove an experimental feature flag ended up removing 265 lines of code.
As another example, when we started supporting IPP in Canada, both the WCPay support for Canada, and Stripe support for the US were behind a feature flag. This meant that our onboarding code had to take both flags into account (among other things), but couldn't read their values synchronously and had to get the stores manager injected, [making the code more complicated than necessary](https://github.com/woocommerce/woocommerce-ios/blob/7f592ed183523205dba8ffdbc683bff0770643ee/WooCommerce/Classes/ViewRelated/Dashboard/Settings/In-Person%20Payments/CardPresentPaymentsOnboardingUseCase.swift#L42).
I think it's time to get app settings out of the stores layer and directly accessible via `ServiceLocator`
| 1 | Direct access to App Settings - Currently, every time we want to read/write from app settings we have to dispatch an `AppSettingsAction` and go through `AppSettingsStore`. The Flux-like model we have in Yosemite has worked well for dealing with data that is stored in Core Data, but the abstraction is making other data access harder.
In the case of app settings, having to though Yosemite means that, for each new setting, we have to:
1. Create a setter action to save a new value
2. Create a getter action with a completion block because actions can't return data
3. Because of that ☝🏽, reading an app setting becomes an async operation, when the reality is that we load settings synchronously
This becomes specially manifest when we are dealing with Experimental Features, which we store as app settings. For instance, [a recent PR](https://github.com/woocommerce/woocommerce-ios/pull/6954) to remove an experimental feature flag ended up removing 265 lines of code.
As another example, when we started supporting IPP in Canada, both the WCPay support for Canada, and Stripe support for the US were behind a feature flag. This meant that our onboarding code had to take both flags into account (among other things), but couldn't read their values synchronously and had to get the stores manager injected, [making the code more complicated than necessary](https://github.com/woocommerce/woocommerce-ios/blob/7f592ed183523205dba8ffdbc683bff0770643ee/WooCommerce/Classes/ViewRelated/Dashboard/Settings/In-Person%20Payments/CardPresentPaymentsOnboardingUseCase.swift#L42).
I think it's time to get app settings out of the stores layer and directly accessible via `ServiceLocator`
| architecture | direct access to app settings currently every time we want to read write from app settings we have to dispatch an appsettingsaction and go through appsettingsstore the flux like model we have in yosemite has worked well for dealing with data that is stored in core data but the abstraction is making other data access harder in the case of app settings having to though yosemite means that for each new setting we have to create a setter action to save a new value create a getter action with a completion block because actions can t return data because of that ☝🏽 reading an app setting becomes an async operation when the reality is that we load settings synchronously this becomes specially manifest when we are dealing with experimental features which we store as app settings for instance to remove an experimental feature flag ended up removing lines of code as another example when we started supporting ipp in canada both the wcpay support for canada and stripe support for the us were behind a feature flag this meant that our onboarding code had to take both flags into account among other things but couldn t read their values synchronously and had to get the stores manager injected i think it s time to get app settings out of the stores layer and directly accessible via servicelocator |
49 | 23,189,582,559 | IssuesEvent | 2022-08-01 11:26:35 | woocommerce/woocommerce-android | https://api.github.com/repos/woocommerce/woocommerce-android | closed | IOException: PHONE_REGISTRATION_ERROR | category: architecture type: technical debt | Sentry Issue: [WOOCOMMERCE-ANDROID-23R](https://sentry.io/organizations/a8c/issues/2574418765/?referrer=github_integration).
First seen in 7.3-rc-2 on August 13, log shows "Fetching FCM registration token failed" and "FIS_AUTH_ERROR."
```
IOException: PHONE_REGISTRATION_ERROR
at com.google.firebase.iid.GmsRpc.handleResponse(com.google.firebase:firebase-iid@@21.0.0:84)
at com.google.firebase.iid.GmsRpc.lambda$extractResponseWhenComplete$0$GmsRpc(com.google.firebase:firebase-iid@@21.0.0:94)
at com.google.firebase.iid.GmsRpc$$Lambda$0.then
at com.google.android.gms.tasks.zzd.run
at com.google.firebase.iid.FirebaseIidExecutors$$Lambda$0.execute
...
(19 additional frame(s) were not displayed)
Fetching FCM registration token failed
``` | 1 | IOException: PHONE_REGISTRATION_ERROR - Sentry Issue: [WOOCOMMERCE-ANDROID-23R](https://sentry.io/organizations/a8c/issues/2574418765/?referrer=github_integration).
First seen in 7.3-rc-2 on August 13, log shows "Fetching FCM registration token failed" and "FIS_AUTH_ERROR."
```
IOException: PHONE_REGISTRATION_ERROR
at com.google.firebase.iid.GmsRpc.handleResponse(com.google.firebase:firebase-iid@@21.0.0:84)
at com.google.firebase.iid.GmsRpc.lambda$extractResponseWhenComplete$0$GmsRpc(com.google.firebase:firebase-iid@@21.0.0:94)
at com.google.firebase.iid.GmsRpc$$Lambda$0.then
at com.google.android.gms.tasks.zzd.run
at com.google.firebase.iid.FirebaseIidExecutors$$Lambda$0.execute
...
(19 additional frame(s) were not displayed)
Fetching FCM registration token failed
``` | architecture | ioexception phone registration error sentry issue first seen in rc on august log shows fetching fcm registration token failed and fis auth error ioexception phone registration error at com google firebase iid gmsrpc handleresponse com google firebase firebase iid at com google firebase iid gmsrpc lambda extractresponsewhencomplete gmsrpc com google firebase firebase iid at com google firebase iid gmsrpc lambda then at com google android gms tasks zzd run at com google firebase iid firebaseiidexecutors lambda execute additional frame s were not displayed fetching fcm registration token failed |
50 | 23,676,730,498 | IssuesEvent | 2022-08-28 07:33:10 | eurofurence/ef-app_ios | https://api.github.com/repos/eurofurence/ef-app_ios | opened | Migration to SwiftUI | enhancement rearchitecture technical debt | Aim to improve the overall architecture of the app (with respect to modern Cocoa development) with the side benefit of making it easier to understand through simplification of layers.
Rather than do a full rewrite (and risk losing time for next year/introducing a slew of bugs) we should do this in stages:
- Port existing views to use SwiftUI, each view consuming a view model dependency. Implementations of the dependency bridge into the existing model. Leave the existing routing tier as it is, with component factories returning UIHostingController objects wrapping the SwiftUI views.
- Drop abstraction layers in model around objects and expose new NSManagedObject subclasses for entities. The model package will continue to own how we create, update and fetch them. Gradually move more behaviour into these objects (as with the current model refactor).
- When our minimum OS dependency hits iOS 15, directly read objects into views using SectionFetchRequest and drop the view models.
All the satellite app services that use the model but do not have a visual presence in the app - e.g. notification scheduling - can continue to be driven with tests using the model. These consumers of the model can act as a sanity check for the APIs as we mov things around. | 1 | Migration to SwiftUI - Aim to improve the overall architecture of the app (with respect to modern Cocoa development) with the side benefit of making it easier to understand through simplification of layers.
Rather than do a full rewrite (and risk losing time for next year/introducing a slew of bugs) we should do this in stages:
- Port existing views to use SwiftUI, each view consuming a view model dependency. Implementations of the dependency bridge into the existing model. Leave the existing routing tier as it is, with component factories returning UIHostingController objects wrapping the SwiftUI views.
- Drop abstraction layers in model around objects and expose new NSManagedObject subclasses for entities. The model package will continue to own how we create, update and fetch them. Gradually move more behaviour into these objects (as with the current model refactor).
- When our minimum OS dependency hits iOS 15, directly read objects into views using SectionFetchRequest and drop the view models.
All the satellite app services that use the model but do not have a visual presence in the app - e.g. notification scheduling - can continue to be driven with tests using the model. These consumers of the model can act as a sanity check for the APIs as we mov things around. | architecture | migration to swiftui aim to improve the overall architecture of the app with respect to modern cocoa development with the side benefit of making it easier to understand through simplification of layers rather than do a full rewrite and risk losing time for next year introducing a slew of bugs we should do this in stages port existing views to use swiftui each view consuming a view model dependency implementations of the dependency bridge into the existing model leave the existing routing tier as it is with component factories returning uihostingcontroller objects wrapping the swiftui views drop abstraction layers in model around objects and expose new nsmanagedobject subclasses for entities the model package will continue to own how we create update and fetch them gradually move more behaviour into these objects as with the current model refactor when our minimum os dependency hits ios directly read objects into views using sectionfetchrequest and drop the view models all the satellite app services that use the model but do not have a visual presence in the app e g notification scheduling can continue to be driven with tests using the model these consumers of the model can act as a sanity check for the apis as we mov things around |
51 | 23,754,645,409 | IssuesEvent | 2022-09-01 01:04:28 | meltano/sdk | https://api.github.com/repos/meltano/sdk | closed | Add reference paginator implementations | architecture decision kind/Tech Debt valuestream/SDK migrated from gitlab | Migrated from GitLab: https://gitlab.com/meltano/sdk/-/issues/318
Originally created by @edgarrmondragon on 2022-01-29 01:34:13
---
## Summary
[//]: # (Concisely summarize the feature you are proposing.)
Add reference and common pagination implementations in a similar fashion to _authenticators_.
## Proposed benefits
[//]: # (Concisely summarize the benefits this feature would bring to yourself and other users.)
There is a limited number of pagination standards, which may only differ in minute details, much like authentication headers (`Authorization: Bearer <token>`, `Authorization: Token <>`), so having a sensible set of pre-built implementations may simplify things further for users by allowing them to pick one off-the-shelf.
Another benefit is that having dedicated pagination classes makes unit-testing them much easier.
Yet another benefit of moving to a dedicated class for pagination, is that the paginator state doesn't need to be limited to the previous value but can include arbitrary attributes, like the last-seen record (https://gitlab.com/meltano/sdk/-/issues/124+).
## Proposal details
[//]: # (In as much detail as you are able, describe the feature you'd like to build or would like to see built.)
I have a reference implementation [here](https://github.com/edgarrmondragon/tap-readthedocs/pull/13/files).
**TL;DR**
```python
class APIPaginator:
"""An API paginator object."""
@property
def current_value(self) -> TPageToken:
"""Get the current pagination value."""
...
@property
def finished(self) -> bool:
"""Get a flag that indicates if the last page of data has been reached."""
...
@property
def count(self) -> int:
"""Count the number of pages traversed so far."""
...
def advance(self, response: Response) -> None:
"""Get a new page value and advance the current one."""
...
def has_more(self, response: Response) -> bool:
"""Override this method to check if the endpoint has any pages left."""
...
@abstractmethod
def get_next(self, response: Response) -> Optional[TPageToken]:
"""Get the next pagination token or index from the API response."""
...
```
## Best reasons not to build
[//]: # (Will this negatively affect any existing functionality? Do you anticipate any breaking changes versus what may already be working today? Make the counter-argument to your proposal here.)
Can't think of any. The current `RESTStream.get_next_page_token` can be slowly deprecated with the introduction of a paginator that wraps the stream (as in [tap-readthedocs/client.py at 09dca8c653cd73e51ce265e239c94c68479481b1 · edgarrmondragon/tap-readthedocs · GitHub](https://github.com/edgarrmondragon/tap-readthedocs/blob/09dca8c653cd73e51ce265e239c94c68479481b1/tap_readthedocs/client.py#L17-L48)). | 1 | Add reference paginator implementations - Migrated from GitLab: https://gitlab.com/meltano/sdk/-/issues/318
Originally created by @edgarrmondragon on 2022-01-29 01:34:13
---
## Summary
[//]: # (Concisely summarize the feature you are proposing.)
Add reference and common pagination implementations in a similar fashion to _authenticators_.
## Proposed benefits
[//]: # (Concisely summarize the benefits this feature would bring to yourself and other users.)
There is a limited number of pagination standards, which may only differ in minute details, much like authentication headers (`Authorization: Bearer <token>`, `Authorization: Token <>`), so having a sensible set of pre-built implementations may simplify things further for users by allowing them to pick one off-the-shelf.
Another benefit is that having dedicated pagination classes makes unit-testing them much easier.
Yet another benefit of moving to a dedicated class for pagination, is that the paginator state doesn't need to be limited to the previous value but can include arbitrary attributes, like the last-seen record (https://gitlab.com/meltano/sdk/-/issues/124+).
## Proposal details
[//]: # (In as much detail as you are able, describe the feature you'd like to build or would like to see built.)
I have a reference implementation [here](https://github.com/edgarrmondragon/tap-readthedocs/pull/13/files).
**TL;DR**
```python
class APIPaginator:
"""An API paginator object."""
@property
def current_value(self) -> TPageToken:
"""Get the current pagination value."""
...
@property
def finished(self) -> bool:
"""Get a flag that indicates if the last page of data has been reached."""
...
@property
def count(self) -> int:
"""Count the number of pages traversed so far."""
...
def advance(self, response: Response) -> None:
"""Get a new page value and advance the current one."""
...
def has_more(self, response: Response) -> bool:
"""Override this method to check if the endpoint has any pages left."""
...
@abstractmethod
def get_next(self, response: Response) -> Optional[TPageToken]:
"""Get the next pagination token or index from the API response."""
...
```
## Best reasons not to build
[//]: # (Will this negatively affect any existing functionality? Do you anticipate any breaking changes versus what may already be working today? Make the counter-argument to your proposal here.)
Can't think of any. The current `RESTStream.get_next_page_token` can be slowly deprecated with the introduction of a paginator that wraps the stream (as in [tap-readthedocs/client.py at 09dca8c653cd73e51ce265e239c94c68479481b1 · edgarrmondragon/tap-readthedocs · GitHub](https://github.com/edgarrmondragon/tap-readthedocs/blob/09dca8c653cd73e51ce265e239c94c68479481b1/tap_readthedocs/client.py#L17-L48)). | architecture | add reference paginator implementations migrated from gitlab originally created by edgarrmondragon on summary concisely summarize the feature you are proposing add reference and common pagination implementations in a similar fashion to authenticators proposed benefits concisely summarize the benefits this feature would bring to yourself and other users there is a limited number of pagination standards which may only differ in minute details much like authentication headers authorization bearer authorization token so having a sensible set of pre built implementations may simplify things further for users by allowing them to pick one off the shelf another benefit is that having dedicated pagination classes makes unit testing them much easier yet another benefit of moving to a dedicated class for pagination is that the paginator state doesn t need to be limited to the previous value but can include arbitrary attributes like the last seen record proposal details in as much detail as you are able describe the feature you d like to build or would like to see built i have a reference implementation tl dr python class apipaginator an api paginator object property def current value self tpagetoken get the current pagination value property def finished self bool get a flag that indicates if the last page of data has been reached property def count self int count the number of pages traversed so far def advance self response response none get a new page value and advance the current one def has more self response response bool override this method to check if the endpoint has any pages left abstractmethod def get next self response response optional get the next pagination token or index from the api response best reasons not to build will this negatively affect any existing functionality do you anticipate any breaking changes versus what may already be working today make the counter argument to your proposal here can t think of any the current reststream get next page token can be slowly deprecated with the introduction of a paginator that wraps the stream as in |
52 | 24,159,351,708 | IssuesEvent | 2022-09-22 10:17:16 | centerofci/mathesar | https://api.github.com/repos/centerofci/mathesar | opened | Improve readability of column moving logic | type: enhancement affects: architecture affects: technical debt status: triage | ## Problem
Our column moving logic is difficult to get into. There's a related bug I've encountered and am of the opinion that a focused effort to improve column moving logic's readability is warranted.
## Proposed solution
@silentninja proposed that we do this together on a call.
## Additional context
@silentninja and @mathemancer seem to have insight into that logic.
| 1 | Improve readability of column moving logic - ## Problem
Our column moving logic is difficult to get into. There's a related bug I've encountered and am of the opinion that a focused effort to improve column moving logic's readability is warranted.
## Proposed solution
@silentninja proposed that we do this together on a call.
## Additional context
@silentninja and @mathemancer seem to have insight into that logic.
| architecture | improve readability of column moving logic problem our column moving logic is difficult to get into there s a related bug i ve encountered and am of the opinion that a focused effort to improve column moving logic s readability is warranted proposed solution silentninja proposed that we do this together on a call additional context silentninja and mathemancer seem to have insight into that logic |
53 | 24,443,675,959 | IssuesEvent | 2022-10-06 16:11:34 | centerofci/mathesar | https://api.github.com/repos/centerofci/mathesar | closed | Optimize MetaData use | type: enhancement affects: architecture affects: technical debt work: backend work: database restricted: maintainers status: started | ### Observation
- Metadata reflection by SqlAlchemy results in 10 query each time a new metadata object is created/reflected,
- and, most of the methods don't reuse metadata,
- currently loading the table page requires ~1600 queries to the database;
- `mathesar.reflection.reflect_db_objects`:
- reflecting each database object in our model (tables, schemas, databases, columns, constraints) adds 12 query for each object as it uses a new MetaData object instead of reusing the existing MetaData;
- `db.columns.operations.select.get_columns_name_from_attnums`:
- fetching column name is needed for accessing the SA column object and each call to fetch the column name results in a metadata reflection call.
### Cause
- `db` module methods not reusing metadata.
### Solution
- Use a single `MetaData` instance;
- Possibly use a multi-request `MetaData` cache;
- Possibilities
- Use session-level cache;
- Django cache;
- Use multi-session cache;
- Maybe file or db to store pickled metadata object;
- How to keep cache's validity up to date?
- Maybe invalidate it when doing mutating operations. | 1 | Optimize MetaData use - ### Observation
- Metadata reflection by SqlAlchemy results in 10 query each time a new metadata object is created/reflected,
- and, most of the methods don't reuse metadata,
- currently loading the table page requires ~1600 queries to the database;
- `mathesar.reflection.reflect_db_objects`:
- reflecting each database object in our model (tables, schemas, databases, columns, constraints) adds 12 query for each object as it uses a new MetaData object instead of reusing the existing MetaData;
- `db.columns.operations.select.get_columns_name_from_attnums`:
- fetching column name is needed for accessing the SA column object and each call to fetch the column name results in a metadata reflection call.
### Cause
- `db` module methods not reusing metadata.
### Solution
- Use a single `MetaData` instance;
- Possibly use a multi-request `MetaData` cache;
- Possibilities
- Use session-level cache;
- Django cache;
- Use multi-session cache;
- Maybe file or db to store pickled metadata object;
- How to keep cache's validity up to date?
- Maybe invalidate it when doing mutating operations. | architecture | optimize metadata use observation metadata reflection by sqlalchemy results in query each time a new metadata object is created reflected and most of the methods don t reuse metadata currently loading the table page requires queries to the database mathesar reflection reflect db objects reflecting each database object in our model tables schemas databases columns constraints adds query for each object as it uses a new metadata object instead of reusing the existing metadata db columns operations select get columns name from attnums fetching column name is needed for accessing the sa column object and each call to fetch the column name results in a metadata reflection call cause db module methods not reusing metadata solution use a single metadata instance possibly use a multi request metadata cache possibilities use session level cache django cache use multi session cache maybe file or db to store pickled metadata object how to keep cache s validity up to date maybe invalidate it when doing mutating operations |
54 | 24,464,940,430 | IssuesEvent | 2022-10-07 14:17:01 | woocommerce/woocommerce-ios | https://api.github.com/repos/woocommerce/woocommerce-ios | closed | Make experimental features easier to add and remove | category: architecture type: technical debt | As mentioned in #7011, adding or removing a toggle for experimental features takes a [bigger code change than one might expect](https://github.com/woocommerce/woocommerce-ios/pull/6954). Ideally we'd want these changes to be leaner, just like [adding or removing a feature flag](https://github.com/woocommerce/woocommerce-ios/pull/6995).
Part of this would be solved by #7011, but we can also improve the Experimental Features screen, making it easier and more automatic to add new flags there.
| 1 | Make experimental features easier to add and remove - As mentioned in #7011, adding or removing a toggle for experimental features takes a [bigger code change than one might expect](https://github.com/woocommerce/woocommerce-ios/pull/6954). Ideally we'd want these changes to be leaner, just like [adding or removing a feature flag](https://github.com/woocommerce/woocommerce-ios/pull/6995).
Part of this would be solved by #7011, but we can also improve the Experimental Features screen, making it easier and more automatic to add new flags there.
| architecture | make experimental features easier to add and remove as mentioned in adding or removing a toggle for experimental features takes a ideally we d want these changes to be leaner just like part of this would be solved by but we can also improve the experimental features screen making it easier and more automatic to add new flags there |
55 | 25,474,118,939 | IssuesEvent | 2022-11-25 12:54:23 | centerofci/mathesar | https://api.github.com/repos/centerofci/mathesar | opened | Fix filter/hint system so that it doesn't have to be stubbed | type: enhancement affects: dx affects: architecture affects: technical debt work: backend status: draft | We're currently hardcoding on the frontend what filters are available and/or how they can be used. We designed the hint system for that, but there was some problem with that, so we ended up hardcoding, and postponed fixing the actual problem.
I don't recall what the actual problem was. @pavish do you have insights? | 1 | Fix filter/hint system so that it doesn't have to be stubbed - We're currently hardcoding on the frontend what filters are available and/or how they can be used. We designed the hint system for that, but there was some problem with that, so we ended up hardcoding, and postponed fixing the actual problem.
I don't recall what the actual problem was. @pavish do you have insights? | architecture | fix filter hint system so that it doesn t have to be stubbed we re currently hardcoding on the frontend what filters are available and or how they can be used we designed the hint system for that but there was some problem with that so we ended up hardcoding and postponed fixing the actual problem i don t recall what the actual problem was pavish do you have insights |
56 | 25,481,399,070 | IssuesEvent | 2022-11-25 21:47:10 | contribute-design/contribute.design | https://api.github.com/repos/contribute-design/contribute.design | opened | Refactor CloudFlare workers | 🏗 architecture 🏗 debt | **Why should it be implemented?**
Our current worker implementation is extremely immature... We're basically commit the workers to the monorepo but deploy them manually to cloudflare
**Describe the solution**
- Ideally we'd use lerna or something similar to manage each worker as an own package
- Each worker would be written in typescript
- Workers should be able to share code between each other
- Building and deployment should happen automatically via a GH action or similar
**Additional context**
There's a nice boilerplate over here:
https://github.com/cmackenzie1/holster
| 1 | Refactor CloudFlare workers - **Why should it be implemented?**
Our current worker implementation is extremely immature... We're basically commit the workers to the monorepo but deploy them manually to cloudflare
**Describe the solution**
- Ideally we'd use lerna or something similar to manage each worker as an own package
- Each worker would be written in typescript
- Workers should be able to share code between each other
- Building and deployment should happen automatically via a GH action or similar
**Additional context**
There's a nice boilerplate over here:
https://github.com/cmackenzie1/holster
| architecture | refactor cloudflare workers why should it be implemented our current worker implementation is extremely immature we re basically commit the workers to the monorepo but deploy them manually to cloudflare describe the solution ideally we d use lerna or something similar to manage each worker as an own package each worker would be written in typescript workers should be able to share code between each other building and deployment should happen automatically via a gh action or similar additional context there s a nice boilerplate over here |
59 | 26,865,654,971 | IssuesEvent | 2023-02-03 23:17:45 | FuelLabs/fuel-core | https://api.github.com/repos/FuelLabs/fuel-core | opened | Test suite is too slow | tech-debt fuel-core architecture | Our full test sweet is taking too long to run. Especially the p2p tests. This slows down our development cycle and leads to avoiding running the entire suite. | 1 | Test suite is too slow - Our full test sweet is taking too long to run. Especially the p2p tests. This slows down our development cycle and leads to avoiding running the entire suite. | architecture | test suite is too slow our full test sweet is taking too long to run especially the tests this slows down our development cycle and leads to avoiding running the entire suite |
60 | 30,629,687,218 | IssuesEvent | 2023-07-24 13:49:23 | woocommerce/woocommerce-ios | https://api.github.com/repos/woocommerce/woocommerce-ios | opened | Make WebKit User Agent run on main thread explicitely | type: enhancement category: architecture type: technical debt | Opening this issue to keep an eye on this potential problem and attempt improvements in the future:
At the moment [our implementation webkit user agent](https://github.com/woocommerce/woocommerce-ios/blob/7c6ecfeefb7c3af6bd6fa328efe3b0aa787e1d6f/Networking/Networking/Settings/UserAgent.swift#L15) (in order to setup, for example, authenticated wpcom requests) relies on the action dispatcher to assure that any request will run on the main thread, however, this is not enforced unless we actually call the remote that create and enqueues these requests through an action.
This means that while works as a side-effect (we happen to call these remotes via actions in the dispatcher) we risk a runtime crash when any remote that involves a webkit user agent instantiation is not ran through the dispatcher, since there's no compiler-check that prevents us from doing so.
One potential solution could be to mark the method as `@MainActor `, but this propagates an error through networking since we would have a main-actor static property attempted to be used across multiple non-main-actor contexts. Another option could be to wrap it in a call to DispatchQueue.main.async.
Ref: p1690199791835769-slack-C03L1NF1EA3 | 1 | Make WebKit User Agent run on main thread explicitely - Opening this issue to keep an eye on this potential problem and attempt improvements in the future:
At the moment [our implementation webkit user agent](https://github.com/woocommerce/woocommerce-ios/blob/7c6ecfeefb7c3af6bd6fa328efe3b0aa787e1d6f/Networking/Networking/Settings/UserAgent.swift#L15) (in order to setup, for example, authenticated wpcom requests) relies on the action dispatcher to assure that any request will run on the main thread, however, this is not enforced unless we actually call the remote that create and enqueues these requests through an action.
This means that while works as a side-effect (we happen to call these remotes via actions in the dispatcher) we risk a runtime crash when any remote that involves a webkit user agent instantiation is not ran through the dispatcher, since there's no compiler-check that prevents us from doing so.
One potential solution could be to mark the method as `@MainActor `, but this propagates an error through networking since we would have a main-actor static property attempted to be used across multiple non-main-actor contexts. Another option could be to wrap it in a call to DispatchQueue.main.async.
Ref: p1690199791835769-slack-C03L1NF1EA3 | architecture | make webkit user agent run on main thread explicitely opening this issue to keep an eye on this potential problem and attempt improvements in the future at the moment in order to setup for example authenticated wpcom requests relies on the action dispatcher to assure that any request will run on the main thread however this is not enforced unless we actually call the remote that create and enqueues these requests through an action this means that while works as a side effect we happen to call these remotes via actions in the dispatcher we risk a runtime crash when any remote that involves a webkit user agent instantiation is not ran through the dispatcher since there s no compiler check that prevents us from doing so one potential solution could be to mark the method as mainactor but this propagates an error through networking since we would have a main actor static property attempted to be used across multiple non main actor contexts another option could be to wrap it in a call to dispatchqueue main async ref slack |
61 | 30,926,869,098 | IssuesEvent | 2023-08-06 15:37:16 | spacemeshos/go-spacemesh | https://api.github.com/repos/spacemeshos/go-spacemesh | closed | Manage running goroutines and graceful shutdown | technical debt devex concurrency architecture | ## Motivation
Graceful shutdown is proving to be a challenge when multiple goroutines are running in the background. This has two parts:
1. Signaling a shutdown and having each service respond by terminating gracefully (this part is working, but each service has its own, slightly different, implementation).
2. being able to know when all services have completed.
This is important in production, where we want to ensure no data is lost due to a dirty shutdown, and in tests where we want to terminate quickly and cleanly without having to add unnecessary waiting periods.
## Method
We want to integrate [Tomb](https://pkg.go.dev/gopkg.in/tomb.v2?tab=doc), a package for handling clean goroutine tracking and termination.
Tomb provides a `Go()` method that's intended to replace calling the `go` keyword directly to start goroutines. Internally, it uses a waitgroup to track how many goroutines have been started and how many have completed. It can also provide the first error that triggered a shutdown, if it was due to an error.
Tomb should be integrated in a single module first and then additional PRs can integrate it into more modules. Eventually:
- No goroutine should be started using a bare `go` keyword.
- All method calls that can accept a `Context` should receive one, provided by Tomb.
- All `select` statements should have an early termination clause using `tomb.Dying()` (a method returning a channel that's closed when the Tomb is killed). If this clause is invoked, an `ErrDying` (a Tomb constant) should be returned by the goroutine and Tomb knows to ignore it as a kill reason.
- All received context objects (specifically in api handlers) should be wrapped using `tomb.Context(ctx)` if used (I think they're never used as of writing this).
- New contexts should never be generated from scratch and we should always use the Tomb for context (this happens mostly in the P2P module, but also in the PoET client).
## WIP
- [ ] Add specific tasks (where to initialize the Tomb, how to do shutdown)
- [ ] List some candidate modules for the first integration | 1 | Manage running goroutines and graceful shutdown - ## Motivation
Graceful shutdown is proving to be a challenge when multiple goroutines are running in the background. This has two parts:
1. Signaling a shutdown and having each service respond by terminating gracefully (this part is working, but each service has its own, slightly different, implementation).
2. being able to know when all services have completed.
This is important in production, where we want to ensure no data is lost due to a dirty shutdown, and in tests where we want to terminate quickly and cleanly without having to add unnecessary waiting periods.
## Method
We want to integrate [Tomb](https://pkg.go.dev/gopkg.in/tomb.v2?tab=doc), a package for handling clean goroutine tracking and termination.
Tomb provides a `Go()` method that's intended to replace calling the `go` keyword directly to start goroutines. Internally, it uses a waitgroup to track how many goroutines have been started and how many have completed. It can also provide the first error that triggered a shutdown, if it was due to an error.
Tomb should be integrated in a single module first and then additional PRs can integrate it into more modules. Eventually:
- No goroutine should be started using a bare `go` keyword.
- All method calls that can accept a `Context` should receive one, provided by Tomb.
- All `select` statements should have an early termination clause using `tomb.Dying()` (a method returning a channel that's closed when the Tomb is killed). If this clause is invoked, an `ErrDying` (a Tomb constant) should be returned by the goroutine and Tomb knows to ignore it as a kill reason.
- All received context objects (specifically in api handlers) should be wrapped using `tomb.Context(ctx)` if used (I think they're never used as of writing this).
- New contexts should never be generated from scratch and we should always use the Tomb for context (this happens mostly in the P2P module, but also in the PoET client).
## WIP
- [ ] Add specific tasks (where to initialize the Tomb, how to do shutdown)
- [ ] List some candidate modules for the first integration | architecture | manage running goroutines and graceful shutdown motivation graceful shutdown is proving to be a challenge when multiple goroutines are running in the background this has two parts signaling a shutdown and having each service respond by terminating gracefully this part is working but each service has its own slightly different implementation being able to know when all services have completed this is important in production where we want to ensure no data is lost due to a dirty shutdown and in tests where we want to terminate quickly and cleanly without having to add unnecessary waiting periods method we want to integrate a package for handling clean goroutine tracking and termination tomb provides a go method that s intended to replace calling the go keyword directly to start goroutines internally it uses a waitgroup to track how many goroutines have been started and how many have completed it can also provide the first error that triggered a shutdown if it was due to an error tomb should be integrated in a single module first and then additional prs can integrate it into more modules eventually no goroutine should be started using a bare go keyword all method calls that can accept a context should receive one provided by tomb all select statements should have an early termination clause using tomb dying a method returning a channel that s closed when the tomb is killed if this clause is invoked an errdying a tomb constant should be returned by the goroutine and tomb knows to ignore it as a kill reason all received context objects specifically in api handlers should be wrapped using tomb context ctx if used i think they re never used as of writing this new contexts should never be generated from scratch and we should always use the tomb for context this happens mostly in the module but also in the poet client wip add specific tasks where to initialize the tomb how to do shutdown list some candidate modules for the first integration |
62 | 31,168,179,526 | IssuesEvent | 2023-08-16 21:39:17 | centerofci/mathesar | https://api.github.com/repos/centerofci/mathesar | closed | `PATCH` requests to the Table API should support changing the table's name and columns at the same time | type: enhancement affects: architecture affects: technical debt work: backend work: database status: blocked | ## Problem
<!-- Please provide a clear and concise description of the problem that this feature request is designed to solve.-->
The current implementation of the table API does not allow you to update both the `name` and `columns` at the same time. This is counterintuitive and it would be ideal if you could update both at once.
## Proposed solution
<!-- A clear and concise description of your proposed solution or feature. -->
We need to figure out how to update the name first, use the updated table name in the column-related changes and roll the whole thing back if any of the operations fail (including the name change).
## Additional context
<!-- Add any other context or screenshots about the feature request here.-->
- See conversation on #562
- We should do #592 first since this involves more single-transaction operations. Marking this issue as blocked by it.
- [Postgres wiki page on transactional DDL](https://wiki.postgresql.org/wiki/Transactional_DDL_in_PostgreSQL:_A_Competitive_Analysis) | 1 | `PATCH` requests to the Table API should support changing the table's name and columns at the same time - ## Problem
<!-- Please provide a clear and concise description of the problem that this feature request is designed to solve.-->
The current implementation of the table API does not allow you to update both the `name` and `columns` at the same time. This is counterintuitive and it would be ideal if you could update both at once.
## Proposed solution
<!-- A clear and concise description of your proposed solution or feature. -->
We need to figure out how to update the name first, use the updated table name in the column-related changes and roll the whole thing back if any of the operations fail (including the name change).
## Additional context
<!-- Add any other context or screenshots about the feature request here.-->
- See conversation on #562
- We should do #592 first since this involves more single-transaction operations. Marking this issue as blocked by it.
- [Postgres wiki page on transactional DDL](https://wiki.postgresql.org/wiki/Transactional_DDL_in_PostgreSQL:_A_Competitive_Analysis) | architecture | patch requests to the table api should support changing the table s name and columns at the same time problem the current implementation of the table api does not allow you to update both the name and columns at the same time this is counterintuitive and it would be ideal if you could update both at once proposed solution we need to figure out how to update the name first use the updated table name in the column related changes and roll the whole thing back if any of the operations fail including the name change additional context see conversation on we should do first since this involves more single transaction operations marking this issue as blocked by it |
0 | 2,490,609,947 | IssuesEvent | 2015-01-02 17:29:17 | FineUploader/fine-uploader | https://api.github.com/repos/FineUploader/fine-uploader | closed | 5 - Get build script under control | 5 - Done build technical debt | Our grunt script is out of control. It's currently just a mess of copy and pasted code with countless instances of code repetition. Furthermore, the majority of the logic resides in one single coffeescript file that is approaching 1000 lines. We'll have to do something(s) to rein in this mess.
<!---
@huboard:{"order":1145.0,"custom_state":""}
-->
| 1 | 5 - Get build script under control - Our grunt script is out of control. It's currently just a mess of copy and pasted code with countless instances of code repetition. Furthermore, the majority of the logic resides in one single coffeescript file that is approaching 1000 lines. We'll have to do something(s) to rein in this mess.
<!---
@huboard:{"order":1145.0,"custom_state":""}
-->
| build | get build script under control our grunt script is out of control it s currently just a mess of copy and pasted code with countless instances of code repetition furthermore the majority of the logic resides in one single coffeescript file that is approaching lines we ll have to do something s to rein in this mess huboard order custom state |
1 | 3,202,827,471 | IssuesEvent | 2015-10-02 15:50:48 | openshift/origin | https://api.github.com/repos/openshift/origin | closed | Unify build strategy fields | area/techdebt area/usability component/build priority/P3 | Currently we're starting to copy&paste fields between different build strategies. So far there's almost two of them that are shared between strategies: `Image` and `Env` (although the latter isn't present in `DockerBuildStrategy`). Additionally when @mfojtik PR https://github.com/openshift/origin/pull/1411 lands there'll be `DockerRegistrySecretRef`. We should move those fields into `BuildStrategy`.
| 1 | Unify build strategy fields - Currently we're starting to copy&paste fields between different build strategies. So far there's almost two of them that are shared between strategies: `Image` and `Env` (although the latter isn't present in `DockerBuildStrategy`). Additionally when @mfojtik PR https://github.com/openshift/origin/pull/1411 lands there'll be `DockerRegistrySecretRef`. We should move those fields into `BuildStrategy`.
| build | unify build strategy fields currently we re starting to copy paste fields between different build strategies so far there s almost two of them that are shared between strategies image and env although the latter isn t present in dockerbuildstrategy additionally when mfojtik pr lands there ll be dockerregistrysecretref we should move those fields into buildstrategy |
3 | 3,642,151,696 | IssuesEvent | 2016-02-14 04:26:15 | cortoproject/corto | https://api.github.com/repos/cortoproject/corto | opened | Revise and simplify buildsystem | Corto:BuildSystem Corto:PackageManagement Corto:TechnicalDebt Corto:Usability | With #450 implemented, the buildsystem merged components with packages, and generators with libraries. To further simplify the buildsystem, libraries can be merged with packages as well. Packages are more expressive, allow for better organization of libraries and prevent nameclashes.
Generators are currently stored in the `lib/corto/<version>/libraries` folder. Generators will become packages as well, and shall be stored in `corto/gen/<binding>/<name>`. For example: `corto/gen/c/api`, or `corto/gen/doc/html`.
Since these packages have no definition file, the current implementation of the buildsystem requires a user to specify `NOCORTO`. However, that disables linking with Corto, and also disables automatic management of dependencies & include files- something that was previously done by components.
This existing functionality must be consolidated with the new design. A proposal:
Create a package with a definition file `foo.cx`. For each interface in the definition file a managed implementation file will be generated, along with header files. Include dependencies are managed.
```
corto create package Foo
```
Create a package with automatic dependency (include file) management. Minimal code generation will be required. The package will have a managed `include/Foo.h` header file. The package will link with Corto. The buildsystem will detect that no definition file is available, and will therefore switch to limited code generation.
```
corto create package Foo --empty
```
Create a package for which no code is generated, and which does not link with Corto. This is useful for wrapping 3rd party libraries or for projects where only the Corto buildsystem is required. The generated rakefile will contain the line `NOCORTO = true` which signals the build system to not generate any code.
```
corto create package Foo --nocorto
```
Additionally, a `--local` flag can be provided which will ensure that the package is not installed to an environment (either local or global). This is for example useful for test suites.
| 1 | Revise and simplify buildsystem - With #450 implemented, the buildsystem merged components with packages, and generators with libraries. To further simplify the buildsystem, libraries can be merged with packages as well. Packages are more expressive, allow for better organization of libraries and prevent nameclashes.
Generators are currently stored in the `lib/corto/<version>/libraries` folder. Generators will become packages as well, and shall be stored in `corto/gen/<binding>/<name>`. For example: `corto/gen/c/api`, or `corto/gen/doc/html`.
Since these packages have no definition file, the current implementation of the buildsystem requires a user to specify `NOCORTO`. However, that disables linking with Corto, and also disables automatic management of dependencies & include files- something that was previously done by components.
This existing functionality must be consolidated with the new design. A proposal:
Create a package with a definition file `foo.cx`. For each interface in the definition file a managed implementation file will be generated, along with header files. Include dependencies are managed.
```
corto create package Foo
```
Create a package with automatic dependency (include file) management. Minimal code generation will be required. The package will have a managed `include/Foo.h` header file. The package will link with Corto. The buildsystem will detect that no definition file is available, and will therefore switch to limited code generation.
```
corto create package Foo --empty
```
Create a package for which no code is generated, and which does not link with Corto. This is useful for wrapping 3rd party libraries or for projects where only the Corto buildsystem is required. The generated rakefile will contain the line `NOCORTO = true` which signals the build system to not generate any code.
```
corto create package Foo --nocorto
```
Additionally, a `--local` flag can be provided which will ensure that the package is not installed to an environment (either local or global). This is for example useful for test suites.
| build | revise and simplify buildsystem with implemented the buildsystem merged components with packages and generators with libraries to further simplify the buildsystem libraries can be merged with packages as well packages are more expressive allow for better organization of libraries and prevent nameclashes generators are currently stored in the lib corto libraries folder generators will become packages as well and shall be stored in corto gen for example corto gen c api or corto gen doc html since these packages have no definition file the current implementation of the buildsystem requires a user to specify nocorto however that disables linking with corto and also disables automatic management of dependencies include files something that was previously done by components this existing functionality must be consolidated with the new design a proposal create a package with a definition file foo cx for each interface in the definition file a managed implementation file will be generated along with header files include dependencies are managed corto create package foo create a package with automatic dependency include file management minimal code generation will be required the package will have a managed include foo h header file the package will link with corto the buildsystem will detect that no definition file is available and will therefore switch to limited code generation corto create package foo empty create a package for which no code is generated and which does not link with corto this is useful for wrapping party libraries or for projects where only the corto buildsystem is required the generated rakefile will contain the line nocorto true which signals the build system to not generate any code corto create package foo nocorto additionally a local flag can be provided which will ensure that the package is not installed to an environment either local or global this is for example useful for test suites |
4 | 3,799,420,730 | IssuesEvent | 2016-03-23 15:51:36 | mesosphere/marathon | https://api.github.com/repos/mesosphere/marathon | opened | Resident Tasks: Flaky test: persistent volume will be re-attached and keep state | build debt Epic-217 | ```
[14:24:16][Step 3/3] - persistent volume will be re-attached and keep state *** FAILED *** (30 seconds, 378 milliseconds)
[14:24:16][Step 3/3] java.lang.AssertionError: Waiting for event deployment_success to arrive took longer than 30 seconds. Give up.
[14:24:16][Step 3/3] at mesosphere.marathon.integration.setup.WaitTestSupport$.next$1(WaitTestSupport.scala:30)
[14:24:16][Step 3/3] at mesosphere.marathon.integration.setup.WaitTestSupport$.waitFor(WaitTestSupport.scala:36)
[14:24:16][Step 3/3] at mesosphere.marathon.integration.setup.MarathonCallbackTestSupport$class.waitForEventMatching(MarathonCallbackTestSupport.scala:48)
[14:24:16][Step 3/3] at mesosphere.marathon.integration.ResidentTaskIntegrationTest.waitForEventMatching(ResidentTaskIntegrationTest.scala:16)
[14:24:16][Step 3/3] at mesosphere.marathon.integration.setup.MarathonCallbackTestSupport$class.waitForEventWith(MarathonCallbackTestSupport.scala:52)
[14:24:16][Step 3/3] at mesosphere.marathon.integration.ResidentTaskIntegrationTest.waitForEventWith(ResidentTaskIntegrationTest.scala:16)
[14:24:16][Step 3/3] at mesosphere.marathon.integration.setup.MarathonCallbackTestSupport$class.waitForEvent(MarathonCallbackTestSupport.scala:32)
[14:24:16][Step 3/3] at mesosphere.marathon.integration.ResidentTaskIntegrationTest.waitForEvent(ResidentTaskIntegrationTest.scala:16)
[14:24:16][Step 3/3] at mesosphere.marathon.integration.ResidentTaskIntegrationTest$$anonfun$3.apply(ResidentTaskIntegrationTest.scala:56)
[14:24:16][Step 3/3] at mesosphere.marathon.integration.ResidentTaskIntegrationTest$$anonfun$3.apply(ResidentTaskIntegrationTest.scala:44)
[14:24:16][Step 3/3] at mesosphere.marathon.integration.ResidentTaskIntegrationTest$$anonfun$test$1.apply$mcV$sp(ResidentTaskIntegrationTest.scala:225)
[14:24:16][Step 3/3] at mesosphere.marathon.integration.ResidentTaskIntegrationTest$$anonfun$test$1.apply(ResidentTaskIntegrationTest.scala:225)
[14:24:16][Step 3/3] at mesosphere.marathon.integration.ResidentTaskIntegrationTest$$anonfun$test$1.apply(ResidentTaskIntegrationTest.scala:225)
[14:24:16][Step 3/3] at org.scalatest.Transformer$$anonfun$apply$1.apply$mcV$sp(Transformer.scala:22)
[14:24:16][Step 3/3] at org.scalatest.OutcomeOf$class.outcomeOf(OutcomeOf.scala:85)
[14:24:16][Step 3/3] at org.scalatest.OutcomeOf$.outcomeOf(OutcomeOf.scala:104)
[14:24:16][Step 3/3] at org.scalatest.Transformer.apply(Transformer.scala:22)
[14:24:16][Step 3/3] at org.scalatest.Transformer.apply(Transformer.scala:20)
[14:24:16][Step 3/3] at org.scalatest.FunSuiteLike$$anon$1.apply(FunSuiteLike.scala:158)
[14:24:16][Step 3/3] at org.scalatest.Suite$class.withFixture(Suite.scala:1121)
[14:24:16][Step 3/3] at org.scalatest.FunSuite.withFixture(FunSuite.scala:1559)
[14:24:16][Step 3/3] at org.scalatest.FunSuiteLike$class.invokeWithFixture$1(FunSuiteLike.scala:155)
[14:24:16][Step 3/3] at org.scalatest.FunSuiteLike$$anonfun$runTest$1.apply(FunSuiteLike.scala:167)
[14:24:16][Step 3/3] at org.scalatest.FunSuiteLike$$anonfun$runTest$1.apply(FunSuiteLike.scala:167)
[14:24:16][Step 3/3] at org.scalatest.SuperEngine.runTestImpl(Engine.scala:306)
[14:24:16][Step 3/3] at org.scalatest.FunSuiteLike$class.runTest(FunSuiteLike.scala:167)
[14:24:16][Step 3/3] at mesosphere.marathon.integration.ResidentTaskIntegrationTest.org$scalatest$BeforeAndAfter$$super$runTest(ResidentTaskIntegrationTest.scala:16)
[14:24:16][Step 3/3] at org.scalatest.BeforeAndAfter$class.runTest(BeforeAndAfter.scala:200)
[14:24:16][Step 3/3] at mesosphere.marathon.integration.ResidentTaskIntegrationTest.runTest(ResidentTaskIntegrationTest.scala:16)
[14:24:16][Step 3/3] at org.scalatest.FunSuiteLike$$anonfun$runTests$1.apply(FunSuiteLike.scala:200)
[14:24:16][Step 3/3] at org.scalatest.FunSuiteLike$$anonfun$runTests$1.apply(FunSuiteLike.scala:200)
[14:24:16][Step 3/3] at org.scalatest.SuperEngine$$anonfun$traverseSubNodes$1$1.apply(Engine.scala:413)
[14:24:16][Step 3/3] at org.scalatest.SuperEngine$$anonfun$traverseSubNodes$1$1.apply(Engine.scala:401)
[14:24:16][Step 3/3] at scala.collection.immutable.List.foreach(List.scala:381)
[14:24:16][Step 3/3] at org.scalatest.SuperEngine.traverseSubNodes$1(Engine.scala:401)
[14:24:16][Step 3/3] at org.scalatest.SuperEngine.org$scalatest$SuperEngine$$runTestsInBranch(Engine.scala:396)
[14:24:16][Step 3/3] at org.scalatest.SuperEngine.runTestsImpl(Engine.scala:483)
[14:24:16][Step 3/3] at org.scalatest.FunSuiteLike$class.runTests(FunSuiteLike.scala:200)
[14:24:16][Step 3/3] at org.scalatest.FunSuite.runTests(FunSuite.scala:1559)
[14:24:16][Step 3/3] at org.scalatest.Suite$class.run(Suite.scala:1423)
[14:24:16][Step 3/3] at org.scalatest.FunSuite.org$scalatest$FunSuiteLike$$super$run(FunSuite.scala:1559)
[14:24:16][Step 3/3] at org.scalatest.FunSuiteLike$$anonfun$run$1.apply(FunSuiteLike.scala:204)
[14:24:16][Step 3/3] at org.scalatest.FunSuiteLike$$anonfun$run$1.apply(FunSuiteLike.scala:204)
[14:24:16][Step 3/3] at org.scalatest.SuperEngine.runImpl(Engine.scala:545)
[14:24:16][Step 3/3] at org.scalatest.FunSuiteLike$class.run(FunSuiteLike.scala:204)
[14:24:16][Step 3/3] at mesosphere.marathon.integration.ResidentTaskIntegrationTest.org$scalatest$BeforeAndAfterAllConfigMap$$super$run(ResidentTaskIntegrationTest.scala:16)
[14:24:16][Step 3/3] at org.scalatest.BeforeAndAfterAllConfigMap$class.liftedTree1$1(BeforeAndAfterAllConfigMap.scala:248)
[14:24:16][Step 3/3] at org.scalatest.BeforeAndAfterAllConfigMap$class.run(BeforeAndAfterAllConfigMap.scala:247)
[14:24:16][Step 3/3] at mesosphere.marathon.integration.ResidentTaskIntegrationTest.org$scalatest$BeforeAndAfter$$super$run(ResidentTaskIntegrationTest.scala:16)
[14:24:16][Step 3/3] at org.scalatest.BeforeAndAfter$class.run(BeforeAndAfter.scala:241)
[14:24:16][Step 3/3] at mesosphere.marathon.integration.ResidentTaskIntegrationTest.run(ResidentTaskIntegrationTest.scala:16)
[14:24:16][Step 3/3] at org.scalatest.tools.Framework.org$scalatest$tools$Framework$$runSuite(Framework.scala:444)
[14:24:16][Step 3/3] at org.scalatest.tools.Framework$ScalaTestTask.execute(Framework.scala:651)
[14:24:16][Step 3/3] at sbt.ForkMain$Run$2.call(ForkMain.java:294)
[14:24:16][Step 3/3] at sbt.ForkMain$Run$2.call(ForkMain.java:284)
[14:24:16][Step 3/3] at java.util.concurrent.FutureTask.run(FutureTask.java:266)
[14:24:16][Step 3/3] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
[14:24:16][Step 3/3] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
[14:24:16][Step 3/3] at java.lang.Thread.run(Thread.java:745)
[14:24:16][Step 3/3] + Given An app that writes into a persistent volume
[14:24:16][Step 3/3] + When a task is launched
[14:24:16][Step 3/3] + Then it successfully writes to the persistent volume and then finishes
``` | 1 | Resident Tasks: Flaky test: persistent volume will be re-attached and keep state - ```
[14:24:16][Step 3/3] - persistent volume will be re-attached and keep state *** FAILED *** (30 seconds, 378 milliseconds)
[14:24:16][Step 3/3] java.lang.AssertionError: Waiting for event deployment_success to arrive took longer than 30 seconds. Give up.
[14:24:16][Step 3/3] at mesosphere.marathon.integration.setup.WaitTestSupport$.next$1(WaitTestSupport.scala:30)
[14:24:16][Step 3/3] at mesosphere.marathon.integration.setup.WaitTestSupport$.waitFor(WaitTestSupport.scala:36)
[14:24:16][Step 3/3] at mesosphere.marathon.integration.setup.MarathonCallbackTestSupport$class.waitForEventMatching(MarathonCallbackTestSupport.scala:48)
[14:24:16][Step 3/3] at mesosphere.marathon.integration.ResidentTaskIntegrationTest.waitForEventMatching(ResidentTaskIntegrationTest.scala:16)
[14:24:16][Step 3/3] at mesosphere.marathon.integration.setup.MarathonCallbackTestSupport$class.waitForEventWith(MarathonCallbackTestSupport.scala:52)
[14:24:16][Step 3/3] at mesosphere.marathon.integration.ResidentTaskIntegrationTest.waitForEventWith(ResidentTaskIntegrationTest.scala:16)
[14:24:16][Step 3/3] at mesosphere.marathon.integration.setup.MarathonCallbackTestSupport$class.waitForEvent(MarathonCallbackTestSupport.scala:32)
[14:24:16][Step 3/3] at mesosphere.marathon.integration.ResidentTaskIntegrationTest.waitForEvent(ResidentTaskIntegrationTest.scala:16)
[14:24:16][Step 3/3] at mesosphere.marathon.integration.ResidentTaskIntegrationTest$$anonfun$3.apply(ResidentTaskIntegrationTest.scala:56)
[14:24:16][Step 3/3] at mesosphere.marathon.integration.ResidentTaskIntegrationTest$$anonfun$3.apply(ResidentTaskIntegrationTest.scala:44)
[14:24:16][Step 3/3] at mesosphere.marathon.integration.ResidentTaskIntegrationTest$$anonfun$test$1.apply$mcV$sp(ResidentTaskIntegrationTest.scala:225)
[14:24:16][Step 3/3] at mesosphere.marathon.integration.ResidentTaskIntegrationTest$$anonfun$test$1.apply(ResidentTaskIntegrationTest.scala:225)
[14:24:16][Step 3/3] at mesosphere.marathon.integration.ResidentTaskIntegrationTest$$anonfun$test$1.apply(ResidentTaskIntegrationTest.scala:225)
[14:24:16][Step 3/3] at org.scalatest.Transformer$$anonfun$apply$1.apply$mcV$sp(Transformer.scala:22)
[14:24:16][Step 3/3] at org.scalatest.OutcomeOf$class.outcomeOf(OutcomeOf.scala:85)
[14:24:16][Step 3/3] at org.scalatest.OutcomeOf$.outcomeOf(OutcomeOf.scala:104)
[14:24:16][Step 3/3] at org.scalatest.Transformer.apply(Transformer.scala:22)
[14:24:16][Step 3/3] at org.scalatest.Transformer.apply(Transformer.scala:20)
[14:24:16][Step 3/3] at org.scalatest.FunSuiteLike$$anon$1.apply(FunSuiteLike.scala:158)
[14:24:16][Step 3/3] at org.scalatest.Suite$class.withFixture(Suite.scala:1121)
[14:24:16][Step 3/3] at org.scalatest.FunSuite.withFixture(FunSuite.scala:1559)
[14:24:16][Step 3/3] at org.scalatest.FunSuiteLike$class.invokeWithFixture$1(FunSuiteLike.scala:155)
[14:24:16][Step 3/3] at org.scalatest.FunSuiteLike$$anonfun$runTest$1.apply(FunSuiteLike.scala:167)
[14:24:16][Step 3/3] at org.scalatest.FunSuiteLike$$anonfun$runTest$1.apply(FunSuiteLike.scala:167)
[14:24:16][Step 3/3] at org.scalatest.SuperEngine.runTestImpl(Engine.scala:306)
[14:24:16][Step 3/3] at org.scalatest.FunSuiteLike$class.runTest(FunSuiteLike.scala:167)
[14:24:16][Step 3/3] at mesosphere.marathon.integration.ResidentTaskIntegrationTest.org$scalatest$BeforeAndAfter$$super$runTest(ResidentTaskIntegrationTest.scala:16)
[14:24:16][Step 3/3] at org.scalatest.BeforeAndAfter$class.runTest(BeforeAndAfter.scala:200)
[14:24:16][Step 3/3] at mesosphere.marathon.integration.ResidentTaskIntegrationTest.runTest(ResidentTaskIntegrationTest.scala:16)
[14:24:16][Step 3/3] at org.scalatest.FunSuiteLike$$anonfun$runTests$1.apply(FunSuiteLike.scala:200)
[14:24:16][Step 3/3] at org.scalatest.FunSuiteLike$$anonfun$runTests$1.apply(FunSuiteLike.scala:200)
[14:24:16][Step 3/3] at org.scalatest.SuperEngine$$anonfun$traverseSubNodes$1$1.apply(Engine.scala:413)
[14:24:16][Step 3/3] at org.scalatest.SuperEngine$$anonfun$traverseSubNodes$1$1.apply(Engine.scala:401)
[14:24:16][Step 3/3] at scala.collection.immutable.List.foreach(List.scala:381)
[14:24:16][Step 3/3] at org.scalatest.SuperEngine.traverseSubNodes$1(Engine.scala:401)
[14:24:16][Step 3/3] at org.scalatest.SuperEngine.org$scalatest$SuperEngine$$runTestsInBranch(Engine.scala:396)
[14:24:16][Step 3/3] at org.scalatest.SuperEngine.runTestsImpl(Engine.scala:483)
[14:24:16][Step 3/3] at org.scalatest.FunSuiteLike$class.runTests(FunSuiteLike.scala:200)
[14:24:16][Step 3/3] at org.scalatest.FunSuite.runTests(FunSuite.scala:1559)
[14:24:16][Step 3/3] at org.scalatest.Suite$class.run(Suite.scala:1423)
[14:24:16][Step 3/3] at org.scalatest.FunSuite.org$scalatest$FunSuiteLike$$super$run(FunSuite.scala:1559)
[14:24:16][Step 3/3] at org.scalatest.FunSuiteLike$$anonfun$run$1.apply(FunSuiteLike.scala:204)
[14:24:16][Step 3/3] at org.scalatest.FunSuiteLike$$anonfun$run$1.apply(FunSuiteLike.scala:204)
[14:24:16][Step 3/3] at org.scalatest.SuperEngine.runImpl(Engine.scala:545)
[14:24:16][Step 3/3] at org.scalatest.FunSuiteLike$class.run(FunSuiteLike.scala:204)
[14:24:16][Step 3/3] at mesosphere.marathon.integration.ResidentTaskIntegrationTest.org$scalatest$BeforeAndAfterAllConfigMap$$super$run(ResidentTaskIntegrationTest.scala:16)
[14:24:16][Step 3/3] at org.scalatest.BeforeAndAfterAllConfigMap$class.liftedTree1$1(BeforeAndAfterAllConfigMap.scala:248)
[14:24:16][Step 3/3] at org.scalatest.BeforeAndAfterAllConfigMap$class.run(BeforeAndAfterAllConfigMap.scala:247)
[14:24:16][Step 3/3] at mesosphere.marathon.integration.ResidentTaskIntegrationTest.org$scalatest$BeforeAndAfter$$super$run(ResidentTaskIntegrationTest.scala:16)
[14:24:16][Step 3/3] at org.scalatest.BeforeAndAfter$class.run(BeforeAndAfter.scala:241)
[14:24:16][Step 3/3] at mesosphere.marathon.integration.ResidentTaskIntegrationTest.run(ResidentTaskIntegrationTest.scala:16)
[14:24:16][Step 3/3] at org.scalatest.tools.Framework.org$scalatest$tools$Framework$$runSuite(Framework.scala:444)
[14:24:16][Step 3/3] at org.scalatest.tools.Framework$ScalaTestTask.execute(Framework.scala:651)
[14:24:16][Step 3/3] at sbt.ForkMain$Run$2.call(ForkMain.java:294)
[14:24:16][Step 3/3] at sbt.ForkMain$Run$2.call(ForkMain.java:284)
[14:24:16][Step 3/3] at java.util.concurrent.FutureTask.run(FutureTask.java:266)
[14:24:16][Step 3/3] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
[14:24:16][Step 3/3] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
[14:24:16][Step 3/3] at java.lang.Thread.run(Thread.java:745)
[14:24:16][Step 3/3] + Given An app that writes into a persistent volume
[14:24:16][Step 3/3] + When a task is launched
[14:24:16][Step 3/3] + Then it successfully writes to the persistent volume and then finishes
``` | build | resident tasks flaky test persistent volume will be re attached and keep state persistent volume will be re attached and keep state failed seconds milliseconds java lang assertionerror waiting for event deployment success to arrive took longer than seconds give up at mesosphere marathon integration setup waittestsupport next waittestsupport scala at mesosphere marathon integration setup waittestsupport waitfor waittestsupport scala at mesosphere marathon integration setup marathoncallbacktestsupport class waitforeventmatching marathoncallbacktestsupport scala at mesosphere marathon integration residenttaskintegrationtest waitforeventmatching residenttaskintegrationtest scala at mesosphere marathon integration setup marathoncallbacktestsupport class waitforeventwith marathoncallbacktestsupport scala at mesosphere marathon integration residenttaskintegrationtest waitforeventwith residenttaskintegrationtest scala at mesosphere marathon integration setup marathoncallbacktestsupport class waitforevent marathoncallbacktestsupport scala at mesosphere marathon integration residenttaskintegrationtest waitforevent residenttaskintegrationtest scala at mesosphere marathon integration residenttaskintegrationtest anonfun apply residenttaskintegrationtest scala at mesosphere marathon integration residenttaskintegrationtest anonfun apply residenttaskintegrationtest scala at mesosphere marathon integration residenttaskintegrationtest anonfun test apply mcv sp residenttaskintegrationtest scala at mesosphere marathon integration residenttaskintegrationtest anonfun test apply residenttaskintegrationtest scala at mesosphere marathon integration residenttaskintegrationtest anonfun test apply residenttaskintegrationtest scala at org scalatest transformer anonfun apply apply mcv sp transformer scala at org scalatest outcomeof class outcomeof outcomeof scala at org scalatest outcomeof outcomeof outcomeof scala at org scalatest transformer apply transformer scala at org scalatest transformer apply transformer scala at org scalatest funsuitelike anon apply funsuitelike scala at org scalatest suite class withfixture suite scala at org scalatest funsuite withfixture funsuite scala at org scalatest funsuitelike class invokewithfixture funsuitelike scala at org scalatest funsuitelike anonfun runtest apply funsuitelike scala at org scalatest funsuitelike anonfun runtest apply funsuitelike scala at org scalatest superengine runtestimpl engine scala at org scalatest funsuitelike class runtest funsuitelike scala at mesosphere marathon integration residenttaskintegrationtest org scalatest beforeandafter super runtest residenttaskintegrationtest scala at org scalatest beforeandafter class runtest beforeandafter scala at mesosphere marathon integration residenttaskintegrationtest runtest residenttaskintegrationtest scala at org scalatest funsuitelike anonfun runtests apply funsuitelike scala at org scalatest funsuitelike anonfun runtests apply funsuitelike scala at org scalatest superengine anonfun traversesubnodes apply engine scala at org scalatest superengine anonfun traversesubnodes apply engine scala at scala collection immutable list foreach list scala at org scalatest superengine traversesubnodes engine scala at org scalatest superengine org scalatest superengine runtestsinbranch engine scala at org scalatest superengine runtestsimpl engine scala at org scalatest funsuitelike class runtests funsuitelike scala at org scalatest funsuite runtests funsuite scala at org scalatest suite class run suite scala at org scalatest funsuite org scalatest funsuitelike super run funsuite scala at org scalatest funsuitelike anonfun run apply funsuitelike scala at org scalatest funsuitelike anonfun run apply funsuitelike scala at org scalatest superengine runimpl engine scala at org scalatest funsuitelike class run funsuitelike scala at mesosphere marathon integration residenttaskintegrationtest org scalatest beforeandafterallconfigmap super run residenttaskintegrationtest scala at org scalatest beforeandafterallconfigmap class beforeandafterallconfigmap scala at org scalatest beforeandafterallconfigmap class run beforeandafterallconfigmap scala at mesosphere marathon integration residenttaskintegrationtest org scalatest beforeandafter super run residenttaskintegrationtest scala at org scalatest beforeandafter class run beforeandafter scala at mesosphere marathon integration residenttaskintegrationtest run residenttaskintegrationtest scala at org scalatest tools framework org scalatest tools framework runsuite framework scala at org scalatest tools framework scalatesttask execute framework scala at sbt forkmain run call forkmain java at sbt forkmain run call forkmain java at java util concurrent futuretask run futuretask java at java util concurrent threadpoolexecutor runworker threadpoolexecutor java at java util concurrent threadpoolexecutor worker run threadpoolexecutor java at java lang thread run thread java given an app that writes into a persistent volume when a task is launched then it successfully writes to the persistent volume and then finishes |
6 | 3,874,113,297 | IssuesEvent | 2016-04-11 19:20:57 | openshift/origin | https://api.github.com/repos/openshift/origin | closed | Update build config name annotation on builds when renaming build config | area/techdebt component/build priority/P2 | We should update the build config name annotation when renaming the BC, so we don't loose the reference to BC the builds were started from.
This might affect removing the builds when removing the build config. It might also break serial builds when we can't check what scheduling policy the BC associated with build has. | 1 | Update build config name annotation on builds when renaming build config - We should update the build config name annotation when renaming the BC, so we don't loose the reference to BC the builds were started from.
This might affect removing the builds when removing the build config. It might also break serial builds when we can't check what scheduling policy the BC associated with build has. | build | update build config name annotation on builds when renaming build config we should update the build config name annotation when renaming the bc so we don t loose the reference to bc the builds were started from this might affect removing the builds when removing the build config it might also break serial builds when we can t check what scheduling policy the bc associated with build has |
7 | 3,908,821,241 | IssuesEvent | 2016-04-19 17:07:05 | ManageIQ/manageiq | https://api.github.com/repos/ManageIQ/manageiq | closed | Docker: switch the clone path from /manageiq to /var/www/miq/vmdb | build technical debt | cc @bazulay @fbladilo | 1 | Docker: switch the clone path from /manageiq to /var/www/miq/vmdb - cc @bazulay @fbladilo | build | docker switch the clone path from manageiq to var www miq vmdb cc bazulay fbladilo |
8 | 4,282,752,050 | IssuesEvent | 2016-07-15 10:27:48 | CartoDB/cartodb | https://api.github.com/repos/CartoDB/cartodb | opened | Separate models for different layers | Builder technical-debt | Right now we basically have at least three different kind of layers (in the builder context, not to be confused with the layer types in the map!), each with their own kind of view and behavior:
There are various [kinds/types that defines a layer](https://github.com/CartoDB/cartodb/blob/f3d7f4be2af99974f2bfb0c71aa3eefbcc1614b0/lib/assets/javascripts/cartodb3/data/layer-types-and-kinds.js#L1-L17), but they can basically be put in three distinct categories:
- layer-on-top (tiled)
- data layers (cartodb, torque)
- basemaps - (tiled, wms etc.)
Right now we only have [one model](https://github.com/CartoDB/cartodb/blob/master/lib/assets/javascripts/cartodb3/data/layer-definition-model.js) that represents them all, even if their behavior and data varies vastly.
Just like we have done with some other models I think it would make sense to have separate implementations for these three categories.
This way we can extract some business logic related to basemaps out from the views (see #8901 for an example), as well as for basemaps not having to know anything about analysis.
cc @xavijam @matallo @alonsogarciapablo | 1 | Separate models for different layers - Right now we basically have at least three different kind of layers (in the builder context, not to be confused with the layer types in the map!), each with their own kind of view and behavior:
There are various [kinds/types that defines a layer](https://github.com/CartoDB/cartodb/blob/f3d7f4be2af99974f2bfb0c71aa3eefbcc1614b0/lib/assets/javascripts/cartodb3/data/layer-types-and-kinds.js#L1-L17), but they can basically be put in three distinct categories:
- layer-on-top (tiled)
- data layers (cartodb, torque)
- basemaps - (tiled, wms etc.)
Right now we only have [one model](https://github.com/CartoDB/cartodb/blob/master/lib/assets/javascripts/cartodb3/data/layer-definition-model.js) that represents them all, even if their behavior and data varies vastly.
Just like we have done with some other models I think it would make sense to have separate implementations for these three categories.
This way we can extract some business logic related to basemaps out from the views (see #8901 for an example), as well as for basemaps not having to know anything about analysis.
cc @xavijam @matallo @alonsogarciapablo | build | separate models for different layers right now we basically have at least three different kind of layers in the builder context not to be confused with the layer types in the map each with their own kind of view and behavior there are various but they can basically be put in three distinct categories layer on top tiled data layers cartodb torque basemaps tiled wms etc right now we only have that represents them all even if their behavior and data varies vastly just like we have done with some other models i think it would make sense to have separate implementations for these three categories this way we can extract some business logic related to basemaps out from the views see for an example as well as for basemaps not having to know anything about analysis cc xavijam matallo alonsogarciapablo |
9 | 4,388,555,625 | IssuesEvent | 2016-08-08 19:15:45 | FineUploader/fine-uploader | https://api.github.com/repos/FineUploader/fine-uploader | closed | Allow more current version of node/npm for build, & clean up build | 5 - Done build technical debt | - [x] npm script to clean the build artifacts
- [x] ... to run linter(s)
- [x] ... to build all.fineuploader.js for manual/automated tests + source maps
- [x] ... to run unit tests (FF only for now)
- [x] ... to generate zip files for all builds of FU for download
- [x] ... to generate build for distribution via npm
- [x] ensure current version of npm/node can be used to build
- [x] Remove grunt & all grunt-related dependencies.
- [x] update travis-ci scripts
- [x] check build files again and [increase minification](https://davidwalsh.name/compress-uglify).
- [x] update build instructions
- [x] update downloads section of fineuploader.com
It would be nice to allow more modern versions of node to work for development. The current
solution is to use nvm to install 0.10.33 for FU development.
I'll take this opportunity to clean up the build entirely and remove all of the grunt-related cruft too. | 1 | Allow more current version of node/npm for build, & clean up build - - [x] npm script to clean the build artifacts
- [x] ... to run linter(s)
- [x] ... to build all.fineuploader.js for manual/automated tests + source maps
- [x] ... to run unit tests (FF only for now)
- [x] ... to generate zip files for all builds of FU for download
- [x] ... to generate build for distribution via npm
- [x] ensure current version of npm/node can be used to build
- [x] Remove grunt & all grunt-related dependencies.
- [x] update travis-ci scripts
- [x] check build files again and [increase minification](https://davidwalsh.name/compress-uglify).
- [x] update build instructions
- [x] update downloads section of fineuploader.com
It would be nice to allow more modern versions of node to work for development. The current
solution is to use nvm to install 0.10.33 for FU development.
I'll take this opportunity to clean up the build entirely and remove all of the grunt-related cruft too. | build | allow more current version of node npm for build clean up build npm script to clean the build artifacts to run linter s to build all fineuploader js for manual automated tests source maps to run unit tests ff only for now to generate zip files for all builds of fu for download to generate build for distribution via npm ensure current version of npm node can be used to build remove grunt all grunt related dependencies update travis ci scripts check build files again and update build instructions update downloads section of fineuploader com it would be nice to allow more modern versions of node to work for development the current solution is to use nvm to install for fu development i ll take this opportunity to clean up the build entirely and remove all of the grunt related cruft too |
10 | 4,928,457,071 | IssuesEvent | 2016-11-27 10:10:39 | GoogleCloudPlatform/google-cloud-eclipse | https://api.github.com/repos/GoogleCloudPlatform/google-cloud-eclipse | closed | Hamcrest jars in release | build high priority Tech Debt | Our release build bundles in 3 hamcrest jars. These should only be needed for tests, not for the production code. Can we get rid of them? | 1 | Hamcrest jars in release - Our release build bundles in 3 hamcrest jars. These should only be needed for tests, not for the production code. Can we get rid of them? | build | hamcrest jars in release our release build bundles in hamcrest jars these should only be needed for tests not for the production code can we get rid of them |
13 | 5,194,460,423 | IssuesEvent | 2017-01-23 03:52:09 | openshift/origin | https://api.github.com/repos/openshift/origin | closed | Refactor build controllers by controller type to match deployments | area/techdebt component/build kind/enhancement priority/P3 | Create a separate package per controller/factory pair.
| 1 | Refactor build controllers by controller type to match deployments - Create a separate package per controller/factory pair.
| build | refactor build controllers by controller type to match deployments create a separate package per controller factory pair |
14 | 5,206,063,714 | IssuesEvent | 2017-01-24 19:36:31 | weaveworks/scope | https://api.github.com/repos/weaveworks/scope | closed | Add unit tests for the ECS reporter | component/build ecs techdebt | Leftover from https://github.com/weaveworks/scope/pull/2026 | 1 | Add unit tests for the ECS reporter - Leftover from https://github.com/weaveworks/scope/pull/2026 | build | add unit tests for the ecs reporter leftover from |
15 | 5,207,015,869 | IssuesEvent | 2017-01-24 22:13:54 | GoogleCloudPlatform/google-cloud-eclipse | https://api.github.com/repos/GoogleCloudPlatform/google-cloud-eclipse | closed | Get freemarker from maven central if possible | build Tech Debt | rather than bundling it | 1 | Get freemarker from maven central if possible - rather than bundling it | build | get freemarker from maven central if possible rather than bundling it |
16 | 5,280,869,280 | IssuesEvent | 2017-02-07 15:14:45 | weaveworks/scope | https://api.github.com/repos/weaveworks/scope | closed | Improve debuggability of integration tests | component/build techdebt | Adding and debugging integration tests is a bit hairy. I tend to enable ssh in the CircleCI builds.
It would be good if we had the local VM creation automated through `vagrant up`, like in weave: https://github.com/weaveworks/weave/blob/master/test/README.md
| 1 | Improve debuggability of integration tests - Adding and debugging integration tests is a bit hairy. I tend to enable ssh in the CircleCI builds.
It would be good if we had the local VM creation automated through `vagrant up`, like in weave: https://github.com/weaveworks/weave/blob/master/test/README.md
| build | improve debuggability of integration tests adding and debugging integration tests is a bit hairy i tend to enable ssh in the circleci builds it would be good if we had the local vm creation automated through vagrant up like in weave |
17 | 5,397,069,445 | IssuesEvent | 2017-02-27 13:43:44 | weaveworks/scope | https://api.github.com/repos/weaveworks/scope | closed | Move common directory to its own repo | component/build techdebt | - Move it to https://github.com/weaveworks/common
- Move the packages out of the common directory
- Preserve history
- Remove stuff that doesn't make sense to be in common
- Make sure it builds (locally & with CI)
- Update scope
Motivation is that:
- we are using this code in a variety of other code bases which have little to do with scope
- we have *duplicated* some of this code in other code bases, leading to bugs and unexpected behaviour | 1 | Move common directory to its own repo - - Move it to https://github.com/weaveworks/common
- Move the packages out of the common directory
- Preserve history
- Remove stuff that doesn't make sense to be in common
- Make sure it builds (locally & with CI)
- Update scope
Motivation is that:
- we are using this code in a variety of other code bases which have little to do with scope
- we have *duplicated* some of this code in other code bases, leading to bugs and unexpected behaviour | build | move common directory to its own repo move it to move the packages out of the common directory preserve history remove stuff that doesn t make sense to be in common make sure it builds locally with ci update scope motivation is that we are using this code in a variety of other code bases which have little to do with scope we have duplicated some of this code in other code bases leading to bugs and unexpected behaviour |
18 | 5,475,708,465 | IssuesEvent | 2017-03-11 14:02:40 | FakeItEasy/FakeItEasy | https://api.github.com/repos/FakeItEasy/FakeItEasy | opened | Drop resource files and move message strings to the code | build tech-debt | As discussed in #1013
Error messages are currently in a resource file (.resx), but are not localized and there's no plan to localize them.
Also, the custom tool that generates code from the resource files produces code that is not compatible with .NET Standard 1.6, so it's currently disconnected. This means that adding a string to the resource file no longer updates the corresponding code file.
Since we don't need the features provided by resx files and can't use the built-in custom tool, it would make sense to convert the resources to string constants in a class. We could also use the approach suggested by @blairconrad in https://github.com/FakeItEasy/FakeItEasy/pull/1013#issuecomment-285865877 | 1 | Drop resource files and move message strings to the code - As discussed in #1013
Error messages are currently in a resource file (.resx), but are not localized and there's no plan to localize them.
Also, the custom tool that generates code from the resource files produces code that is not compatible with .NET Standard 1.6, so it's currently disconnected. This means that adding a string to the resource file no longer updates the corresponding code file.
Since we don't need the features provided by resx files and can't use the built-in custom tool, it would make sense to convert the resources to string constants in a class. We could also use the approach suggested by @blairconrad in https://github.com/FakeItEasy/FakeItEasy/pull/1013#issuecomment-285865877 | build | drop resource files and move message strings to the code as discussed in error messages are currently in a resource file resx but are not localized and there s no plan to localize them also the custom tool that generates code from the resource files produces code that is not compatible with net standard so it s currently disconnected this means that adding a string to the resource file no longer updates the corresponding code file since we don t need the features provided by resx files and can t use the built in custom tool it would make sense to convert the resources to string constants in a class we could also use the approach suggested by blairconrad in |
19 | 5,868,275,987 | IssuesEvent | 2017-05-14 11:09:27 | FakeItEasy/FakeItEasy | https://api.github.com/repos/FakeItEasy/FakeItEasy | closed | Generate external assembly with extension points for integration tests | build in-progress P2 tech-debt | As noted in https://github.com/FakeItEasy/FakeItEasy/pull/1090#issuecomment-300454071, the current assembly will not be compiled if a user runs the integration tests from the IDE without doing a command-line build, and then the tests would fail.
We could generate the assembly as part of the test fixture setup. | 1 | Generate external assembly with extension points for integration tests - As noted in https://github.com/FakeItEasy/FakeItEasy/pull/1090#issuecomment-300454071, the current assembly will not be compiled if a user runs the integration tests from the IDE without doing a command-line build, and then the tests would fail.
We could generate the assembly as part of the test fixture setup. | build | generate external assembly with extension points for integration tests as noted in the current assembly will not be compiled if a user runs the integration tests from the ide without doing a command line build and then the tests would fail we could generate the assembly as part of the test fixture setup |
21 | 6,188,154,399 | IssuesEvent | 2017-07-04 09:26:51 | hyperledger/composer | https://api.github.com/repos/hyperledger/composer | closed | Link checking within docs should be automated | build debt docs qa | Related to #449
Documentation links need to be checked via a tool as part of an automated test. This needs to be part of the test pipe-line so that we can gain confidence in the documentation quality with regards to internal linking.
## Context
Issues are being raised relating to broken links; this can be avoided by incorporating a link checker within the automated tests
## Expected Behavior
If there is a broken link in the documentation, a test will fail
## Actual Behavior
No tests to fail, documents released with unintentional broken links
## Possible Fix
Incorporate a link checker into testing pipeline | 1 | Link checking within docs should be automated - Related to #449
Documentation links need to be checked via a tool as part of an automated test. This needs to be part of the test pipe-line so that we can gain confidence in the documentation quality with regards to internal linking.
## Context
Issues are being raised relating to broken links; this can be avoided by incorporating a link checker within the automated tests
## Expected Behavior
If there is a broken link in the documentation, a test will fail
## Actual Behavior
No tests to fail, documents released with unintentional broken links
## Possible Fix
Incorporate a link checker into testing pipeline | build | link checking within docs should be automated related to documentation links need to be checked via a tool as part of an automated test this needs to be part of the test pipe line so that we can gain confidence in the documentation quality with regards to internal linking context issues are being raised relating to broken links this can be avoided by incorporating a link checker within the automated tests expected behavior if there is a broken link in the documentation a test will fail actual behavior no tests to fail documents released with unintentional broken links possible fix incorporate a link checker into testing pipeline |
24 | 6,342,832,192 | IssuesEvent | 2017-07-27 16:15:38 | yahoo/fili | https://api.github.com/repos/yahoo/fili | closed | Maven warnings on build | BUILD TECH-DEBT | Parent POM warnings from `mvn install`
Version: Apache Maven 3.5.0
```
[WARNING]
[WARNING] Some problems were encountered while building the effective model for com.yahoo.fili:fili-system-config:jar:0.9-SNAPSHOT
[WARNING] Reporting configuration should be done in <reporting> section, not in maven-site-plugin <configuration> as reportPlugins parameter.
[WARNING]
[WARNING] Some problems were encountered while building the effective model for com.yahoo.fili:fili-core:jar:0.9-SNAPSHOT
[WARNING] Reporting configuration should be done in <reporting> section, not in maven-site-plugin <configuration> as reportPlugins parameter.
[WARNING]
[WARNING] Some problems were encountered while building the effective model for com.yahoo.fili:fili-navi:jar:0.9-SNAPSHOT
[WARNING] Reporting configuration should be done in <reporting> section, not in maven-site-plugin <configuration> as reportPlugins parameter.
[WARNING]
[WARNING] Some problems were encountered while building the effective model for com.yahoo.fili:fili:jar:0.9-SNAPSHOT
[WARNING] Reporting configuration should be done in <reporting> section, not in maven-site-plugin <configuration> as reportPlugins parameter.
[WARNING]
[WARNING] Some problems were encountered while building the effective model for com.yahoo.fili:fili-wikipedia-example:war:0.9-SNAPSHOT
[WARNING] Reporting configuration should be done in <reporting> section, not in maven-site-plugin <configuration> as reportPlugins parameter.
[WARNING]
[WARNING] Some problems were encountered while building the effective model for com.yahoo.fili:fili-generic-example:war:0.9-SNAPSHOT
[WARNING] Reporting configuration should be done in <reporting> section, not in maven-site-plugin <configuration> as reportPlugins parameter.
[WARNING]
[WARNING] Some problems were encountered while building the effective model for com.yahoo.fili:fili-parent-pom:pom:0.9-SNAPSHOT
[WARNING] Reporting configuration should be done in <reporting> section, not in maven-site-plugin <configuration> as reportPlugins parameter. @ line 768, column 32
[WARNING]
[WARNING] It is highly recommended to fix these problems because they threaten the stability of your build.
[WARNING]
[WARNING] For this reason, future Maven versions might no longer support building such malformed projects.
```
it looks like we have some potential build issues with the current config. | 1 | Maven warnings on build - Parent POM warnings from `mvn install`
Version: Apache Maven 3.5.0
```
[WARNING]
[WARNING] Some problems were encountered while building the effective model for com.yahoo.fili:fili-system-config:jar:0.9-SNAPSHOT
[WARNING] Reporting configuration should be done in <reporting> section, not in maven-site-plugin <configuration> as reportPlugins parameter.
[WARNING]
[WARNING] Some problems were encountered while building the effective model for com.yahoo.fili:fili-core:jar:0.9-SNAPSHOT
[WARNING] Reporting configuration should be done in <reporting> section, not in maven-site-plugin <configuration> as reportPlugins parameter.
[WARNING]
[WARNING] Some problems were encountered while building the effective model for com.yahoo.fili:fili-navi:jar:0.9-SNAPSHOT
[WARNING] Reporting configuration should be done in <reporting> section, not in maven-site-plugin <configuration> as reportPlugins parameter.
[WARNING]
[WARNING] Some problems were encountered while building the effective model for com.yahoo.fili:fili:jar:0.9-SNAPSHOT
[WARNING] Reporting configuration should be done in <reporting> section, not in maven-site-plugin <configuration> as reportPlugins parameter.
[WARNING]
[WARNING] Some problems were encountered while building the effective model for com.yahoo.fili:fili-wikipedia-example:war:0.9-SNAPSHOT
[WARNING] Reporting configuration should be done in <reporting> section, not in maven-site-plugin <configuration> as reportPlugins parameter.
[WARNING]
[WARNING] Some problems were encountered while building the effective model for com.yahoo.fili:fili-generic-example:war:0.9-SNAPSHOT
[WARNING] Reporting configuration should be done in <reporting> section, not in maven-site-plugin <configuration> as reportPlugins parameter.
[WARNING]
[WARNING] Some problems were encountered while building the effective model for com.yahoo.fili:fili-parent-pom:pom:0.9-SNAPSHOT
[WARNING] Reporting configuration should be done in <reporting> section, not in maven-site-plugin <configuration> as reportPlugins parameter. @ line 768, column 32
[WARNING]
[WARNING] It is highly recommended to fix these problems because they threaten the stability of your build.
[WARNING]
[WARNING] For this reason, future Maven versions might no longer support building such malformed projects.
```
it looks like we have some potential build issues with the current config. | build | maven warnings on build parent pom warnings from mvn install version apache maven some problems were encountered while building the effective model for com yahoo fili fili system config jar snapshot reporting configuration should be done in section not in maven site plugin as reportplugins parameter some problems were encountered while building the effective model for com yahoo fili fili core jar snapshot reporting configuration should be done in section not in maven site plugin as reportplugins parameter some problems were encountered while building the effective model for com yahoo fili fili navi jar snapshot reporting configuration should be done in section not in maven site plugin as reportplugins parameter some problems were encountered while building the effective model for com yahoo fili fili jar snapshot reporting configuration should be done in section not in maven site plugin as reportplugins parameter some problems were encountered while building the effective model for com yahoo fili fili wikipedia example war snapshot reporting configuration should be done in section not in maven site plugin as reportplugins parameter some problems were encountered while building the effective model for com yahoo fili fili generic example war snapshot reporting configuration should be done in section not in maven site plugin as reportplugins parameter some problems were encountered while building the effective model for com yahoo fili fili parent pom pom snapshot reporting configuration should be done in section not in maven site plugin as reportplugins parameter line column it is highly recommended to fix these problems because they threaten the stability of your build for this reason future maven versions might no longer support building such malformed projects it looks like we have some potential build issues with the current config |
25 | 6,497,910,317 | IssuesEvent | 2017-08-22 15:28:58 | openshift/origin | https://api.github.com/repos/openshift/origin | closed | refactor build admission plugins | area/techdebt component/build priority/P3 | they're not run at admission any more, and they run in the context of the build controller - so the code should move location, and they should take advantage of having access to the build object rather than doing lots of serialisation/deserialisation | 1 | refactor build admission plugins - they're not run at admission any more, and they run in the context of the build controller - so the code should move location, and they should take advantage of having access to the build object rather than doing lots of serialisation/deserialisation | build | refactor build admission plugins they re not run at admission any more and they run in the context of the build controller so the code should move location and they should take advantage of having access to the build object rather than doing lots of serialisation deserialisation |
26 | 6,498,503,681 | IssuesEvent | 2017-08-22 17:41:14 | vterm/vterm | https://api.github.com/repos/vterm/vterm | opened | Build for deb | debt errors/build platform/linux | Because of #58, we had to remove the "deb" target for Linux builds.
This needs to be solved. | 1 | Build for deb - Because of #58, we had to remove the "deb" target for Linux builds.
This needs to be solved. | build | build for deb because of we had to remove the deb target for linux builds this needs to be solved |
27 | 6,515,403,042 | IssuesEvent | 2017-08-26 15:18:17 | openshift/ansible-service-broker | https://api.github.com/repos/openshift/ansible-service-broker | closed | Fix make prep_local to handle the auth directory and files. | bug build tech-debt | `make run` won't work quite right with auth enabled. | 1 | Fix make prep_local to handle the auth directory and files. - `make run` won't work quite right with auth enabled. | build | fix make prep local to handle the auth directory and files make run won t work quite right with auth enabled |
28 | 6,552,986,166 | IssuesEvent | 2017-09-05 20:33:42 | openshift/origin | https://api.github.com/repos/openshift/origin | closed | build controller: use retry in buildconfig policy to determine what build to run next | area/techdebt component/build priority/P2 | With https://github.com/openshift/origin/pull/16055 we are reverting to using the REST client to determine which build to run next when transitioning a build to a completed phase.
A better solution would be to keep using the cache and simply retry when determining the build to run next and a build is still running. | 1 | build controller: use retry in buildconfig policy to determine what build to run next - With https://github.com/openshift/origin/pull/16055 we are reverting to using the REST client to determine which build to run next when transitioning a build to a completed phase.
A better solution would be to keep using the cache and simply retry when determining the build to run next and a build is still running. | build | build controller use retry in buildconfig policy to determine what build to run next with we are reverting to using the rest client to determine which build to run next when transitioning a build to a completed phase a better solution would be to keep using the cache and simply retry when determining the build to run next and a build is still running |
29 | 6,843,161,771 | IssuesEvent | 2017-11-12 12:22:20 | adamralph/liteguard | https://api.github.com/repos/adamralph/liteguard | closed | Fix up test project | build in-progress tech-debt | - Reference Test SDK
- Reference xunit VS runner
- Replace xunit.runner.console in build packages with .NET CLI tool reference in project
- Simplify test build target | 1 | Fix up test project - - Reference Test SDK
- Reference xunit VS runner
- Replace xunit.runner.console in build packages with .NET CLI tool reference in project
- Simplify test build target | build | fix up test project reference test sdk reference xunit vs runner replace xunit runner console in build packages with net cli tool reference in project simplify test build target |