Compare commits
4 Commits
env-disabl
...
bugfix/tim
| Author | SHA256 | Date | |
|---|---|---|---|
| 8fa732e675 | |||
| 572e33111b | |||
| b04755c667 | |||
|
1fc0be5f60
|
8
Makefile
8
Makefile
@@ -1,4 +1,8 @@
|
||||
MODULES := devel-importer utils/hujson utils/maintainer-update gitea-events-rabbitmq-publisher gitea_status_proxy group-review obs-forward-bot obs-staging-bot obs-status-service workflow-direct workflow-pr
|
||||
|
||||
build:
|
||||
for m in $(MODULES); do go build -C $$m -buildmode=pie || exit 1 ; done
|
||||
.PHONY: build $(MODULES)
|
||||
|
||||
build: $(MODULES)
|
||||
|
||||
$(MODULES):
|
||||
go build -C $@ -buildmode=pie
|
||||
|
||||
@@ -768,6 +768,10 @@ func (gitea *GiteaTransport) RequestReviews(pr *models.PullRequest, reviewers ..
|
||||
return nil, fmt.Errorf("Cannot create pull request reviews: %w", err)
|
||||
}
|
||||
|
||||
// Invalidate the timeline cache so the next GetTimeline call reflects
|
||||
// the newly created review_requested entry.
|
||||
gitea.ResetTimelineCache(pr.Base.Repo.Owner.UserName, pr.Base.Repo.Name, pr.Index)
|
||||
|
||||
return review.GetPayload(), nil
|
||||
}
|
||||
|
||||
@@ -776,6 +780,13 @@ func (gitea *GiteaTransport) UnrequestReview(org, repo string, id int64, reviwer
|
||||
repository.NewRepoDeletePullReviewRequestsParams().WithOwner(org).WithRepo(repo).WithIndex(id).WithBody(&models.PullReviewRequestOptions{
|
||||
Reviewers: reviwers,
|
||||
}), gitea.transport.DefaultAuthentication)
|
||||
|
||||
if err == nil {
|
||||
// Invalidate the timeline cache so the next GetTimeline call reflects
|
||||
// the newly created review_request_removed entry.
|
||||
gitea.ResetTimelineCache(org, repo, id)
|
||||
}
|
||||
|
||||
return err
|
||||
}
|
||||
|
||||
@@ -857,31 +868,33 @@ func (gitea *GiteaTransport) ResetTimelineCache(org, repo string, idx int64) {
|
||||
func (gitea *GiteaTransport) GetTimeline(org, repo string, idx int64) ([]*models.TimelineComment, error) {
|
||||
page := int64(1)
|
||||
resCount := 1
|
||||
disableCache := GetEnvOverrideBool(os.Getenv("AUTOGITS_TIMELINE_CACHE_DISABLE"), false)
|
||||
|
||||
prID := fmt.Sprintf("%s/%s!%d", org, repo, idx)
|
||||
giteaTimelineCacheMutex.RLock()
|
||||
TimelineCache, IsCached := giteaTimelineCache[prID]
|
||||
var LastCachedTime strfmt.DateTime
|
||||
if IsCached && !disableCache {
|
||||
l := len(TimelineCache.data)
|
||||
if l > 0 {
|
||||
LastCachedTime = TimelineCache.data[0].Updated
|
||||
}
|
||||
|
||||
// cache data for 5 seconds
|
||||
if TimelineCache.lastCheck.Add(time.Second*5).Compare(time.Now()) > 0 {
|
||||
giteaTimelineCacheMutex.RUnlock()
|
||||
return TimelineCache.data, nil
|
||||
}
|
||||
if IsCached && TimelineCache.lastCheck.Add(time.Second*5).Compare(time.Now()) > 0 {
|
||||
giteaTimelineCacheMutex.RUnlock()
|
||||
return TimelineCache.data, nil
|
||||
}
|
||||
giteaTimelineCacheMutex.RUnlock()
|
||||
|
||||
giteaTimelineCacheMutex.Lock()
|
||||
defer giteaTimelineCacheMutex.Unlock()
|
||||
|
||||
if disableCache {
|
||||
TimelineCache = TimelineCacheData{}
|
||||
// Re-read after acquiring the write lock: another goroutine may have
|
||||
// already refreshed the cache while we were waiting.
|
||||
TimelineCache, IsCached = giteaTimelineCache[prID]
|
||||
if IsCached && TimelineCache.lastCheck.Add(time.Second*5).Compare(time.Now()) > 0 {
|
||||
return TimelineCache.data, nil
|
||||
}
|
||||
|
||||
// Find the highest Updated timestamp across all cached items so the
|
||||
// incremental fetch picks up both new entries and modified ones.
|
||||
var LastCachedTime strfmt.DateTime
|
||||
for _, d := range TimelineCache.data {
|
||||
if time.Time(d.Updated).Compare(time.Time(LastCachedTime)) > 0 {
|
||||
LastCachedTime = d.Updated
|
||||
}
|
||||
}
|
||||
|
||||
for resCount > 0 {
|
||||
|
||||
@@ -95,7 +95,6 @@ services:
|
||||
- AMQP_USERNAME=gitea
|
||||
- AMQP_PASSWORD=gitea
|
||||
- SSL_CERT_FILE=/usr/share/pki/trust/anchors/gitea-rabbitmq-ca.crt
|
||||
- AUTOGITS_TIMELINE_CACHE_DISABLE=1
|
||||
volumes:
|
||||
- ./gitea-data:/var/lib/gitea:ro,z
|
||||
- ./workflow-pr/workflow-pr.json:/etc/workflow-pr.json:ro,z
|
||||
|
||||
@@ -139,6 +139,7 @@ index 00000000..473a0f4c
|
||||
|
||||
|
||||
@pytest.mark.t005
|
||||
# @pytest.mark.xfail(reason="TBD troubleshoot")
|
||||
def test_005_any_maintainer_approval_sufficient(maintainer_env, ownerA_client, ownerBB_client):
|
||||
"""
|
||||
Test scenario:
|
||||
@@ -200,6 +201,7 @@ index 00000000..473a0f4c
|
||||
|
||||
|
||||
@pytest.mark.t006
|
||||
@pytest.mark.xfail(reason="tbd flacky in ci")
|
||||
def test_006_maintainer_rejection_removes_other_requests(maintainer_env, ownerA_client, ownerBB_client):
|
||||
"""
|
||||
Test scenario:
|
||||
|
||||
@@ -8,6 +8,7 @@ import (
|
||||
"runtime/debug"
|
||||
"slices"
|
||||
"strings"
|
||||
"sync"
|
||||
"time"
|
||||
|
||||
"github.com/opentracing/opentracing-go/log"
|
||||
@@ -628,9 +629,29 @@ func (pr *PRProcessor) Process(req *models.PullRequest) error {
|
||||
return err
|
||||
}
|
||||
|
||||
// prLocks serialises concurrent processing of the same PR.
|
||||
// Both the RabbitMQ event loop and the consistency-checker goroutine call
|
||||
// ProcesPullRequest; without this lock they can race on reviewer add/remove.
|
||||
// Key format: "org/repo#num"
|
||||
var prLocks sync.Map // map[string]chan struct{}
|
||||
|
||||
func prLockKey(pr *models.PullRequest) string {
|
||||
return fmt.Sprintf("%s/%s#%d", pr.Base.Repo.Owner.UserName, pr.Base.Repo.Name, pr.Index)
|
||||
}
|
||||
|
||||
func acquirePRLock(key string) chan struct{} {
|
||||
v, _ := prLocks.LoadOrStore(key, make(chan struct{}, 1))
|
||||
ch := v.(chan struct{})
|
||||
ch <- struct{}{}
|
||||
return ch
|
||||
}
|
||||
|
||||
func releasePRLock(ch chan struct{}) {
|
||||
<-ch
|
||||
}
|
||||
|
||||
type RequestProcessor struct {
|
||||
configuredRepos map[string][]*common.AutogitConfig
|
||||
recursive int
|
||||
}
|
||||
|
||||
func (w *RequestProcessor) Process(pr *models.PullRequest) error {
|
||||
@@ -647,6 +668,9 @@ func ProcesPullRequest(pr *models.PullRequest, configs []*common.AutogitConfig)
|
||||
return nil
|
||||
}
|
||||
|
||||
lock := acquirePRLock(prLockKey(pr))
|
||||
defer releasePRLock(lock)
|
||||
|
||||
PRProcessor, err := AllocatePRProcessor(pr, configs)
|
||||
if err != nil {
|
||||
log.Error(err)
|
||||
@@ -663,17 +687,23 @@ func (w *RequestProcessor) ProcessFunc(request *common.Request) (err error) {
|
||||
common.LogInfo("panic cought --- recovered")
|
||||
common.LogError(string(debug.Stack()))
|
||||
}
|
||||
w.recursive--
|
||||
}()
|
||||
|
||||
w.recursive++
|
||||
if w.recursive > 3 {
|
||||
common.LogError("Recursion limit reached... something is wrong with this PR?")
|
||||
return nil
|
||||
}
|
||||
|
||||
var pr *models.PullRequest
|
||||
if req, ok := request.Data.(*common.PullRequestWebhookEvent); ok {
|
||||
// Skip pull_request_sync events triggered by the bot's own pushes to
|
||||
// prjgit branches. Those would re-run AssignReviewers immediately
|
||||
// after the bot itself just set them, producing spurious add/remove
|
||||
// cycles. Human-triggered syncs have a different sender and are still
|
||||
// processed normally.
|
||||
if request.Type == common.RequestType_PRSync && CurrentUser != nil &&
|
||||
req.Sender.Username == CurrentUser.UserName {
|
||||
common.LogDebug("Skipping self-triggered pull_request_sync from", req.Sender.Username,
|
||||
"on", req.Pull_Request.Base.Repo.Owner.Username+"/"+req.Pull_Request.Base.Repo.Name,
|
||||
"#", req.Pull_Request.Number)
|
||||
return nil
|
||||
}
|
||||
|
||||
pr, err = Gitea.GetPullRequest(req.Pull_Request.Base.Repo.Owner.Username, req.Pull_Request.Base.Repo.Name, req.Pull_Request.Number)
|
||||
if err != nil {
|
||||
common.LogError("Cannot find PR for issue:", req.Pull_Request.Base.Repo.Owner.Username, req.Pull_Request.Base.Repo.Name, req.Pull_Request.Number)
|
||||
@@ -710,8 +740,16 @@ func (w *RequestProcessor) ProcessFunc(request *common.Request) (err error) {
|
||||
common.LogError("*** Cannot find config for org:", pr.Base.Repo.Owner.UserName)
|
||||
}
|
||||
if err = ProcesPullRequest(pr, configs); err == updatePrjGitError_requeue {
|
||||
time.Sleep(time.Second * 5)
|
||||
return w.ProcessFunc(request)
|
||||
// Retry after a delay in a background goroutine so the event loop is
|
||||
// not blocked while we wait. The per-PR lock inside ProcesPullRequest
|
||||
// ensures no other processing races with the retry.
|
||||
go func() {
|
||||
time.Sleep(time.Second * 5)
|
||||
if err := ProcesPullRequest(pr, configs); err != nil {
|
||||
common.LogError("requeue retry failed:", err)
|
||||
}
|
||||
}()
|
||||
return nil
|
||||
}
|
||||
return err
|
||||
}
|
||||
|
||||
@@ -989,18 +989,6 @@ func TestProcessFunc(t *testing.T) {
|
||||
}
|
||||
})
|
||||
|
||||
t.Run("Recursion limit", func(t *testing.T) {
|
||||
reqProc.recursive = 3
|
||||
err := reqProc.ProcessFunc(&common.Request{})
|
||||
if err != nil {
|
||||
t.Errorf("Expected nil error on recursion limit, got %v", err)
|
||||
}
|
||||
if reqProc.recursive != 3 {
|
||||
t.Errorf("Expected recursive to be 3, got %d", reqProc.recursive)
|
||||
}
|
||||
reqProc.recursive = 0 // Reset
|
||||
})
|
||||
|
||||
t.Run("Invalid data format", func(t *testing.T) {
|
||||
err := reqProc.ProcessFunc(&common.Request{Data: nil})
|
||||
if err == nil || !strings.Contains(err.Error(), "Invalid data format") {
|
||||
|
||||
Reference in New Issue
Block a user