summaryrefslogtreecommitdiffstats
path: root/modules
diff options
context:
space:
mode:
authorLunny Xiao <xiaolunwen@gmail.com>2019-09-27 08:22:36 +0800
committerGitHub <noreply@github.com>2019-09-27 08:22:36 +0800
commiteb11ca68470e4ab60ed4ec31f4d170a4b36a528e (patch)
tree48347da8ba6c78d6ca30ac6ae5949fa93bb4c459 /modules
parentd958b9db4fa0f1910b3ca82338e3d68a70efedd9 (diff)
downloadgitea-eb11ca68470e4ab60ed4ec31f4d170a4b36a528e.tar.gz
gitea-eb11ca68470e4ab60ed4ec31f4d170a4b36a528e.zip
Extract actions on new pull request from models to pulls service and move code.gitea.io/gitea/modules/pull to code.gitea.io/gitea/services/pull (#8218)
* extract actions on new pull request from models to pulls service * improve code * move code.gitea.io/gitea/modules/pull to code.gitea.io/gitea/services/pull * fix fmt * Rename pulls.go to pull.go
Diffstat (limited to 'modules')
-rw-r--r--modules/pull/commit_status.go70
-rw-r--r--modules/pull/lfs.go226
-rw-r--r--modules/pull/merge.go320
-rw-r--r--modules/pull/review.go57
4 files changed, 0 insertions, 673 deletions
diff --git a/modules/pull/commit_status.go b/modules/pull/commit_status.go
deleted file mode 100644
index bdadc329d6..0000000000
--- a/modules/pull/commit_status.go
+++ /dev/null
@@ -1,70 +0,0 @@
-// Copyright 2019 The Gitea Authors.
-// All rights reserved.
-// Use of this source code is governed by a MIT-style
-// license that can be found in the LICENSE file.
-
-package pull
-
-import (
- "code.gitea.io/gitea/models"
- "code.gitea.io/gitea/modules/git"
-
- "github.com/pkg/errors"
-)
-
-// IsCommitStatusContextSuccess returns true if all required status check contexts succeed.
-func IsCommitStatusContextSuccess(commitStatuses []*models.CommitStatus, requiredContexts []string) bool {
- for _, ctx := range requiredContexts {
- var found bool
- for _, commitStatus := range commitStatuses {
- if commitStatus.Context == ctx {
- if commitStatus.State != models.CommitStatusSuccess {
- return false
- }
-
- found = true
- break
- }
- }
- if !found {
- return false
- }
- }
- return true
-}
-
-// IsPullCommitStatusPass returns if all required status checks PASS
-func IsPullCommitStatusPass(pr *models.PullRequest) (bool, error) {
- if err := pr.LoadProtectedBranch(); err != nil {
- return false, errors.Wrap(err, "GetLatestCommitStatus")
- }
- if pr.ProtectedBranch == nil || !pr.ProtectedBranch.EnableStatusCheck {
- return true, nil
- }
-
- // check if all required status checks are successful
- headGitRepo, err := git.OpenRepository(pr.HeadRepo.RepoPath())
- if err != nil {
- return false, errors.Wrap(err, "OpenRepository")
- }
-
- if !headGitRepo.IsBranchExist(pr.HeadBranch) {
- return false, errors.New("Head branch does not exist, can not merge")
- }
-
- sha, err := headGitRepo.GetBranchCommitID(pr.HeadBranch)
- if err != nil {
- return false, errors.Wrap(err, "GetBranchCommitID")
- }
-
- if err := pr.LoadBaseRepo(); err != nil {
- return false, errors.Wrap(err, "LoadBaseRepo")
- }
-
- commitStatuses, err := models.GetLatestCommitStatus(pr.BaseRepo, sha, 0)
- if err != nil {
- return false, errors.Wrap(err, "GetLatestCommitStatus")
- }
-
- return IsCommitStatusContextSuccess(commitStatuses, pr.ProtectedBranch.StatusCheckContexts), nil
-}
diff --git a/modules/pull/lfs.go b/modules/pull/lfs.go
deleted file mode 100644
index 2706d3a200..0000000000
--- a/modules/pull/lfs.go
+++ /dev/null
@@ -1,226 +0,0 @@
-// Copyright 2019 The Gitea Authors.
-// All rights reserved.
-// Use of this source code is governed by a MIT-style
-// license that can be found in the LICENSE file.
-
-package pull
-
-import (
- "bufio"
- "bytes"
- "fmt"
- "io"
- "strconv"
- "strings"
- "sync"
-
- "code.gitea.io/gitea/models"
- "code.gitea.io/gitea/modules/git"
- "code.gitea.io/gitea/modules/lfs"
- "code.gitea.io/gitea/modules/log"
-)
-
-// LFSPush pushes lfs objects referred to in new commits in the head repository from the base repository
-func LFSPush(tmpBasePath, mergeHeadSHA, mergeBaseSHA string, pr *models.PullRequest) error {
- // Now we have to implement git lfs push
- // git rev-list --objects --filter=blob:limit=1k HEAD --not base
- // pass blob shas in to git cat-file --batch-check (possibly unnecessary)
- // ensure only blobs and <=1k size then pass in to git cat-file --batch
- // to read each sha and check each as a pointer
- // Then if they are lfs -> add them to the baseRepo
- revListReader, revListWriter := io.Pipe()
- shasToCheckReader, shasToCheckWriter := io.Pipe()
- catFileCheckReader, catFileCheckWriter := io.Pipe()
- shasToBatchReader, shasToBatchWriter := io.Pipe()
- catFileBatchReader, catFileBatchWriter := io.Pipe()
- errChan := make(chan error, 1)
- wg := sync.WaitGroup{}
- wg.Add(6)
- // Create the go-routines in reverse order.
-
- // 6. Take the output of cat-file --batch and check if each file in turn
- // to see if they're pointers to files in the LFS store associated with
- // the head repo and add them to the base repo if so
- go readCatFileBatch(catFileBatchReader, &wg, pr)
-
- // 5. Take the shas of the blobs and batch read them
- go doCatFileBatch(shasToBatchReader, catFileBatchWriter, &wg, tmpBasePath)
-
- // 4. From the provided objects restrict to blobs <=1k
- go readCatFileBatchCheck(catFileCheckReader, shasToBatchWriter, &wg)
-
- // 3. Run batch-check on the objects retrieved from rev-list
- go doCatFileBatchCheck(shasToCheckReader, catFileCheckWriter, &wg, tmpBasePath)
-
- // 2. Check each object retrieved rejecting those without names as they will be commits or trees
- go readRevListObjects(revListReader, shasToCheckWriter, &wg)
-
- // 1. Run rev-list objects from mergeHead to mergeBase
- go doRevListObjects(revListWriter, &wg, tmpBasePath, mergeHeadSHA, mergeBaseSHA, errChan)
-
- wg.Wait()
- select {
- case err, has := <-errChan:
- if has {
- return err
- }
- default:
- }
- return nil
-}
-
-func doRevListObjects(revListWriter *io.PipeWriter, wg *sync.WaitGroup, tmpBasePath, headSHA, baseSHA string, errChan chan<- error) {
- defer wg.Done()
- defer revListWriter.Close()
- stderr := new(bytes.Buffer)
- var errbuf strings.Builder
- cmd := git.NewCommand("rev-list", "--objects", headSHA, "--not", baseSHA)
- if err := cmd.RunInDirPipeline(tmpBasePath, revListWriter, stderr); err != nil {
- log.Error("git rev-list [%s]: %v - %s", tmpBasePath, err, errbuf.String())
- errChan <- fmt.Errorf("git rev-list [%s]: %v - %s", tmpBasePath, err, errbuf.String())
- }
-}
-
-func readRevListObjects(revListReader *io.PipeReader, shasToCheckWriter *io.PipeWriter, wg *sync.WaitGroup) {
- defer wg.Done()
- defer revListReader.Close()
- defer shasToCheckWriter.Close()
- scanner := bufio.NewScanner(revListReader)
- for scanner.Scan() {
- line := scanner.Text()
- if len(line) == 0 {
- continue
- }
- fields := strings.Split(line, " ")
- if len(fields) < 2 || len(fields[1]) == 0 {
- continue
- }
- toWrite := []byte(fields[0] + "\n")
- for len(toWrite) > 0 {
- n, err := shasToCheckWriter.Write(toWrite)
- if err != nil {
- _ = revListReader.CloseWithError(err)
- break
- }
- toWrite = toWrite[n:]
- }
- }
- _ = shasToCheckWriter.CloseWithError(scanner.Err())
-}
-
-func doCatFileBatchCheck(shasToCheckReader *io.PipeReader, catFileCheckWriter *io.PipeWriter, wg *sync.WaitGroup, tmpBasePath string) {
- defer wg.Done()
- defer shasToCheckReader.Close()
- defer catFileCheckWriter.Close()
-
- stderr := new(bytes.Buffer)
- var errbuf strings.Builder
- cmd := git.NewCommand("cat-file", "--batch-check")
- if err := cmd.RunInDirFullPipeline(tmpBasePath, catFileCheckWriter, stderr, shasToCheckReader); err != nil {
- _ = catFileCheckWriter.CloseWithError(fmt.Errorf("git cat-file --batch-check [%s]: %v - %s", tmpBasePath, err, errbuf.String()))
- }
-}
-
-func readCatFileBatchCheck(catFileCheckReader *io.PipeReader, shasToBatchWriter *io.PipeWriter, wg *sync.WaitGroup) {
- defer wg.Done()
- defer catFileCheckReader.Close()
-
- scanner := bufio.NewScanner(catFileCheckReader)
- defer func() {
- _ = shasToBatchWriter.CloseWithError(scanner.Err())
- }()
- for scanner.Scan() {
- line := scanner.Text()
- if len(line) == 0 {
- continue
- }
- fields := strings.Split(line, " ")
- if len(fields) < 3 || fields[1] != "blob" {
- continue
- }
- size, _ := strconv.Atoi(fields[2])
- if size > 1024 {
- continue
- }
- toWrite := []byte(fields[0] + "\n")
- for len(toWrite) > 0 {
- n, err := shasToBatchWriter.Write(toWrite)
- if err != nil {
- _ = catFileCheckReader.CloseWithError(err)
- break
- }
- toWrite = toWrite[n:]
- }
- }
-}
-
-func doCatFileBatch(shasToBatchReader *io.PipeReader, catFileBatchWriter *io.PipeWriter, wg *sync.WaitGroup, tmpBasePath string) {
- defer wg.Done()
- defer shasToBatchReader.Close()
- defer catFileBatchWriter.Close()
-
- stderr := new(bytes.Buffer)
- var errbuf strings.Builder
- if err := git.NewCommand("cat-file", "--batch").RunInDirFullPipeline(tmpBasePath, catFileBatchWriter, stderr, shasToBatchReader); err != nil {
- _ = shasToBatchReader.CloseWithError(fmt.Errorf("git rev-list [%s]: %v - %s", tmpBasePath, err, errbuf.String()))
- }
-}
-
-func readCatFileBatch(catFileBatchReader *io.PipeReader, wg *sync.WaitGroup, pr *models.PullRequest) {
- defer wg.Done()
- defer catFileBatchReader.Close()
-
- bufferedReader := bufio.NewReader(catFileBatchReader)
- buf := make([]byte, 1025)
- for {
- // File descriptor line: sha
- _, err := bufferedReader.ReadString(' ')
- if err != nil {
- _ = catFileBatchReader.CloseWithError(err)
- break
- }
- // Throw away the blob
- if _, err := bufferedReader.ReadString(' '); err != nil {
- _ = catFileBatchReader.CloseWithError(err)
- break
- }
- sizeStr, err := bufferedReader.ReadString('\n')
- if err != nil {
- _ = catFileBatchReader.CloseWithError(err)
- break
- }
- size, err := strconv.Atoi(sizeStr[:len(sizeStr)-1])
- if err != nil {
- _ = catFileBatchReader.CloseWithError(err)
- break
- }
- pointerBuf := buf[:size+1]
- if _, err := io.ReadFull(bufferedReader, pointerBuf); err != nil {
- _ = catFileBatchReader.CloseWithError(err)
- break
- }
- pointerBuf = pointerBuf[:size]
- // Now we need to check if the pointerBuf is an LFS pointer
- pointer := lfs.IsPointerFile(&pointerBuf)
- if pointer == nil {
- continue
- }
- // Then we need to check that this pointer is in the db
- if _, err := pr.HeadRepo.GetLFSMetaObjectByOid(pointer.Oid); err != nil {
- if err == models.ErrLFSObjectNotExist {
- log.Warn("During merge of: %d in %-v, there is a pointer to LFS Oid: %s which although present in the LFS store is not associated with the head repo %-v", pr.Index, pr.BaseRepo, pointer.Oid, pr.HeadRepo)
- continue
- }
- _ = catFileBatchReader.CloseWithError(err)
- break
- }
- // OK we have a pointer that is associated with the head repo
- // and is actually a file in the LFS
- // Therefore it should be associated with the base repo
- pointer.RepositoryID = pr.BaseRepoID
- if _, err := models.NewLFSMetaObject(pointer); err != nil {
- _ = catFileBatchReader.CloseWithError(err)
- break
- }
- }
-}
diff --git a/modules/pull/merge.go b/modules/pull/merge.go
deleted file mode 100644
index 6150c1518e..0000000000
--- a/modules/pull/merge.go
+++ /dev/null
@@ -1,320 +0,0 @@
-// Copyright 2019 The Gitea Authors.
-// All rights reserved.
-// Use of this source code is governed by a MIT-style
-// license that can be found in the LICENSE file.
-
-package pull
-
-import (
- "bufio"
- "bytes"
- "fmt"
- "io/ioutil"
- "os"
- "path"
- "path/filepath"
- "strings"
-
- "code.gitea.io/gitea/models"
- "code.gitea.io/gitea/modules/cache"
- "code.gitea.io/gitea/modules/git"
- "code.gitea.io/gitea/modules/log"
- "code.gitea.io/gitea/modules/setting"
- api "code.gitea.io/gitea/modules/structs"
- "code.gitea.io/gitea/modules/timeutil"
-)
-
-// Merge merges pull request to base repository.
-// FIXME: add repoWorkingPull make sure two merges does not happen at same time.
-func Merge(pr *models.PullRequest, doer *models.User, baseGitRepo *git.Repository, mergeStyle models.MergeStyle, message string) (err error) {
- if err = pr.GetHeadRepo(); err != nil {
- return fmt.Errorf("GetHeadRepo: %v", err)
- } else if err = pr.GetBaseRepo(); err != nil {
- return fmt.Errorf("GetBaseRepo: %v", err)
- }
-
- prUnit, err := pr.BaseRepo.GetUnit(models.UnitTypePullRequests)
- if err != nil {
- return err
- }
- prConfig := prUnit.PullRequestsConfig()
-
- if err := pr.CheckUserAllowedToMerge(doer); err != nil {
- return fmt.Errorf("CheckUserAllowedToMerge: %v", err)
- }
-
- // Check if merge style is correct and allowed
- if !prConfig.IsMergeStyleAllowed(mergeStyle) {
- return models.ErrInvalidMergeStyle{ID: pr.BaseRepo.ID, Style: mergeStyle}
- }
-
- defer func() {
- go models.AddTestPullRequestTask(doer, pr.BaseRepo.ID, pr.BaseBranch, false)
- }()
-
- // Clone base repo.
- tmpBasePath, err := models.CreateTemporaryPath("merge")
- if err != nil {
- return err
- }
-
- defer func() {
- if err := models.RemoveTemporaryPath(tmpBasePath); err != nil {
- log.Error("Merge: RemoveTemporaryPath: %s", err)
- }
- }()
-
- headRepoPath := models.RepoPath(pr.HeadUserName, pr.HeadRepo.Name)
-
- if err := git.Clone(baseGitRepo.Path, tmpBasePath, git.CloneRepoOptions{
- Shared: true,
- NoCheckout: true,
- Branch: pr.BaseBranch,
- }); err != nil {
- return fmt.Errorf("git clone: %v", err)
- }
-
- remoteRepoName := "head_repo"
-
- // Add head repo remote.
- addCacheRepo := func(staging, cache string) error {
- p := filepath.Join(staging, ".git", "objects", "info", "alternates")
- f, err := os.OpenFile(p, os.O_APPEND|os.O_WRONLY, 0600)
- if err != nil {
- return err
- }
- defer f.Close()
- data := filepath.Join(cache, "objects")
- if _, err := fmt.Fprintln(f, data); err != nil {
- return err
- }
- return nil
- }
-
- if err := addCacheRepo(tmpBasePath, headRepoPath); err != nil {
- return fmt.Errorf("addCacheRepo [%s -> %s]: %v", headRepoPath, tmpBasePath, err)
- }
-
- var errbuf strings.Builder
- if err := git.NewCommand("remote", "add", remoteRepoName, headRepoPath).RunInDirPipeline(tmpBasePath, nil, &errbuf); err != nil {
- return fmt.Errorf("git remote add [%s -> %s]: %s", headRepoPath, tmpBasePath, errbuf.String())
- }
-
- // Fetch head branch
- if err := git.NewCommand("fetch", remoteRepoName, fmt.Sprintf("%s:refs/remotes/%s/%s", pr.HeadBranch, remoteRepoName, pr.HeadBranch)).RunInDirPipeline(tmpBasePath, nil, &errbuf); err != nil {
- return fmt.Errorf("git fetch [%s -> %s]: %s", headRepoPath, tmpBasePath, errbuf.String())
- }
-
- trackingBranch := path.Join(remoteRepoName, pr.HeadBranch)
- stagingBranch := fmt.Sprintf("%s_%s", remoteRepoName, pr.HeadBranch)
-
- // Enable sparse-checkout
- sparseCheckoutList, err := getDiffTree(tmpBasePath, pr.BaseBranch, trackingBranch)
- if err != nil {
- return fmt.Errorf("getDiffTree: %v", err)
- }
-
- infoPath := filepath.Join(tmpBasePath, ".git", "info")
- if err := os.MkdirAll(infoPath, 0700); err != nil {
- return fmt.Errorf("creating directory failed [%s]: %v", infoPath, err)
- }
- sparseCheckoutListPath := filepath.Join(infoPath, "sparse-checkout")
- if err := ioutil.WriteFile(sparseCheckoutListPath, []byte(sparseCheckoutList), 0600); err != nil {
- return fmt.Errorf("Writing sparse-checkout file to %s: %v", sparseCheckoutListPath, err)
- }
-
- // Switch off LFS process (set required, clean and smudge here also)
- if err := git.NewCommand("config", "--local", "filter.lfs.process", "").RunInDirPipeline(tmpBasePath, nil, &errbuf); err != nil {
- return fmt.Errorf("git config [filter.lfs.process -> <> ]: %v", errbuf.String())
- }
- if err := git.NewCommand("config", "--local", "filter.lfs.required", "false").RunInDirPipeline(tmpBasePath, nil, &errbuf); err != nil {
- return fmt.Errorf("git config [filter.lfs.required -> <false> ]: %v", errbuf.String())
- }
- if err := git.NewCommand("config", "--local", "filter.lfs.clean", "").RunInDirPipeline(tmpBasePath, nil, &errbuf); err != nil {
- return fmt.Errorf("git config [filter.lfs.clean -> <> ]: %v", errbuf.String())
- }
- if err := git.NewCommand("config", "--local", "filter.lfs.smudge", "").RunInDirPipeline(tmpBasePath, nil, &errbuf); err != nil {
- return fmt.Errorf("git config [filter.lfs.smudge -> <> ]: %v", errbuf.String())
- }
-
- if err := git.NewCommand("config", "--local", "core.sparseCheckout", "true").RunInDirPipeline(tmpBasePath, nil, &errbuf); err != nil {
- return fmt.Errorf("git config [core.sparsecheckout -> true]: %v", errbuf.String())
- }
-
- // Read base branch index
- if err := git.NewCommand("read-tree", "HEAD").RunInDirPipeline(tmpBasePath, nil, &errbuf); err != nil {
- return fmt.Errorf("git read-tree HEAD: %s", errbuf.String())
- }
-
- // Merge commits.
- switch mergeStyle {
- case models.MergeStyleMerge:
- if err := git.NewCommand("merge", "--no-ff", "--no-commit", trackingBranch).RunInDirPipeline(tmpBasePath, nil, &errbuf); err != nil {
- return fmt.Errorf("git merge --no-ff --no-commit [%s]: %v - %s", tmpBasePath, err, errbuf.String())
- }
-
- sig := doer.NewGitSig()
- if err := git.NewCommand("commit", fmt.Sprintf("--author='%s <%s>'", sig.Name, sig.Email), "-m", message).RunInDirPipeline(tmpBasePath, nil, &errbuf); err != nil {
- return fmt.Errorf("git commit [%s]: %v - %s", tmpBasePath, err, errbuf.String())
- }
- case models.MergeStyleRebase:
- // Checkout head branch
- if err := git.NewCommand("checkout", "-b", stagingBranch, trackingBranch).RunInDirPipeline(tmpBasePath, nil, &errbuf); err != nil {
- return fmt.Errorf("git checkout: %s", errbuf.String())
- }
- // Rebase before merging
- if err := git.NewCommand("rebase", "-q", pr.BaseBranch).RunInDirPipeline(tmpBasePath, nil, &errbuf); err != nil {
- return fmt.Errorf("git rebase [%s -> %s]: %s", headRepoPath, tmpBasePath, errbuf.String())
- }
- // Checkout base branch again
- if err := git.NewCommand("checkout", pr.BaseBranch).RunInDirPipeline(tmpBasePath, nil, &errbuf); err != nil {
- return fmt.Errorf("git checkout: %s", errbuf.String())
- }
- // Merge fast forward
- if err := git.NewCommand("merge", "--ff-only", "-q", stagingBranch).RunInDirPipeline(tmpBasePath, nil, &errbuf); err != nil {
- return fmt.Errorf("git merge --ff-only [%s -> %s]: %s", headRepoPath, tmpBasePath, errbuf.String())
- }
- case models.MergeStyleRebaseMerge:
- // Checkout head branch
- if err := git.NewCommand("checkout", "-b", stagingBranch, trackingBranch).RunInDirPipeline(tmpBasePath, nil, &errbuf); err != nil {
- return fmt.Errorf("git checkout: %s", errbuf.String())
- }
- // Rebase before merging
- if err := git.NewCommand("rebase", "-q", pr.BaseBranch).RunInDirPipeline(tmpBasePath, nil, &errbuf); err != nil {
- return fmt.Errorf("git rebase [%s -> %s]: %s", headRepoPath, tmpBasePath, errbuf.String())
- }
- // Checkout base branch again
- if err := git.NewCommand("checkout", pr.BaseBranch).RunInDirPipeline(tmpBasePath, nil, &errbuf); err != nil {
- return fmt.Errorf("git checkout: %s", errbuf.String())
- }
- // Prepare merge with commit
- if err := git.NewCommand("merge", "--no-ff", "--no-commit", "-q", stagingBranch).RunInDirPipeline(tmpBasePath, nil, &errbuf); err != nil {
- return fmt.Errorf("git merge --no-ff [%s -> %s]: %s", headRepoPath, tmpBasePath, errbuf.String())
- }
-
- // Set custom message and author and create merge commit
- sig := doer.NewGitSig()
- if err := git.NewCommand("commit", fmt.Sprintf("--author='%s <%s>'", sig.Name, sig.Email), "-m", message).RunInDirPipeline(tmpBasePath, nil, &errbuf); err != nil {
- return fmt.Errorf("git commit [%s]: %v - %s", tmpBasePath, err, errbuf.String())
- }
-
- case models.MergeStyleSquash:
- // Merge with squash
- if err := git.NewCommand("merge", "-q", "--squash", trackingBranch).RunInDirPipeline(tmpBasePath, nil, &errbuf); err != nil {
- return fmt.Errorf("git merge --squash [%s -> %s]: %s", headRepoPath, tmpBasePath, errbuf.String())
- }
- sig := pr.Issue.Poster.NewGitSig()
- if err := git.NewCommand("commit", fmt.Sprintf("--author='%s <%s>'", sig.Name, sig.Email), "-m", message).RunInDirPipeline(tmpBasePath, nil, &errbuf); err != nil {
- return fmt.Errorf("git commit [%s]: %v - %s", tmpBasePath, err, errbuf.String())
- }
- default:
- return models.ErrInvalidMergeStyle{ID: pr.BaseRepo.ID, Style: mergeStyle}
- }
-
- // OK we should cache our current head and origin/headbranch
- mergeHeadSHA, err := git.GetFullCommitID(tmpBasePath, "HEAD")
- if err != nil {
- return fmt.Errorf("Failed to get full commit id for HEAD: %v", err)
- }
- mergeBaseSHA, err := git.GetFullCommitID(tmpBasePath, "origin/"+pr.BaseBranch)
- if err != nil {
- return fmt.Errorf("Failed to get full commit id for origin/%s: %v", pr.BaseBranch, err)
- }
-
- // Now it's questionable about where this should go - either after or before the push
- // I think in the interests of data safety - failures to push to the lfs should prevent
- // the merge as you can always remerge.
- if setting.LFS.StartServer {
- if err := LFSPush(tmpBasePath, mergeHeadSHA, mergeBaseSHA, pr); err != nil {
- return err
- }
- }
-
- headUser, err := models.GetUserByName(pr.HeadUserName)
- if err != nil {
- if !models.IsErrUserNotExist(err) {
- log.Error("Can't find user: %s for head repository - %v", pr.HeadUserName, err)
- return err
- }
- log.Error("Can't find user: %s for head repository - defaulting to doer: %s - %v", pr.HeadUserName, doer.Name, err)
- headUser = doer
- }
-
- env := models.FullPushingEnvironment(
- headUser,
- doer,
- pr.BaseRepo,
- pr.BaseRepo.Name,
- pr.ID,
- )
-
- // Push back to upstream.
- if err := git.NewCommand("push", "origin", pr.BaseBranch).RunInDirTimeoutEnvPipeline(env, -1, tmpBasePath, nil, &errbuf); err != nil {
- return fmt.Errorf("git push: %s", errbuf.String())
- }
-
- pr.MergedCommitID, err = baseGitRepo.GetBranchCommitID(pr.BaseBranch)
- if err != nil {
- return fmt.Errorf("GetBranchCommit: %v", err)
- }
-
- pr.MergedUnix = timeutil.TimeStampNow()
- pr.Merger = doer
- pr.MergerID = doer.ID
-
- if err = pr.SetMerged(); err != nil {
- log.Error("setMerged [%d]: %v", pr.ID, err)
- }
-
- if err = models.MergePullRequestAction(doer, pr.Issue.Repo, pr.Issue); err != nil {
- log.Error("MergePullRequestAction [%d]: %v", pr.ID, err)
- }
-
- // Reset cached commit count
- cache.Remove(pr.Issue.Repo.GetCommitsCountCacheKey(pr.BaseBranch, true))
-
- // Reload pull request information.
- if err = pr.LoadAttributes(); err != nil {
- log.Error("LoadAttributes: %v", err)
- return nil
- }
-
- mode, _ := models.AccessLevel(doer, pr.Issue.Repo)
- if err = models.PrepareWebhooks(pr.Issue.Repo, models.HookEventPullRequest, &api.PullRequestPayload{
- Action: api.HookIssueClosed,
- Index: pr.Index,
- PullRequest: pr.APIFormat(),
- Repository: pr.Issue.Repo.APIFormat(mode),
- Sender: doer.APIFormat(),
- }); err != nil {
- log.Error("PrepareWebhooks: %v", err)
- } else {
- go models.HookQueue.Add(pr.Issue.Repo.ID)
- }
-
- return nil
-}
-
-func getDiffTree(repoPath, baseBranch, headBranch string) (string, error) {
- getDiffTreeFromBranch := func(repoPath, baseBranch, headBranch string) (string, error) {
- var outbuf, errbuf strings.Builder
- // Compute the diff-tree for sparse-checkout
- if err := git.NewCommand("diff-tree", "--no-commit-id", "--name-only", "-r", "--root", baseBranch, headBranch, "--").RunInDirPipeline(repoPath, &outbuf, &errbuf); err != nil {
- return "", fmt.Errorf("git diff-tree [%s base:%s head:%s]: %s", repoPath, baseBranch, headBranch, errbuf.String())
- }
- return outbuf.String(), nil
- }
-
- list, err := getDiffTreeFromBranch(repoPath, baseBranch, headBranch)
- if err != nil {
- return "", err
- }
-
- // Prefixing '/' for each entry, otherwise all files with the same name in subdirectories would be matched.
- out := bytes.Buffer{}
- scanner := bufio.NewScanner(strings.NewReader(list))
- for scanner.Scan() {
- fmt.Fprintf(&out, "/%s\n", scanner.Text())
- }
- return out.String(), nil
-}
diff --git a/modules/pull/review.go b/modules/pull/review.go
deleted file mode 100644
index 3fdfaaff84..0000000000
--- a/modules/pull/review.go
+++ /dev/null
@@ -1,57 +0,0 @@
-// Copyright 2019 The Gitea Authors.
-// All rights reserved.
-// Use of this source code is governed by a MIT-style
-// license that can be found in the LICENSE file.
-
-package pull
-
-import (
- "code.gitea.io/gitea/models"
- api "code.gitea.io/gitea/modules/structs"
-)
-
-// CreateReview creates a new review based on opts
-func CreateReview(opts models.CreateReviewOptions) (*models.Review, error) {
- review, err := models.CreateReview(opts)
- if err != nil {
- return nil, err
- }
-
- var reviewHookType models.HookEventType
-
- switch opts.Type {
- case models.ReviewTypeApprove:
- reviewHookType = models.HookEventPullRequestApproved
- case models.ReviewTypeComment:
- reviewHookType = models.HookEventPullRequestComment
- case models.ReviewTypeReject:
- reviewHookType = models.HookEventPullRequestRejected
- default:
- // unsupported review webhook type here
- return review, nil
- }
-
- pr := opts.Issue.PullRequest
-
- if err := pr.LoadIssue(); err != nil {
- return nil, err
- }
-
- mode, err := models.AccessLevel(opts.Issue.Poster, opts.Issue.Repo)
- if err != nil {
- return nil, err
- }
-
- if err := models.PrepareWebhooks(opts.Issue.Repo, reviewHookType, &api.PullRequestPayload{
- Action: api.HookIssueSynchronized,
- Index: opts.Issue.Index,
- PullRequest: pr.APIFormat(),
- Repository: opts.Issue.Repo.APIFormat(mode),
- Sender: opts.Reviewer.APIFormat(),
- }); err != nil {
- return nil, err
- }
- go models.HookQueue.Add(opts.Issue.Repo.ID)
-
- return review, nil
-}