summaryrefslogtreecommitdiffstats
path: root/modules/migrations/gitea_uploader.go
diff options
context:
space:
mode:
author6543 <6543@obermui.de>2020-10-14 06:06:00 +0200
committerGitHub <noreply@github.com>2020-10-14 07:06:00 +0300
commit49b1948cb11f46b8dabc6c977079ca9a187c92e0 (patch)
tree4c582fd4a53ce6ee3d24b1ef902b3f242187eb50 /modules/migrations/gitea_uploader.go
parentdfa7291f8fafd2aac032ef73a78b588e4d0e9a36 (diff)
downloadgitea-49b1948cb11f46b8dabc6c977079ca9a187c92e0.tar.gz
gitea-49b1948cb11f46b8dabc6c977079ca9a187c92e0.zip
Gitea 2 Gitea migration (#12657)
* first draft * update gitea sdk to 9e280adb4da * adapt feat of updated sdk * releases now works * break the Reactions loop * use convertGiteaLabel * fix endless loop because paggination is not supported there !!! * rename gitea local uploader files * pagination can bite you in the ass * Version Checks * lint * docs * rename gitea sdk import to miss future conficts * go-swagger: dont scan the sdk structs * make sure gitea can shutdown gracefully * make GetPullRequests and GetIssues similar * rm useles * Add Test: started ... * ... add tests ... * Add tests and Fixing things * Workaround missing SHA * Adapt: Ensure that all migration requests are cancellable (714ab71ddc4260937b1480519d453d2dc4e77dd6) * LINT: fix misspells in test set * adapt ListMergeRequestAwardEmoji * update sdk * Return error when creating giteadownloader failed * update sdk * adapt new sdk * adopt new features * check version before err * adapt: 'migrate service type switch page' * optimize * Fix DefaultBranch * impruve * handle subPath * fix test * Fix ReviewCommentPosition * test GetReviews * add DefaultBranch int test set * rm unused * Update SDK to v0.13.0 * addopt sdk changes * found better link * format template * Update Docs * Update Gitea SDK (v0.13.1)
Diffstat (limited to 'modules/migrations/gitea_uploader.go')
-rw-r--r--modules/migrations/gitea_uploader.go849
1 files changed, 849 insertions, 0 deletions
diff --git a/modules/migrations/gitea_uploader.go b/modules/migrations/gitea_uploader.go
new file mode 100644
index 0000000000..cd1fd5cb8d
--- /dev/null
+++ b/modules/migrations/gitea_uploader.go
@@ -0,0 +1,849 @@
+// Copyright 2019 The Gitea Authors. All rights reserved.
+// Copyright 2018 Jonas Franz. All rights reserved.
+// Use of this source code is governed by a MIT-style
+// license that can be found in the LICENSE file.
+
+package migrations
+
+import (
+ "bytes"
+ "context"
+ "fmt"
+ "io"
+ "net/http"
+ "net/url"
+ "os"
+ "path/filepath"
+ "strings"
+ "sync"
+ "time"
+
+ "code.gitea.io/gitea/models"
+ "code.gitea.io/gitea/modules/git"
+ "code.gitea.io/gitea/modules/log"
+ "code.gitea.io/gitea/modules/migrations/base"
+ "code.gitea.io/gitea/modules/repository"
+ repo_module "code.gitea.io/gitea/modules/repository"
+ "code.gitea.io/gitea/modules/setting"
+ "code.gitea.io/gitea/modules/storage"
+ "code.gitea.io/gitea/modules/structs"
+ "code.gitea.io/gitea/modules/timeutil"
+
+ gouuid "github.com/google/uuid"
+)
+
+var (
+ _ base.Uploader = &GiteaLocalUploader{}
+)
+
+// GiteaLocalUploader implements an Uploader to gitea sites
+type GiteaLocalUploader struct {
+ ctx context.Context
+ doer *models.User
+ repoOwner string
+ repoName string
+ repo *models.Repository
+ labels sync.Map
+ milestones sync.Map
+ issues sync.Map
+ gitRepo *git.Repository
+ prHeadCache map[string]struct{}
+ userMap map[int64]int64 // external user id mapping to user id
+ prCache map[int64]*models.PullRequest
+ gitServiceType structs.GitServiceType
+}
+
+// NewGiteaLocalUploader creates an gitea Uploader via gitea API v1
+func NewGiteaLocalUploader(ctx context.Context, doer *models.User, repoOwner, repoName string) *GiteaLocalUploader {
+ return &GiteaLocalUploader{
+ ctx: ctx,
+ doer: doer,
+ repoOwner: repoOwner,
+ repoName: repoName,
+ prHeadCache: make(map[string]struct{}),
+ userMap: make(map[int64]int64),
+ prCache: make(map[int64]*models.PullRequest),
+ }
+}
+
+// MaxBatchInsertSize returns the table's max batch insert size
+func (g *GiteaLocalUploader) MaxBatchInsertSize(tp string) int {
+ switch tp {
+ case "issue":
+ return models.MaxBatchInsertSize(new(models.Issue))
+ case "comment":
+ return models.MaxBatchInsertSize(new(models.Comment))
+ case "milestone":
+ return models.MaxBatchInsertSize(new(models.Milestone))
+ case "label":
+ return models.MaxBatchInsertSize(new(models.Label))
+ case "release":
+ return models.MaxBatchInsertSize(new(models.Release))
+ case "pullrequest":
+ return models.MaxBatchInsertSize(new(models.PullRequest))
+ }
+ return 10
+}
+
+// CreateRepo creates a repository
+func (g *GiteaLocalUploader) CreateRepo(repo *base.Repository, opts base.MigrateOptions) error {
+ owner, err := models.GetUserByName(g.repoOwner)
+ if err != nil {
+ return err
+ }
+
+ var remoteAddr = repo.CloneURL
+ if len(opts.AuthToken) > 0 || len(opts.AuthUsername) > 0 {
+ u, err := url.Parse(repo.CloneURL)
+ if err != nil {
+ return err
+ }
+ u.User = url.UserPassword(opts.AuthUsername, opts.AuthPassword)
+ if len(opts.AuthToken) > 0 {
+ u.User = url.UserPassword("oauth2", opts.AuthToken)
+ }
+ remoteAddr = u.String()
+ }
+
+ var r *models.Repository
+ if opts.MigrateToRepoID <= 0 {
+ r, err = repo_module.CreateRepository(g.doer, owner, models.CreateRepoOptions{
+ Name: g.repoName,
+ Description: repo.Description,
+ OriginalURL: repo.OriginalURL,
+ GitServiceType: opts.GitServiceType,
+ IsPrivate: opts.Private,
+ IsMirror: opts.Mirror,
+ Status: models.RepositoryBeingMigrated,
+ })
+ } else {
+ r, err = models.GetRepositoryByID(opts.MigrateToRepoID)
+ }
+ if err != nil {
+ return err
+ }
+ r.DefaultBranch = repo.DefaultBranch
+
+ r, err = repository.MigrateRepositoryGitData(g.doer, owner, r, base.MigrateOptions{
+ RepoName: g.repoName,
+ Description: repo.Description,
+ OriginalURL: repo.OriginalURL,
+ GitServiceType: opts.GitServiceType,
+ Mirror: repo.IsMirror,
+ CloneAddr: remoteAddr,
+ Private: repo.IsPrivate,
+ Wiki: opts.Wiki,
+ Releases: opts.Releases, // if didn't get releases, then sync them from tags
+ })
+
+ g.repo = r
+ if err != nil {
+ return err
+ }
+ g.gitRepo, err = git.OpenRepository(r.RepoPath())
+ return err
+}
+
+// Close closes this uploader
+func (g *GiteaLocalUploader) Close() {
+ if g.gitRepo != nil {
+ g.gitRepo.Close()
+ }
+}
+
+// CreateTopics creates topics
+func (g *GiteaLocalUploader) CreateTopics(topics ...string) error {
+ return models.SaveTopics(g.repo.ID, topics...)
+}
+
+// CreateMilestones creates milestones
+func (g *GiteaLocalUploader) CreateMilestones(milestones ...*base.Milestone) error {
+ var mss = make([]*models.Milestone, 0, len(milestones))
+ for _, milestone := range milestones {
+ var deadline timeutil.TimeStamp
+ if milestone.Deadline != nil {
+ deadline = timeutil.TimeStamp(milestone.Deadline.Unix())
+ }
+ if deadline == 0 {
+ deadline = timeutil.TimeStamp(time.Date(9999, 1, 1, 0, 0, 0, 0, setting.DefaultUILocation).Unix())
+ }
+ var ms = models.Milestone{
+ RepoID: g.repo.ID,
+ Name: milestone.Title,
+ Content: milestone.Description,
+ IsClosed: milestone.State == "closed",
+ DeadlineUnix: deadline,
+ }
+ if ms.IsClosed && milestone.Closed != nil {
+ ms.ClosedDateUnix = timeutil.TimeStamp(milestone.Closed.Unix())
+ }
+ mss = append(mss, &ms)
+ }
+
+ err := models.InsertMilestones(mss...)
+ if err != nil {
+ return err
+ }
+
+ for _, ms := range mss {
+ g.milestones.Store(ms.Name, ms.ID)
+ }
+ return nil
+}
+
+// CreateLabels creates labels
+func (g *GiteaLocalUploader) CreateLabels(labels ...*base.Label) error {
+ var lbs = make([]*models.Label, 0, len(labels))
+ for _, label := range labels {
+ lbs = append(lbs, &models.Label{
+ RepoID: g.repo.ID,
+ Name: label.Name,
+ Description: label.Description,
+ Color: fmt.Sprintf("#%s", label.Color),
+ })
+ }
+
+ err := models.NewLabels(lbs...)
+ if err != nil {
+ return err
+ }
+ for _, lb := range lbs {
+ g.labels.Store(lb.Name, lb)
+ }
+ return nil
+}
+
+// CreateReleases creates releases
+func (g *GiteaLocalUploader) CreateReleases(downloader base.Downloader, releases ...*base.Release) error {
+ var rels = make([]*models.Release, 0, len(releases))
+ for _, release := range releases {
+ var rel = models.Release{
+ RepoID: g.repo.ID,
+ TagName: release.TagName,
+ LowerTagName: strings.ToLower(release.TagName),
+ Target: release.TargetCommitish,
+ Title: release.Name,
+ Sha1: release.TargetCommitish,
+ Note: release.Body,
+ IsDraft: release.Draft,
+ IsPrerelease: release.Prerelease,
+ IsTag: false,
+ CreatedUnix: timeutil.TimeStamp(release.Created.Unix()),
+ }
+
+ userid, ok := g.userMap[release.PublisherID]
+ tp := g.gitServiceType.Name()
+ if !ok && tp != "" {
+ var err error
+ userid, err = models.GetUserIDByExternalUserID(tp, fmt.Sprintf("%v", release.PublisherID))
+ if err != nil {
+ log.Error("GetUserIDByExternalUserID: %v", err)
+ }
+ if userid > 0 {
+ g.userMap[release.PublisherID] = userid
+ }
+ }
+
+ if userid > 0 {
+ rel.PublisherID = userid
+ } else {
+ rel.PublisherID = g.doer.ID
+ rel.OriginalAuthor = release.PublisherName
+ rel.OriginalAuthorID = release.PublisherID
+ }
+
+ // calc NumCommits
+ commit, err := g.gitRepo.GetCommit(rel.TagName)
+ if err != nil {
+ return fmt.Errorf("GetCommit: %v", err)
+ }
+ rel.NumCommits, err = commit.CommitsCount()
+ if err != nil {
+ return fmt.Errorf("CommitsCount: %v", err)
+ }
+
+ for _, asset := range release.Assets {
+ var attach = models.Attachment{
+ UUID: gouuid.New().String(),
+ Name: asset.Name,
+ DownloadCount: int64(*asset.DownloadCount),
+ Size: int64(*asset.Size),
+ CreatedUnix: timeutil.TimeStamp(asset.Created.Unix()),
+ }
+
+ // download attachment
+ err = func() error {
+ var rc io.ReadCloser
+ if asset.DownloadURL == nil {
+ rc, err = downloader.GetAsset(rel.TagName, rel.ID, asset.ID)
+ if err != nil {
+ return err
+ }
+ } else {
+ resp, err := http.Get(*asset.DownloadURL)
+ if err != nil {
+ return err
+ }
+ rc = resp.Body
+ }
+ _, err = storage.Attachments.Save(attach.RelativePath(), rc)
+ return err
+ }()
+ if err != nil {
+ return err
+ }
+ rel.Attachments = append(rel.Attachments, &attach)
+ }
+
+ rels = append(rels, &rel)
+ }
+
+ return models.InsertReleases(rels...)
+}
+
+// SyncTags syncs releases with tags in the database
+func (g *GiteaLocalUploader) SyncTags() error {
+ return repository.SyncReleasesWithTags(g.repo, g.gitRepo)
+}
+
+// CreateIssues creates issues
+func (g *GiteaLocalUploader) CreateIssues(issues ...*base.Issue) error {
+ var iss = make([]*models.Issue, 0, len(issues))
+ for _, issue := range issues {
+ var labels []*models.Label
+ for _, label := range issue.Labels {
+ lb, ok := g.labels.Load(label.Name)
+ if ok {
+ labels = append(labels, lb.(*models.Label))
+ }
+ }
+
+ var milestoneID int64
+ if issue.Milestone != "" {
+ milestone, ok := g.milestones.Load(issue.Milestone)
+ if ok {
+ milestoneID = milestone.(int64)
+ }
+ }
+
+ var is = models.Issue{
+ RepoID: g.repo.ID,
+ Repo: g.repo,
+ Index: issue.Number,
+ Title: issue.Title,
+ Content: issue.Content,
+ IsClosed: issue.State == "closed",
+ IsLocked: issue.IsLocked,
+ MilestoneID: milestoneID,
+ Labels: labels,
+ CreatedUnix: timeutil.TimeStamp(issue.Created.Unix()),
+ UpdatedUnix: timeutil.TimeStamp(issue.Updated.Unix()),
+ }
+
+ userid, ok := g.userMap[issue.PosterID]
+ tp := g.gitServiceType.Name()
+ if !ok && tp != "" {
+ var err error
+ userid, err = models.GetUserIDByExternalUserID(tp, fmt.Sprintf("%v", issue.PosterID))
+ if err != nil {
+ log.Error("GetUserIDByExternalUserID: %v", err)
+ }
+ if userid > 0 {
+ g.userMap[issue.PosterID] = userid
+ }
+ }
+
+ if userid > 0 {
+ is.PosterID = userid
+ } else {
+ is.PosterID = g.doer.ID
+ is.OriginalAuthor = issue.PosterName
+ is.OriginalAuthorID = issue.PosterID
+ }
+
+ if issue.Closed != nil {
+ is.ClosedUnix = timeutil.TimeStamp(issue.Closed.Unix())
+ }
+ // add reactions
+ for _, reaction := range issue.Reactions {
+ userid, ok := g.userMap[reaction.UserID]
+ if !ok && tp != "" {
+ var err error
+ userid, err = models.GetUserIDByExternalUserID(tp, fmt.Sprintf("%v", reaction.UserID))
+ if err != nil {
+ log.Error("GetUserIDByExternalUserID: %v", err)
+ }
+ if userid > 0 {
+ g.userMap[reaction.UserID] = userid
+ }
+ }
+ var res = models.Reaction{
+ Type: reaction.Content,
+ CreatedUnix: timeutil.TimeStampNow(),
+ }
+ if userid > 0 {
+ res.UserID = userid
+ } else {
+ res.UserID = g.doer.ID
+ res.OriginalAuthorID = reaction.UserID
+ res.OriginalAuthor = reaction.UserName
+ }
+ is.Reactions = append(is.Reactions, &res)
+ }
+ iss = append(iss, &is)
+ }
+
+ if len(iss) > 0 {
+ if err := models.InsertIssues(iss...); err != nil {
+ return err
+ }
+
+ for _, is := range iss {
+ g.issues.Store(is.Index, is.ID)
+ }
+ }
+
+ return nil
+}
+
+// CreateComments creates comments of issues
+func (g *GiteaLocalUploader) CreateComments(comments ...*base.Comment) error {
+ var cms = make([]*models.Comment, 0, len(comments))
+ for _, comment := range comments {
+ var issueID int64
+ if issueIDStr, ok := g.issues.Load(comment.IssueIndex); !ok {
+ issue, err := models.GetIssueByIndex(g.repo.ID, comment.IssueIndex)
+ if err != nil {
+ return err
+ }
+ issueID = issue.ID
+ g.issues.Store(comment.IssueIndex, issueID)
+ } else {
+ issueID = issueIDStr.(int64)
+ }
+
+ userid, ok := g.userMap[comment.PosterID]
+ tp := g.gitServiceType.Name()
+ if !ok && tp != "" {
+ var err error
+ userid, err = models.GetUserIDByExternalUserID(tp, fmt.Sprintf("%v", comment.PosterID))
+ if err != nil {
+ log.Error("GetUserIDByExternalUserID: %v", err)
+ }
+ if userid > 0 {
+ g.userMap[comment.PosterID] = userid
+ }
+ }
+
+ cm := models.Comment{
+ IssueID: issueID,
+ Type: models.CommentTypeComment,
+ Content: comment.Content,
+ CreatedUnix: timeutil.TimeStamp(comment.Created.Unix()),
+ UpdatedUnix: timeutil.TimeStamp(comment.Updated.Unix()),
+ }
+
+ if userid > 0 {
+ cm.PosterID = userid
+ } else {
+ cm.PosterID = g.doer.ID
+ cm.OriginalAuthor = comment.PosterName
+ cm.OriginalAuthorID = comment.PosterID
+ }
+
+ // add reactions
+ for _, reaction := range comment.Reactions {
+ userid, ok := g.userMap[reaction.UserID]
+ if !ok && tp != "" {
+ var err error
+ userid, err = models.GetUserIDByExternalUserID(tp, fmt.Sprintf("%v", reaction.UserID))
+ if err != nil {
+ log.Error("GetUserIDByExternalUserID: %v", err)
+ }
+ if userid > 0 {
+ g.userMap[reaction.UserID] = userid
+ }
+ }
+ var res = models.Reaction{
+ Type: reaction.Content,
+ CreatedUnix: timeutil.TimeStampNow(),
+ }
+ if userid > 0 {
+ res.UserID = userid
+ } else {
+ res.UserID = g.doer.ID
+ res.OriginalAuthorID = reaction.UserID
+ res.OriginalAuthor = reaction.UserName
+ }
+ cm.Reactions = append(cm.Reactions, &res)
+ }
+
+ cms = append(cms, &cm)
+ }
+
+ if len(cms) == 0 {
+ return nil
+ }
+ return models.InsertIssueComments(cms)
+}
+
+// CreatePullRequests creates pull requests
+func (g *GiteaLocalUploader) CreatePullRequests(prs ...*base.PullRequest) error {
+ var gprs = make([]*models.PullRequest, 0, len(prs))
+ for _, pr := range prs {
+ gpr, err := g.newPullRequest(pr)
+ if err != nil {
+ return err
+ }
+
+ userid, ok := g.userMap[pr.PosterID]
+ tp := g.gitServiceType.Name()
+ if !ok && tp != "" {
+ var err error
+ userid, err = models.GetUserIDByExternalUserID(tp, fmt.Sprintf("%v", pr.PosterID))
+ if err != nil {
+ log.Error("GetUserIDByExternalUserID: %v", err)
+ }
+ if userid > 0 {
+ g.userMap[pr.PosterID] = userid
+ }
+ }
+
+ if userid > 0 {
+ gpr.Issue.PosterID = userid
+ } else {
+ gpr.Issue.PosterID = g.doer.ID
+ gpr.Issue.OriginalAuthor = pr.PosterName
+ gpr.Issue.OriginalAuthorID = pr.PosterID
+ }
+
+ gprs = append(gprs, gpr)
+ }
+ if err := models.InsertPullRequests(gprs...); err != nil {
+ return err
+ }
+ for _, pr := range gprs {
+ g.issues.Store(pr.Issue.Index, pr.Issue.ID)
+ }
+ return nil
+}
+
+func (g *GiteaLocalUploader) newPullRequest(pr *base.PullRequest) (*models.PullRequest, error) {
+ var labels []*models.Label
+ for _, label := range pr.Labels {
+ lb, ok := g.labels.Load(label.Name)
+ if ok {
+ labels = append(labels, lb.(*models.Label))
+ }
+ }
+
+ var milestoneID int64
+ if pr.Milestone != "" {
+ milestone, ok := g.milestones.Load(pr.Milestone)
+ if ok {
+ milestoneID = milestone.(int64)
+ }
+ }
+
+ // download patch file
+ err := func() error {
+ resp, err := http.Get(pr.PatchURL)
+ if err != nil {
+ return err
+ }
+ defer resp.Body.Close()
+ pullDir := filepath.Join(g.repo.RepoPath(), "pulls")
+ if err = os.MkdirAll(pullDir, os.ModePerm); err != nil {
+ return err
+ }
+ f, err := os.Create(filepath.Join(pullDir, fmt.Sprintf("%d.patch", pr.Number)))
+ if err != nil {
+ return err
+ }
+ defer f.Close()
+ _, err = io.Copy(f, resp.Body)
+ return err
+ }()
+ if err != nil {
+ return nil, err
+ }
+
+ // set head information
+ pullHead := filepath.Join(g.repo.RepoPath(), "refs", "pull", fmt.Sprintf("%d", pr.Number))
+ if err := os.MkdirAll(pullHead, os.ModePerm); err != nil {
+ return nil, err
+ }
+ p, err := os.Create(filepath.Join(pullHead, "head"))
+ if err != nil {
+ return nil, err
+ }
+ _, err = p.WriteString(pr.Head.SHA)
+ p.Close()
+ if err != nil {
+ return nil, err
+ }
+
+ var head = "unknown repository"
+ if pr.IsForkPullRequest() && pr.State != "closed" {
+ if pr.Head.OwnerName != "" {
+ remote := pr.Head.OwnerName
+ _, ok := g.prHeadCache[remote]
+ if !ok {
+ // git remote add
+ err := g.gitRepo.AddRemote(remote, pr.Head.CloneURL, true)
+ if err != nil {
+ log.Error("AddRemote failed: %s", err)
+ } else {
+ g.prHeadCache[remote] = struct{}{}
+ ok = true
+ }
+ }
+
+ if ok {
+ _, err = git.NewCommand("fetch", remote, pr.Head.Ref).RunInDir(g.repo.RepoPath())
+ if err != nil {
+ log.Error("Fetch branch from %s failed: %v", pr.Head.CloneURL, err)
+ } else {
+ headBranch := filepath.Join(g.repo.RepoPath(), "refs", "heads", pr.Head.OwnerName, pr.Head.Ref)
+ if err := os.MkdirAll(filepath.Dir(headBranch), os.ModePerm); err != nil {
+ return nil, err
+ }
+ b, err := os.Create(headBranch)
+ if err != nil {
+ return nil, err
+ }
+ _, err = b.WriteString(pr.Head.SHA)
+ b.Close()
+ if err != nil {
+ return nil, err
+ }
+ head = pr.Head.OwnerName + "/" + pr.Head.Ref
+ }
+ }
+ }
+ } else {
+ head = pr.Head.Ref
+ }
+
+ var issue = models.Issue{
+ RepoID: g.repo.ID,
+ Repo: g.repo,
+ Title: pr.Title,
+ Index: pr.Number,
+ Content: pr.Content,
+ MilestoneID: milestoneID,
+ IsPull: true,
+ IsClosed: pr.State == "closed",
+ IsLocked: pr.IsLocked,
+ Labels: labels,
+ CreatedUnix: timeutil.TimeStamp(pr.Created.Unix()),
+ UpdatedUnix: timeutil.TimeStamp(pr.Updated.Unix()),
+ }
+
+ tp := g.gitServiceType.Name()
+
+ userid, ok := g.userMap[pr.PosterID]
+ if !ok && tp != "" {
+ var err error
+ userid, err = models.GetUserIDByExternalUserID(tp, fmt.Sprintf("%v", pr.PosterID))
+ if err != nil {
+ log.Error("GetUserIDByExternalUserID: %v", err)
+ }
+ if userid > 0 {
+ g.userMap[pr.PosterID] = userid
+ }
+ }
+
+ if userid > 0 {
+ issue.PosterID = userid
+ } else {
+ issue.PosterID = g.doer.ID
+ issue.OriginalAuthor = pr.PosterName
+ issue.OriginalAuthorID = pr.PosterID
+ }
+
+ // add reactions
+ for _, reaction := range pr.Reactions {
+ userid, ok := g.userMap[reaction.UserID]
+ if !ok && tp != "" {
+ var err error
+ userid, err = models.GetUserIDByExternalUserID(tp, fmt.Sprintf("%v", reaction.UserID))
+ if err != nil {
+ log.Error("GetUserIDByExternalUserID: %v", err)
+ }
+ if userid > 0 {
+ g.userMap[reaction.UserID] = userid
+ }
+ }
+ var res = models.Reaction{
+ Type: reaction.Content,
+ CreatedUnix: timeutil.TimeStampNow(),
+ }
+ if userid > 0 {
+ res.UserID = userid
+ } else {
+ res.UserID = g.doer.ID
+ res.OriginalAuthorID = reaction.UserID
+ res.OriginalAuthor = reaction.UserName
+ }
+ issue.Reactions = append(issue.Reactions, &res)
+ }
+
+ var pullRequest = models.PullRequest{
+ HeadRepoID: g.repo.ID,
+ HeadBranch: head,
+ BaseRepoID: g.repo.ID,
+ BaseBranch: pr.Base.Ref,
+ MergeBase: pr.Base.SHA,
+ Index: pr.Number,
+ HasMerged: pr.Merged,
+
+ Issue: &issue,
+ }
+
+ if pullRequest.Issue.IsClosed && pr.Closed != nil {
+ pullRequest.Issue.ClosedUnix = timeutil.TimeStamp(pr.Closed.Unix())
+ }
+ if pullRequest.HasMerged && pr.MergedTime != nil {
+ pullRequest.MergedUnix = timeutil.TimeStamp(pr.MergedTime.Unix())
+ pullRequest.MergedCommitID = pr.MergeCommitSHA
+ pullRequest.MergerID = g.doer.ID
+ }
+
+ // TODO: assignees
+
+ return &pullRequest, nil
+}
+
+func convertReviewState(state string) models.ReviewType {
+ switch state {
+ case base.ReviewStatePending:
+ return models.ReviewTypePending
+ case base.ReviewStateApproved:
+ return models.ReviewTypeApprove
+ case base.ReviewStateChangesRequested:
+ return models.ReviewTypeReject
+ case base.ReviewStateCommented:
+ return models.ReviewTypeComment
+ default:
+ return models.ReviewTypePending
+ }
+}
+
+// CreateReviews create pull request reviews
+func (g *GiteaLocalUploader) CreateReviews(reviews ...*base.Review) error {
+ var cms = make([]*models.Review, 0, len(reviews))
+ for _, review := range reviews {
+ var issueID int64
+ if issueIDStr, ok := g.issues.Load(review.IssueIndex); !ok {
+ issue, err := models.GetIssueByIndex(g.repo.ID, review.IssueIndex)
+ if err != nil {
+ return err
+ }
+ issueID = issue.ID
+ g.issues.Store(review.IssueIndex, issueID)
+ } else {
+ issueID = issueIDStr.(int64)
+ }
+
+ userid, ok := g.userMap[review.ReviewerID]
+ tp := g.gitServiceType.Name()
+ if !ok && tp != "" {
+ var err error
+ userid, err = models.GetUserIDByExternalUserID(tp, fmt.Sprintf("%v", review.ReviewerID))
+ if err != nil {
+ log.Error("GetUserIDByExternalUserID: %v", err)
+ }
+ if userid > 0 {
+ g.userMap[review.ReviewerID] = userid
+ }
+ }
+
+ var cm = models.Review{
+ Type: convertReviewState(review.State),
+ IssueID: issueID,
+ Content: review.Content,
+ Official: review.Official,
+ CreatedUnix: timeutil.TimeStamp(review.CreatedAt.Unix()),
+ UpdatedUnix: timeutil.TimeStamp(review.CreatedAt.Unix()),
+ }
+
+ if userid > 0 {
+ cm.ReviewerID = userid
+ } else {
+ cm.ReviewerID = g.doer.ID
+ cm.OriginalAuthor = review.ReviewerName
+ cm.OriginalAuthorID = review.ReviewerID
+ }
+
+ // get pr
+ pr, ok := g.prCache[issueID]
+ if !ok {
+ var err error
+ pr, err = models.GetPullRequestByIssueIDWithNoAttributes(issueID)
+ if err != nil {
+ return err
+ }
+ g.prCache[issueID] = pr
+ }
+
+ for _, comment := range review.Comments {
+ line := comment.Line
+ if line != 0 {
+ comment.Position = 1
+ } else {
+ _, _, line, _ = git.ParseDiffHunkString(comment.DiffHunk)
+ }
+ headCommitID, err := g.gitRepo.GetRefCommitID(pr.GetGitRefName())
+ if err != nil {
+ return fmt.Errorf("GetRefCommitID[%s]: %v", pr.GetGitRefName(), err)
+ }
+
+ var patch string
+ patchBuf := new(bytes.Buffer)
+ if err := git.GetRepoRawDiffForFile(g.gitRepo, pr.MergeBase, headCommitID, git.RawDiffNormal, comment.TreePath, patchBuf); err != nil {
+ // We should ignore the error since the commit maybe removed when force push to the pull request
+ log.Warn("GetRepoRawDiffForFile failed when migrating [%s, %s, %s, %s]: %v", g.gitRepo.Path, pr.MergeBase, headCommitID, comment.TreePath, err)
+ } else {
+ patch = git.CutDiffAroundLine(patchBuf, int64((&models.Comment{Line: int64(line + comment.Position - 1)}).UnsignedLine()), line < 0, setting.UI.CodeCommentLines)
+ }
+
+ var c = models.Comment{
+ Type: models.CommentTypeCode,
+ PosterID: comment.PosterID,
+ IssueID: issueID,
+ Content: comment.Content,
+ Line: int64(line + comment.Position - 1),
+ TreePath: comment.TreePath,
+ CommitSHA: comment.CommitID,
+ Patch: patch,
+ CreatedUnix: timeutil.TimeStamp(comment.CreatedAt.Unix()),
+ UpdatedUnix: timeutil.TimeStamp(comment.UpdatedAt.Unix()),
+ }
+
+ if userid > 0 {
+ c.PosterID = userid
+ } else {
+ c.PosterID = g.doer.ID
+ c.OriginalAuthor = review.ReviewerName
+ c.OriginalAuthorID = review.ReviewerID
+ }
+
+ cm.Comments = append(cm.Comments, &c)
+ }
+
+ cms = append(cms, &cm)
+ }
+
+ return models.InsertReviews(cms)
+}
+
+// Rollback when migrating failed, this will rollback all the changes.
+func (g *GiteaLocalUploader) Rollback() error {
+ if g.repo != nil && g.repo.ID > 0 {
+ if err := models.DeleteRepository(g.doer, g.repo.OwnerID, g.repo.ID); err != nil {
+ return err
+ }
+ }
+ return nil
+}