// Package bug contains the bug data model and low-level related functions package bug import ( "encoding/json" "fmt" "strings" "github.com/pkg/errors" "github.com/MichaelMure/git-bug/entity" "github.com/MichaelMure/git-bug/identity" "github.com/MichaelMure/git-bug/repository" "github.com/MichaelMure/git-bug/util/git" "github.com/MichaelMure/git-bug/util/lamport" ) const bugsRefPattern = "refs/bugs/" const bugsRemoteRefPattern = "refs/remotes/%s/bugs/" const opsEntryName = "ops" const rootEntryName = "root" const mediaEntryName = "media" const createClockEntryPrefix = "create-clock-" const createClockEntryPattern = "create-clock-%d" const editClockEntryPrefix = "edit-clock-" const editClockEntryPattern = "edit-clock-%d" var ErrBugNotExist = errors.New("bug doesn't exist") type ErrMultipleMatch struct { Matching []entity.Id } func (e ErrMultipleMatch) Error() string { matching := make([]string, len(e.Matching)) for i, match := range e.Matching { matching[i] = match.String() } return fmt.Sprintf("Multiple matching bug found:\n%s", strings.Join(matching, "\n")) } var _ Interface = &Bug{} var _ entity.Interface = &Bug{} // Bug hold the data of a bug thread, organized in a way close to // how it will be persisted inside Git. This is the data structure // used to merge two different version of the same Bug. type Bug struct { // A Lamport clock is a logical clock that allow to order event // inside a distributed system. // It must be the first field in this struct due to https://github.com/golang/go/issues/599 createTime lamport.Time editTime lamport.Time // Id used as unique identifier id entity.Id lastCommit git.Hash rootPack git.Hash // all the committed operations packs []OperationPack // a temporary pack of operations used for convenience to pile up new operations // before a commit staging OperationPack } // NewBug create a new Bug func NewBug() *Bug { // No id yet // No logical clock yet return &Bug{} } // FindLocalBug find an existing Bug matching a prefix func FindLocalBug(repo repository.ClockedRepo, prefix string) (*Bug, error) { ids, err := ListLocalIds(repo) if err != nil { return nil, err } // preallocate but empty matching := make([]entity.Id, 0, 5) for _, id := range ids { if id.HasPrefix(prefix) { matching = append(matching, id) } } if len(matching) == 0 { return nil, errors.New("no matching bug found.") } if len(matching) > 1 { return nil, ErrMultipleMatch{Matching: matching} } return ReadLocalBug(repo, matching[0]) } // ReadLocalBug will read a local bug from its hash func ReadLocalBug(repo repository.ClockedRepo, id entity.Id) (*Bug, error) { ref := bugsRefPattern + id.String() return readBug(repo, ref) } // ReadRemoteBug will read a remote bug from its hash func ReadRemoteBug(repo repository.ClockedRepo, remote string, id string) (*Bug, error) { ref := fmt.Sprintf(bugsRemoteRefPattern, remote) + id return readBug(repo, ref) } // readBug will read and parse a Bug from git func readBug(repo repository.ClockedRepo, ref string) (*Bug, error) { refSplit := strings.Split(ref, "/") id := entity.Id(refSplit[len(refSplit)-1]) if err := id.Validate(); err != nil { return nil, errors.Wrap(err, "invalid ref ") } hashes, err := repo.ListCommits(ref) // TODO: this is not perfect, it might be a command invoke error if err != nil { return nil, ErrBugNotExist } bug := Bug{ id: id, editTime: 0, } // Load each OperationPack for _, hash := range hashes { entries, err := repo.ListEntries(hash) if err != nil { return nil, errors.Wrap(err, "can't list git tree entries") } bug.lastCommit = hash var opsEntry repository.TreeEntry opsFound := false var rootEntry repository.TreeEntry rootFound := false var createTime uint64 var editTime uint64 for _, entry := range entries { if entry.Name == opsEntryName { opsEntry = entry opsFound = true continue } if entry.Name == rootEntryName { rootEntry = entry rootFound = true } if strings.HasPrefix(entry.Name, createClockEntryPrefix) { n, err := fmt.Sscanf(string(entry.Name), createClockEntryPattern, &createTime) if err != nil { return nil, errors.Wrap(err, "can't read create lamport time") } if n != 1 { return nil, fmt.Errorf("could not parse create time lamport value") } } if strings.HasPrefix(entry.Name, editClockEntryPrefix) { n, err := fmt.Sscanf(string(entry.Name), editClockEntryPattern, &editTime) if err != nil { return nil, errors.Wrap(err, "can't read edit lamport time") } if n != 1 { return nil, fmt.Errorf("could not parse edit time lamport value") } } } if !opsFound { return nil, errors.New("invalid tree, missing the ops entry") } if !rootFound { return nil, errors.New("invalid tree, missing the root entry") } if bug.rootPack == "" { bug.rootPack = rootEntry.Hash bug.createTime = lamport.Time(createTime) } // Due to rebase, edit Lamport time are not necessarily ordered if editTime > uint64(bug.editTime) { bug.editTime = lamport.Time(editTime) } // Update the clocks if err := repo.CreateWitness(bug.createTime); err != nil { return nil, errors.Wrap(err, "failed to update create lamport clock") } if err := repo.EditWitness(bug.editTime); err != nil { return nil, errors.Wrap(err, "failed to update edit lamport clock") } data, err := repo.ReadData(opsEntry.Hash) if err != nil { return nil, errors.Wrap(err, "failed to read git blob data") } opp := &OperationPack{} err = json.Unmarshal(data, &opp) if err != nil { return nil, errors.Wrap(err, "failed to decode OperationPack json") } // tag the pack with the commit hash opp.commitHash = hash bug.packs = append(bug.packs, *opp) } // Make sure that the identities are properly loaded resolver := identity.NewSimpleResolver(repo) err = bug.EnsureIdentities(resolver) if err != nil { return nil, err } return &bug, nil } type StreamedBug struct { Bug *Bug Err error } // ReadAllLocalBugs read and parse all local bugs func ReadAllLocalBugs(repo repository.ClockedRepo) <-chan StreamedBug { return readAllBugs(repo, bugsRefPattern) } // ReadAllRemoteBugs read and parse all remote bugs for a given remote func ReadAllRemoteBugs(repo repository.ClockedRepo, remote string) <-chan StreamedBug { refPrefix := fmt.Sprintf(bugsRemoteRefPattern, remote) return readAllBugs(repo, refPrefix) } // Read and parse all available bug with a given ref prefix func readAllBugs(repo repository.ClockedRepo, refPrefix string) <-chan StreamedBug { out := make(chan StreamedBug) go func() { defer close(out) refs, err := repo.ListRefs(refPrefix) if err != nil { out <- StreamedBug{Err: err} return } for _, ref := range refs { b, err := readBug(repo, ref) if err != nil { out <- StreamedBug{Err: err} return } out <- StreamedBug{Bug: b} } }() return out } // ListLocalIds list all the available local bug ids func ListLocalIds(repo repository.Repo) ([]entity.Id, error) { refs, err := repo.ListRefs(bugsRefPattern) if err != nil { return nil, err } return refsToIds(refs), nil } func refsToIds(refs []string) []entity.Id { ids := make([]entity.Id, len(refs)) for i, ref := range refs { split := strings.Split(ref, "/") ids[i] = entity.Id(split[len(split)-1]) } return ids } // Validate check if the Bug data is valid func (bug *Bug) Validate() error { // non-empty if len(bug.packs) == 0 && bug.staging.IsEmpty() { return fmt.Errorf("bug has no operations") } // check if each pack and operations are valid for _, pack := range bug.packs { if err := pack.Validate(); err != nil { return err } } // check if staging is valid if needed if !bug.staging.IsEmpty() { if err := bug.staging.Validate(); err != nil { return errors.Wrap(err, "staging") } } // The very first Op should be a CreateOp firstOp := bug.FirstOp() if firstOp == nil || firstOp.base().OperationType != CreateOp { return fmt.Errorf("first operation should be a Create op") } // The bug Id should be the hash of the first commit if len(bug.packs) > 0 && string(bug.packs[0].commitHash) != bug.id.String() { return fmt.Errorf("bug id should be the first commit hash") } // Check that there is no more CreateOp op it := NewOperationIterator(bug) createCount := 0 for it.Next() { if it.Value().base().OperationType == CreateOp { createCount++ } } if createCount != 1 { return fmt.Errorf("only one Create op allowed") } return nil } // Append an operation into the staging area, to be committed later func (bug *Bug) Append(op Operation) { bug.staging.Append(op) } // HasPendingOp tell if the bug need to be committed func (bug *Bug) HasPendingOp() bool { return !bug.staging.IsEmpty() } // Commit write the staging area in Git and move the operations to the packs func (bug *Bug) Commit(repo repository.ClockedRepo) error { if !bug.NeedCommit() { return fmt.Errorf("can't commit a bug with no pending operation") } if err := bug.Validate(); err != nil { return errors.Wrap(err, "can't commit a bug with invalid data") } // Write the Ops as a Git blob containing the serialized array hash, err := bug.staging.Write(repo) if err != nil { return err } if bug.rootPack == "" { bug.rootPack = hash } // Make a Git tree referencing this blob tree := []repository.TreeEntry{ // the last pack of ops {ObjectType: repository.Blob, Hash: hash, Name: opsEntryName}, // always the first pack of ops (might be the same) {ObjectType: repository.Blob, Hash: bug.rootPack, Name: rootEntryName}, } // Reference, if any, all the files required by the ops // Git will check that they actually exist in the storage and will make sure // to push/pull them as needed. mediaTree := makeMediaTree(bug.staging) if len(mediaTree) > 0 { mediaTreeHash, err := repo.StoreTree(mediaTree) if err != nil { return err } tree = append(tree, repository.TreeEntry{ ObjectType: repository.Tree, Hash: mediaTreeHash, Name: mediaEntryName, }) } // Store the logical clocks as well // --> edit clock for each OperationPack/commits // --> create clock only for the first OperationPack/commits // // To avoid having one blob for each clock value, clocks are serialized // directly into the entry name emptyBlobHash, err := repo.StoreData([]byte{}) if err != nil { return err } bug.editTime, err = repo.EditTimeIncrement() if err != nil { return err } tree = append(tree, repository.TreeEntry{ ObjectType: repository.Blob, Hash: emptyBlobHash, Name: fmt.Sprintf(editClockEntryPattern, bug.editTime), }) if bug.lastCommit == "" { bug.createTime, err = repo.CreateTimeIncrement() if err != nil { return err } tree = append(tree, repository.TreeEntry{ ObjectType: repository.Blob, Hash: emptyBlobHash, Name: fmt.Sprintf(createClockEntryPattern, bug.createTime), }) } // Store the tree hash, err = repo.StoreTree(tree) if err != nil { return err } // Write a Git commit referencing the tree, with the previous commit as parent if bug.lastCommit != "" { hash, err = repo.StoreCommitWithParent(hash, bug.lastCommit) } else { hash, err = repo.StoreCommit(hash) } if err != nil { return err } bug.lastCommit = hash // if it was the first commit, use the commit hash as bug id if bug.id == "" { bug.id = entity.Id(hash) } // Create or update the Git reference for this bug // When pushing later, the remote will ensure that this ref update // is fast-forward, that is no data has been overwritten ref := fmt.Sprintf("%s%s", bugsRefPattern, bug.id) err = repo.UpdateRef(ref, hash) if err != nil { return err } bug.staging.commitHash = hash bug.packs = append(bug.packs, bug.staging) bug.staging = OperationPack{} return nil } func (bug *Bug) CommitAsNeeded(repo repository.ClockedRepo) error { if !bug.NeedCommit() { return nil } return bug.Commit(repo) } func (bug *Bug) NeedCommit() bool { return !bug.staging.IsEmpty() } func makeMediaTree(pack OperationPack) []repository.TreeEntry { var tree []repository.TreeEntry counter := 0 added := make(map[git.Hash]interface{}) for _, ops := range pack.Operations { for _, file := range ops.GetFiles() { if _, has := added[file]; !has { tree = append(tree, repository.TreeEntry{ ObjectType: repository.Blob, Hash: file, // The name is not important here, we only need to // reference the blob. Name: fmt.Sprintf("file%d", counter), }) counter++ added[file] = struct{}{} } } } return tree } // Merge a different version of the same bug by rebasing operations of this bug // that are not present in the other on top of the chain of operations of the // other version. func (bug *Bug) Merge(repo repository.Repo, other Interface) (bool, error) { var otherBug = bugFromInterface(other) // Note: a faster merge should be possible without actually reading and parsing // all operations pack of our side. // Reading the other side is still necessary to validate remote data, at least // for new operations if bug.id != otherBug.id { return false, errors.New("merging unrelated bugs is not supported") } if len(otherBug.staging.Operations) > 0 { return false, errors.New("merging a bug with a non-empty staging is not supported") } if bug.lastCommit == "" || otherBug.lastCommit == "" { return false, errors.New("can't merge a bug that has never been stored") } ancestor, err := repo.FindCommonAncestor(bug.lastCommit, otherBug.lastCommit) if err != nil { return false, errors.Wrap(err, "can't find common ancestor") } ancestorIndex := 0 newPacks := make([]OperationPack, 0, len(bug.packs)) // Find the root of the rebase for i, pack := range bug.packs { newPacks = append(newPacks, pack) if pack.commitHash == ancestor { ancestorIndex = i break } } if len(otherBug.packs) == ancestorIndex+1 { // Nothing to rebase, return early return false, nil } // get other bug's extra packs for i := ancestorIndex + 1; i < len(otherBug.packs); i++ { // clone is probably not necessary newPack := otherBug.packs[i].Clone() newPacks = append(newPacks, newPack) bug.lastCommit = newPack.commitHash } // rebase our extra packs for i := ancestorIndex + 1; i < len(bug.packs); i++ { pack := bug.packs[i] // get the referenced git tree treeHash, err := repo.GetTreeHash(pack.commitHash) if err != nil { return false, err } // create a new commit with the correct ancestor hash, err := repo.StoreCommitWithParent(treeHash, bug.lastCommit) if err != nil { return false, err } // replace the pack newPack := pack.Clone() newPack.commitHash = hash newPacks = append(newPacks, newPack) // update the bug bug.lastCommit = hash } // Update the git ref err = repo.UpdateRef(bugsRefPattern+bug.id.String(), bug.lastCommit) if err != nil { return false, err } return true, nil } // Id return the Bug identifier func (bug *Bug) Id() entity.Id { if bug.id == "" { // simply panic as it would be a coding error // (using an id of a bug not stored yet) panic("no id yet") } return bug.id } // CreateLamportTime return the Lamport time of creation func (bug *Bug) CreateLamportTime() lamport.Time { return bug.createTime } // EditLamportTime return the Lamport time of the last edit func (bug *Bug) EditLamportTime() lamport.Time { return bug.editTime } // Lookup for the very first operation of the bug. // For a valid Bug, this operation should be a CreateOp func (bug *Bug) FirstOp() Operation { for _, pack := range bug.packs { for _, op := range pack.Operations { return op } } if !bug.staging.IsEmpty() { return bug.staging.Operations[0] } return nil } // Lookup for the very last operation of the bug. // For a valid Bug, should never be nil func (bug *Bug) LastOp() Operation { if !bug.staging.IsEmpty() { return bug.staging.Operations[len(bug.staging.Operations)-1] } if len(bug.packs) == 0 { return nil } lastPack := bug.packs[len(bug.packs)-1] if len(lastPack.Operations) == 0 { return nil } return lastPack.Operations[len(lastPack.Operations)-1] } // Compile a bug in a easily usable snapshot func (bug *Bug) Compile() Snapshot { snap := Snapshot{ id: bug.id, Status: OpenStatus, } it := NewOperationIterator(bug) for it.Next() { op := it.Value() op.Apply(&snap) snap.Operations = append(snap.Operations, op) } return snap } // Sign post method for gqlgen func (bug *Bug) IsAuthored() {}