diff options
author | tjpcc <tjp@ctrl-c.club> | 2023-08-30 10:29:47 -0600 |
---|---|---|
committer | tjpcc <tjp@ctrl-c.club> | 2023-08-30 10:29:47 -0600 |
commit | 77ac83e700415cdbd4635eae91e79f99312ea240 (patch) | |
tree | 65205fbf85c2ba6bf7712f65af49eca564c49365 /iris |
Initial commit
* iris support copied in from iris-news
* a new slog backend
* "metabackend" wraps and routes between multiple backends based on the
groups they support
* better logging than iris-news ever had
Diffstat (limited to 'iris')
-rw-r--r-- | iris/backend.go | 448 |
1 files changed, 448 insertions, 0 deletions
diff --git a/iris/backend.go b/iris/backend.go new file mode 100644 index 0000000..787db96 --- /dev/null +++ b/iris/backend.go @@ -0,0 +1,448 @@ +package iris + +import ( + "bytes" + "crypto/sha1" + "encoding/base64" + "encoding/json" + "fmt" + "io" + "net/mail" + "net/textproto" + "os" + "os/exec" + "path" + "sort" + "strconv" + "strings" + "time" + + "github.com/dustin/go-nntp" + nntpserver "github.com/dustin/go-nntp/server" + "github.com/go-kit/log" + "github.com/go-kit/log/level" +) + +var group = &nntp.Group{ + Name: "ctrl-c.iris", + Description: "The iris message board", + Posting: nntp.PostingPermitted, + Low: 1, +} + +const DefaultWaitTime = 30 * time.Second +const msgfile = ".iris.messages" + +// NewBackend builds an iris nntp backend. +// +// The provided waitTime may be <= 0, in which case DefaultWaitTime will be used. +func NewBackend(logger log.Logger, waitTime time.Duration) (nntpserver.Backend, error) { + if waitTime <= 0 { + waitTime = DefaultWaitTime + } + + b := &backend{logger: logger, waitTime: waitTime} + if err := b.refresh(); err != nil { + return nil, err + } + + return b, nil +} + +type backend struct { + logger log.Logger + waitTime time.Duration + lastRead time.Time + messages []*nntp.Article +} + +func (b backend) debug(keyvals ...any) error { return level.Debug(b.logger).Log(keyvals...) } +func (b backend) info(keyvals ...any) error { return level.Info(b.logger).Log(keyvals...) } +func (b backend) warn(keyvals ...any) error { return level.Warn(b.logger).Log(keyvals...) } +func (b backend) err(keyvals ...any) error { return level.Error(b.logger).Log(keyvals...) } + +func (b backend) ListGroups(max int) ([]*nntp.Group, error) { + return []*nntp.Group{group}, nil +} + +func (b *backend) GetGroup(name string) (*nntp.Group, error) { + if name != group.Name { + return nil, nntpserver.ErrNoSuchGroup + } + if err := b.refresh(); err != nil { + return nil, err + } + + return group, nil +} + +func (b *backend) GetArticles(_ *nntp.Group, from, to int64) ([]nntpserver.NumberedArticle, error) { + if err := b.refresh(); err != nil { + return nil, err + } + + numbered := make([]nntpserver.NumberedArticle, 0, len(b.messages)) + for i, msg := range b.messages { + num := int64(i + 1) + if num >= from && num <= to { + numbered = append(numbered, nntpserver.NumberedArticle{ + Num: num, + Article: copyArticle(msg), + }) + } + } + + return numbered, nil +} + +func (b *backend) GetArticle(_ *nntp.Group, messageID string) (*nntp.Article, error) { + if err := b.refresh(); err != nil { + return nil, err + } + + for _, msg := range b.messages { + if msg.Header.Get("Message-Id") == messageID { + return msg, nil + } + } + + num, err := strconv.Atoi(messageID) + if err == nil && num <= len(b.messages) { + return copyArticle(b.messages[num-1]), nil + } + + return nil, nntpserver.ErrInvalidMessageID +} + +func (b backend) Post(article *nntp.Article) error { + // iris replies are all made to a top-level post, there is no grandchild nesting. + // + // but NNTP supports this, so collapse the provided "References" header up to the OP. + parent := b.findOP(article.Header.Get("References")) + if parent != nil { + article.Header.Set("References", parent.MessageID()) + } + + msg, err := msgToIris(article) + if err != nil { + return err + } + return appendMessage(msg) +} + +func (b backend) Authorized() bool { return true } +func (b backend) AllowPost() bool { return true } +func (b backend) Authenticate(_, _ string) (nntpserver.Backend, error) { return nil, nil } + +func (b backend) findOP(ref string) *nntp.Article { + if ref == "" { + return nil + } + // all references should have the same OP so just take the first + msgID := strings.SplitN(ref, " ", 2)[0] + + // traverse backwards expecting most reply activity concentrated late in the total history + for i := len(b.messages) - 1; i >= 0; i-- { + article := b.messages[i] + if article.MessageID() != msgID { + continue + } + + gpID := article.Header.Get("References") + if gpID != "" { + return b.findOP(gpID) + } + return article + } + + return nil +} + +func (b *backend) refresh() error { + now := time.Now() + if b.lastRead.IsZero() || now.Sub(b.lastRead) > b.waitTime { + b.lastRead = now + } else { + return nil + } + + binpath, err := exec.LookPath("iris") + if err != nil { + return err + } + cmd := exec.Command(binpath, "-d") + buf := &bytes.Buffer{} + cmd.Stdout = buf + + if err := cmd.Run(); err != nil { + return err + } + + msgs := irisDump{} + if err := json.NewDecoder(buf).Decode(&msgs); err != nil { + return err + } + + b.messages, err = msgs.Articles() + if err != nil { + return err + } + group.High = int64(len(b.messages)) + group.Count = int64(len(b.messages)) + return nil +} + +func copyArticle(article *nntp.Article) *nntp.Article { + out := *article + out.Body = bytes.NewBuffer(article.Body.(*bytes.Buffer).Bytes()) + return &out +} + +type irisMsg struct { + Hash string `json:"hash"` + EditHash *string `json:"edit_hash"` + IsDeleted *bool `json:"is_deleted"` + Data struct { + Author string `json:"author"` + Parent *string `json:"parent"` + Timestamp string `json:"timestamp"` + Message string `json:"message"` + } `json:"data"` +} + +func (m irisMsg) calcHash() (string, error) { + /* + Careful coding here to match ruby's hash calculation: + ``` + Base64.encode64(Digest::SHA1.digest(m["data"].to_json)) + ``` + + * have to use an encoder rather than json.Marshal so we can + turn off the default HTML escaping (ruby doesn't do this) + * strip trailing newline from JSON encoding output + * add a trailing newline to base64 encoded form + */ + + b := &bytes.Buffer{} + enc := json.NewEncoder(b) + enc.SetEscapeHTML(false) + if err := enc.Encode(m.Data); err != nil { + return "", err + } + + arr := sha1.Sum(bytes.TrimSuffix(b.Bytes(), []byte("\n"))) + s := base64.StdEncoding.EncodeToString(arr[:]) + if !strings.HasSuffix(s, "\n") { + s += "\n" + } + return s, nil +} + +func msgToIris(article *nntp.Article) (*irisMsg, error) { + postTime := time.Now().UTC().Format(time.RFC3339) + + body, err := io.ReadAll(article.Body) + if err != nil { + return nil, err + } + + var msg irisMsg + msg.Data.Author = irisAuthor(article.Header.Get("From")) + msg.Data.Timestamp = postTime + msg.Data.Message = string(body) + refs := article.Header.Get("References") + if refs != "" { + spl := strings.SplitN(refs, " ", 2) + ref := fromMsgID(spl[0]) + msg.Data.Parent = &ref + } + + hash, err := msg.calcHash() + if err != nil { + return nil, err + } + msg.Hash = hash + + return &msg, nil +} + +func irisAuthor(nntpAuthor string) string { + addr, err := mail.ParseAddress(nntpAuthor) + if err != nil { + return nntpAuthor + } + + return addr.Address +} + +type irisDump []irisMsg + +func (dump irisDump) Articles() ([]*nntp.Article, error) { + // calculate the article replacements due to edits + // + // note: this is only a single "hop", and because there can be edits-of-edits + // and edits-of-edits-of-edits, we must actually resolve replacements with a loop. + // + // we need to keep all the hops though because there could have been replies to + // the original or to any intermediate edits. + replacements := make(map[string]string) + for _, msg := range dump { + if msg.EditHash != nil { + replacements[*msg.EditHash] = msg.Hash + } + } + + articles := make([]*nntp.Article, 0, len(dump)-len(replacements)) + + // index iris hashes -> nntp Articles for reference lookups + idx := make(map[string]*nntp.Article) + + sort.SliceStable(dump, func(i, j int) bool { + return dump[i].Data.Timestamp < dump[j].Data.Timestamp + }) + +outer: + for _, msg := range dump { + if _, ok := replacements[msg.Hash]; ok { + continue + } + if msg.EditHash != nil && *msg.EditHash == msg.Hash { + continue + } + + msgID := msgIDFor(&msg) + ts, err := time.Parse(time.RFC3339, msg.Data.Timestamp) + if err != nil { + return nil, err + } + + article := &nntp.Article{ + Header: textproto.MIMEHeader{ + "Message-Id": []string{msgID}, + "From": []string{msg.Data.Author}, + "Newsgroups": []string{group.Name}, + "Date": []string{ts.Format(time.RFC1123Z)}, + }, + } + + if msg.IsDeleted != nil && *msg.IsDeleted { + article.Header.Set("Subject", "**TOPIC DELETED BY AUTHOR**") + article.Body = &bytes.Buffer{} + article.Bytes = 0 + article.Lines = 0 + } else { + article.Body = bytes.NewBufferString(msg.Data.Message) + article.Bytes = len(msg.Data.Message) + article.Lines = strings.Count(msg.Data.Message, "\n") + + if msg.Data.Parent == nil { + article.Header.Set("Subject", strings.SplitN(msg.Data.Message, "\n", 2)[0]) + } else { + parentHash := *msg.Data.Parent + for { + if p, ok := replacements[parentHash]; ok { + if parentHash == p { + continue outer + } + parentHash = p + } else { + break + } + } + msg.Data.Parent = &parentHash + parent, ok := idx[parentHash] + if !ok { + continue + } + parentSubj := strings.TrimPrefix(parent.Header.Get("Subject"), "Re: ") + article.Header.Set("Subject", "Re: "+parentSubj) + } + } + + if msg.Data.Parent != nil { + parent := idx[*msg.Data.Parent] + if parent == nil { + continue + } + parentRefs := parent.Header.Get("References") + if parentRefs != "" { + article.Header.Set("References", parentRefs) + } else { + article.Header.Set("References", parent.MessageID()) + } + } + + articles = append(articles, article) + idx[msg.Hash] = article + } + + return articles, nil +} + +func (id irisDump) MarshalJSON() ([]byte, error) { + buf := &bytes.Buffer{} + enc := json.NewEncoder(buf) + enc.SetEscapeHTML(false) + + out := bytes.NewBufferString("[\n ") + + for i, msg := range id { + if err := enc.Encode(msg); err != nil { + return nil, err + } + + if i > 0 { + _, _ = out.WriteString(",\n ") + } + _, _ = out.Write(bytes.TrimSuffix(buf.Bytes(), []byte("\n"))) + buf.Reset() + } + _, _ = out.WriteString("\n]") + + return out.Bytes(), nil +} + +func msgIDFor(msg *irisMsg) string { + return fmt.Sprintf("<%s.%s>", + strings.TrimSuffix(msg.Hash, "=\n"), + msg.Data.Author, + ) +} + +func fromMsgID(nntpID string) string { + hash, _, _ := strings.Cut(strings.TrimSuffix(strings.TrimPrefix(nntpID, "<"), ">"), ".") + return hash + "=\n" +} + +func appendMessage(msg *irisMsg) error { + home, err := os.UserHomeDir() + if err != nil { + return err + } + + msgFile, err := os.Open(path.Join(home, msgfile)) + if err != nil { + return err + } + + var msgs irisDump + if err := json.NewDecoder(msgFile).Decode(&msgs); err != nil { + _ = msgFile.Close() + return err + } + _ = msgFile.Close() + msgs = append(msgs, *msg) + + msgFile, err = os.Create(path.Join(home, msgfile)) + if err != nil { + return err + } + defer func() { _ = msgFile.Close() }() + + out, err := msgs.MarshalJSON() + if err != nil { + return err + } + _, err = msgFile.Write(out) + return err +} |