Merge pull request #220 from ssb-ngi-pointer/room-attendants
room.attendants
This commit is contained in:
commit
d8da4ae94b
|
@ -0,0 +1,125 @@
|
|||
// SPDX-License-Identifier: MIT
|
||||
|
||||
package broadcasts
|
||||
|
||||
import (
|
||||
"io"
|
||||
"sync"
|
||||
|
||||
"github.com/ssb-ngi-pointer/go-ssb-room/internal/maybemod/multierror"
|
||||
refs "go.mindeco.de/ssb-refs"
|
||||
)
|
||||
|
||||
type AttendantsEmitter interface {
|
||||
Joined(member refs.FeedRef) error
|
||||
Left(member refs.FeedRef) error
|
||||
|
||||
io.Closer
|
||||
}
|
||||
|
||||
// NewAttendantsEmitter returns the Sink, to write to the broadcaster, and the new
|
||||
// broadcast instance.
|
||||
func NewAttendantsEmitter() (AttendantsEmitter, *AttendantsBroadcast) {
|
||||
bcst := AttendantsBroadcast{
|
||||
mu: &sync.Mutex{},
|
||||
sinks: make(map[*AttendantsEmitter]struct{}),
|
||||
}
|
||||
|
||||
return (*attendantsSink)(&bcst), &bcst
|
||||
}
|
||||
|
||||
// AttendantsBroadcast is an interface for registering one or more Sinks to recieve
|
||||
// updates.
|
||||
type AttendantsBroadcast struct {
|
||||
mu *sync.Mutex
|
||||
sinks map[*AttendantsEmitter]struct{}
|
||||
}
|
||||
|
||||
// Register a Sink for updates to be sent. also returns
|
||||
func (bcst *AttendantsBroadcast) Register(sink AttendantsEmitter) func() {
|
||||
bcst.mu.Lock()
|
||||
defer bcst.mu.Unlock()
|
||||
bcst.sinks[&sink] = struct{}{}
|
||||
|
||||
return func() {
|
||||
bcst.mu.Lock()
|
||||
defer bcst.mu.Unlock()
|
||||
delete(bcst.sinks, &sink)
|
||||
sink.Close()
|
||||
}
|
||||
}
|
||||
|
||||
type attendantsSink AttendantsBroadcast
|
||||
|
||||
func (bcst *attendantsSink) Joined(member refs.FeedRef) error {
|
||||
bcst.mu.Lock()
|
||||
for s := range bcst.sinks {
|
||||
err := (*s).Joined(member)
|
||||
if err != nil {
|
||||
delete(bcst.sinks, s)
|
||||
}
|
||||
}
|
||||
bcst.mu.Unlock()
|
||||
|
||||
return nil
|
||||
}
|
||||
|
||||
func (bcst *attendantsSink) Left(member refs.FeedRef) error {
|
||||
bcst.mu.Lock()
|
||||
for s := range bcst.sinks {
|
||||
err := (*s).Left(member)
|
||||
if err != nil {
|
||||
delete(bcst.sinks, s)
|
||||
}
|
||||
}
|
||||
bcst.mu.Unlock()
|
||||
|
||||
return nil
|
||||
}
|
||||
|
||||
// Close implements the Sink interface.
|
||||
func (bcst *attendantsSink) Close() error {
|
||||
bcst.mu.Lock()
|
||||
defer bcst.mu.Unlock()
|
||||
sinks := make([]AttendantsEmitter, 0, len(bcst.sinks))
|
||||
|
||||
for sink := range bcst.sinks {
|
||||
sinks = append(sinks, *sink)
|
||||
}
|
||||
|
||||
bcst.mu.Lock()
|
||||
defer bcst.mu.Unlock()
|
||||
|
||||
sinks = make([]AttendantsEmitter, 0, len(bcst.sinks))
|
||||
|
||||
for sink := range bcst.sinks {
|
||||
sinks = append(sinks, *sink)
|
||||
}
|
||||
|
||||
var (
|
||||
wg sync.WaitGroup
|
||||
me multierror.List
|
||||
)
|
||||
|
||||
// might be fine without the waitgroup and concurrency
|
||||
|
||||
wg.Add(len(sinks))
|
||||
for _, sink_ := range sinks {
|
||||
go func(sink AttendantsEmitter) {
|
||||
defer wg.Done()
|
||||
|
||||
err := sink.Close()
|
||||
if err != nil {
|
||||
me.Errs = append(me.Errs, err)
|
||||
return
|
||||
}
|
||||
}(sink_)
|
||||
}
|
||||
wg.Wait()
|
||||
|
||||
if len(me.Errs) == 0 {
|
||||
return nil
|
||||
}
|
||||
|
||||
return me
|
||||
}
|
|
@ -0,0 +1,5 @@
|
|||
// SPDX-License-Identifier: MIT
|
||||
|
||||
// Package broadcasts implements custom typed one-to-n facilities for broadcasting messages/calls to multiple subscribers.
|
||||
// They loosely follow from luigi.Broadcasts but using concrete types instead of empty interfaces.
|
||||
package broadcasts
|
|
@ -9,31 +9,31 @@ import (
|
|||
"github.com/ssb-ngi-pointer/go-ssb-room/internal/maybemod/multierror"
|
||||
)
|
||||
|
||||
type RoomChangeSink interface {
|
||||
type EndpointsEmitter interface {
|
||||
Update(members []string) error
|
||||
io.Closer
|
||||
}
|
||||
|
||||
// NewRoomChanger returns the Sink, to write to the broadcaster, and the new
|
||||
// NewEndpointsEmitter returns the Sink, to write to the broadcaster, and the new
|
||||
// broadcast instance.
|
||||
func NewRoomChanger() (RoomChangeSink, *RoomChangeBroadcast) {
|
||||
bcst := RoomChangeBroadcast{
|
||||
func NewEndpointsEmitter() (EndpointsEmitter, *EndpointsBroadcast) {
|
||||
bcst := EndpointsBroadcast{
|
||||
mu: &sync.Mutex{},
|
||||
sinks: make(map[*RoomChangeSink]struct{}),
|
||||
sinks: make(map[*EndpointsEmitter]struct{}),
|
||||
}
|
||||
|
||||
return (*broadcastSink)(&bcst), &bcst
|
||||
return (*endpointsSink)(&bcst), &bcst
|
||||
}
|
||||
|
||||
// RoomChangeBroadcast is an interface for registering one or more Sinks to recieve
|
||||
// EndpointsBroadcast is an interface for registering one or more Sinks to recieve
|
||||
// updates.
|
||||
type RoomChangeBroadcast struct {
|
||||
type EndpointsBroadcast struct {
|
||||
mu *sync.Mutex
|
||||
sinks map[*RoomChangeSink]struct{}
|
||||
sinks map[*EndpointsEmitter]struct{}
|
||||
}
|
||||
|
||||
// Register a Sink for updates to be sent. also returns
|
||||
func (bcst *RoomChangeBroadcast) Register(sink RoomChangeSink) func() {
|
||||
func (bcst *EndpointsBroadcast) Register(sink EndpointsEmitter) func() {
|
||||
bcst.mu.Lock()
|
||||
defer bcst.mu.Unlock()
|
||||
bcst.sinks[&sink] = struct{}{}
|
||||
|
@ -46,10 +46,10 @@ func (bcst *RoomChangeBroadcast) Register(sink RoomChangeSink) func() {
|
|||
}
|
||||
}
|
||||
|
||||
type broadcastSink RoomChangeBroadcast
|
||||
type endpointsSink EndpointsBroadcast
|
||||
|
||||
// Pour implements the Sink interface.
|
||||
func (bcst *broadcastSink) Update(members []string) error {
|
||||
func (bcst *endpointsSink) Update(members []string) error {
|
||||
|
||||
bcst.mu.Lock()
|
||||
for s := range bcst.sinks {
|
||||
|
@ -64,13 +64,13 @@ func (bcst *broadcastSink) Update(members []string) error {
|
|||
}
|
||||
|
||||
// Close implements the Sink interface.
|
||||
func (bcst *broadcastSink) Close() error {
|
||||
var sinks []RoomChangeSink
|
||||
func (bcst *endpointsSink) Close() error {
|
||||
var sinks []EndpointsEmitter
|
||||
|
||||
bcst.mu.Lock()
|
||||
defer bcst.mu.Unlock()
|
||||
|
||||
sinks = make([]RoomChangeSink, 0, len(bcst.sinks))
|
||||
sinks = make([]EndpointsEmitter, 0, len(bcst.sinks))
|
||||
|
||||
for sink := range bcst.sinks {
|
||||
sinks = append(sinks, *sink)
|
||||
|
@ -85,7 +85,7 @@ func (bcst *broadcastSink) Close() error {
|
|||
|
||||
wg.Add(len(sinks))
|
||||
for _, sink_ := range sinks {
|
||||
go func(sink RoomChangeSink) {
|
||||
go func(sink EndpointsEmitter) {
|
||||
defer wg.Done()
|
||||
|
||||
err := sink.Close()
|
|
@ -24,7 +24,7 @@ func (tp testPrinter) Update(members []string) error {
|
|||
func (tp testPrinter) Close() error { return nil }
|
||||
|
||||
func ExampleBroadcast() {
|
||||
sink, bcast := NewRoomChanger()
|
||||
sink, bcast := NewEndpointsEmitter()
|
||||
defer sink.Close()
|
||||
|
||||
var p1, p2 testPrinter
|
||||
|
@ -44,7 +44,7 @@ func ExampleBroadcast() {
|
|||
}
|
||||
|
||||
func ExampleBroadcastCanceled() {
|
||||
sink, bcast := NewRoomChanger()
|
||||
sink, bcast := NewEndpointsEmitter()
|
||||
defer sink.Close()
|
||||
|
||||
var p1, p2 testPrinter
|
||||
|
@ -78,7 +78,7 @@ func (tp erroringPrinter) Close() error { return nil }
|
|||
func TestBroadcastOneErrs(t *testing.T) {
|
||||
var buf = &bytes.Buffer{}
|
||||
|
||||
sink, bcast := NewRoomChanger()
|
||||
sink, bcast := NewEndpointsEmitter()
|
||||
defer sink.Close()
|
||||
|
||||
var p1 testPrinter
|
||||
|
@ -144,7 +144,7 @@ func TestBroadcast(t *testing.T) {
|
|||
tc.rx = tc.tx
|
||||
}
|
||||
|
||||
sink, bcast := NewRoomChanger()
|
||||
sink, bcast := NewEndpointsEmitter()
|
||||
|
||||
mkSink := func() Sink {
|
||||
var (
|
|
@ -0,0 +1,101 @@
|
|||
// SPDX-License-Identifier: MIT
|
||||
|
||||
package server
|
||||
|
||||
import (
|
||||
"context"
|
||||
"encoding/json"
|
||||
"fmt"
|
||||
"sync"
|
||||
|
||||
"github.com/ssb-ngi-pointer/go-ssb-room/internal/network"
|
||||
"github.com/ssb-ngi-pointer/go-ssb-room/roomdb"
|
||||
"go.cryptoscope.co/muxrpc/v2"
|
||||
refs "go.mindeco.de/ssb-refs"
|
||||
)
|
||||
|
||||
// AttendantsUpdate is emitted if a single member joins or leaves.
|
||||
// Type is either 'joined' or 'left'.
|
||||
type AttendantsUpdate struct {
|
||||
Type string `json:"type"`
|
||||
ID refs.FeedRef `json:"id"`
|
||||
}
|
||||
|
||||
// AttendantsInitialState is emitted the first time the stream is opened
|
||||
type AttendantsInitialState struct {
|
||||
Type string `json:"type"`
|
||||
IDs []refs.FeedRef `json:"ids"`
|
||||
}
|
||||
|
||||
func (h *Handler) attendants(ctx context.Context, req *muxrpc.Request, snk *muxrpc.ByteSink) error {
|
||||
// for future updates
|
||||
toPeer := newAttendantsEncoder(snk)
|
||||
h.state.RegisterAttendantsUpdates(toPeer)
|
||||
|
||||
// get public key from the calling peer
|
||||
peer, err := network.GetFeedRefFromAddr(req.RemoteAddr())
|
||||
if err != nil {
|
||||
return err
|
||||
}
|
||||
|
||||
pm, err := h.config.GetPrivacyMode(ctx)
|
||||
if err != nil {
|
||||
return fmt.Errorf("running with unknown privacy mode")
|
||||
}
|
||||
|
||||
if pm == roomdb.ModeCommunity || pm == roomdb.ModeRestricted {
|
||||
_, err := h.members.GetByFeed(ctx, *peer)
|
||||
if err != nil {
|
||||
return fmt.Errorf("external user are not allowed to enumerate members")
|
||||
}
|
||||
}
|
||||
err = json.NewEncoder(snk).Encode(AttendantsInitialState{
|
||||
Type: "state",
|
||||
IDs: h.state.ListAsRefs(),
|
||||
})
|
||||
if err != nil {
|
||||
return err
|
||||
}
|
||||
|
||||
return nil
|
||||
}
|
||||
|
||||
// a muxrpc json encoder for endpoints broadcasts
|
||||
type attendantsJSONEncoder struct {
|
||||
mu sync.Mutex // only one caller to forwarder at a time
|
||||
snk *muxrpc.ByteSink
|
||||
enc *json.Encoder
|
||||
}
|
||||
|
||||
func newAttendantsEncoder(snk *muxrpc.ByteSink) *attendantsJSONEncoder {
|
||||
enc := json.NewEncoder(snk)
|
||||
snk.SetEncoding(muxrpc.TypeJSON)
|
||||
return &attendantsJSONEncoder{
|
||||
snk: snk,
|
||||
enc: enc,
|
||||
}
|
||||
}
|
||||
|
||||
func (uf *attendantsJSONEncoder) Joined(member refs.FeedRef) error {
|
||||
uf.mu.Lock()
|
||||
defer uf.mu.Unlock()
|
||||
return uf.enc.Encode(AttendantsUpdate{
|
||||
Type: "joined",
|
||||
ID: member,
|
||||
})
|
||||
}
|
||||
|
||||
func (uf *attendantsJSONEncoder) Left(member refs.FeedRef) error {
|
||||
uf.mu.Lock()
|
||||
defer uf.mu.Unlock()
|
||||
return uf.enc.Encode(AttendantsUpdate{
|
||||
Type: "left",
|
||||
ID: member,
|
||||
})
|
||||
}
|
||||
|
||||
func (uf *attendantsJSONEncoder) Close() error {
|
||||
uf.mu.Lock()
|
||||
defer uf.mu.Unlock()
|
||||
return uf.snk.Close()
|
||||
}
|
|
@ -34,7 +34,8 @@ func New(log kitlog.Logger, netInfo network.ServerEndpointDetails, m *roomstate.
|
|||
return h
|
||||
}
|
||||
|
||||
func (h *Handler) Register(mux typemux.HandlerMux, namespace muxrpc.Method) {
|
||||
func (h *Handler) RegisterTunnel(mux typemux.HandlerMux) {
|
||||
var namespace = muxrpc.Method{"tunnel"}
|
||||
mux.RegisterAsync(append(namespace, "isRoom"), typemux.AsyncFunc(h.metadata))
|
||||
mux.RegisterAsync(append(namespace, "ping"), typemux.AsyncFunc(h.ping))
|
||||
|
||||
|
@ -49,3 +50,20 @@ func (h *Handler) Register(mux typemux.HandlerMux, namespace muxrpc.Method) {
|
|||
state: h.state,
|
||||
})
|
||||
}
|
||||
|
||||
func (h *Handler) RegisterRoom(mux typemux.HandlerMux) {
|
||||
var namespace = muxrpc.Method{"room"}
|
||||
mux.RegisterAsync(append(namespace, "metadata"), typemux.AsyncFunc(h.metadata))
|
||||
mux.RegisterAsync(append(namespace, "ping"), typemux.AsyncFunc(h.ping))
|
||||
|
||||
mux.RegisterAsync(append(namespace, "announce"), typemux.AsyncFunc(h.announce))
|
||||
mux.RegisterAsync(append(namespace, "leave"), typemux.AsyncFunc(h.leave))
|
||||
|
||||
mux.RegisterSource(append(namespace, "attendants"), typemux.SourceFunc(h.attendants))
|
||||
|
||||
mux.RegisterDuplex(append(namespace, "connect"), connectHandler{
|
||||
logger: h.logger,
|
||||
self: h.netInfo.RoomID,
|
||||
state: h.state,
|
||||
})
|
||||
}
|
||||
|
|
|
@ -105,10 +105,10 @@ func (h *Handler) leave(_ context.Context, req *muxrpc.Request) (interface{}, er
|
|||
}
|
||||
|
||||
func (h *Handler) endpoints(ctx context.Context, req *muxrpc.Request, snk *muxrpc.ByteSink) error {
|
||||
toPeer := newForwarder(snk)
|
||||
toPeer := newEndpointsForwarder(snk)
|
||||
|
||||
// for future updates
|
||||
h.state.Register(toPeer)
|
||||
h.state.RegisterLegacyEndpoints(toPeer)
|
||||
|
||||
// get public key from the calling peer
|
||||
peer, err := network.GetFeedRefFromAddr(req.RemoteAddr())
|
||||
|
@ -140,28 +140,29 @@ func (h *Handler) endpoints(ctx context.Context, req *muxrpc.Request, snk *muxrp
|
|||
return nil
|
||||
}
|
||||
|
||||
type updateForwarder struct {
|
||||
// a muxrpc json encoder for endpoints broadcasts
|
||||
type endpointsJSONEncoder struct {
|
||||
mu sync.Mutex // only one caller to forwarder at a time
|
||||
snk *muxrpc.ByteSink
|
||||
enc *json.Encoder
|
||||
}
|
||||
|
||||
func newForwarder(snk *muxrpc.ByteSink) *updateForwarder {
|
||||
func newEndpointsForwarder(snk *muxrpc.ByteSink) *endpointsJSONEncoder {
|
||||
enc := json.NewEncoder(snk)
|
||||
snk.SetEncoding(muxrpc.TypeJSON)
|
||||
return &updateForwarder{
|
||||
return &endpointsJSONEncoder{
|
||||
snk: snk,
|
||||
enc: enc,
|
||||
}
|
||||
}
|
||||
|
||||
func (uf *updateForwarder) Update(members []string) error {
|
||||
func (uf *endpointsJSONEncoder) Update(members []string) error {
|
||||
uf.mu.Lock()
|
||||
defer uf.mu.Unlock()
|
||||
return uf.enc.Encode(members)
|
||||
}
|
||||
|
||||
func (uf *updateForwarder) Close() error {
|
||||
func (uf *endpointsJSONEncoder) Close() error {
|
||||
uf.mu.Lock()
|
||||
defer uf.mu.Unlock()
|
||||
return uf.snk.Close()
|
||||
|
|
|
@ -0,0 +1,208 @@
|
|||
// SPDX-License-Identifier: MIT
|
||||
|
||||
package go_test
|
||||
|
||||
import (
|
||||
"context"
|
||||
"encoding/json"
|
||||
"io"
|
||||
"os"
|
||||
"path/filepath"
|
||||
"testing"
|
||||
"time"
|
||||
|
||||
"github.com/ssb-ngi-pointer/go-ssb-room/muxrpc/handlers/tunnel/server"
|
||||
"github.com/stretchr/testify/assert"
|
||||
"github.com/stretchr/testify/require"
|
||||
"go.cryptoscope.co/muxrpc/v2"
|
||||
refs "go.mindeco.de/ssb-refs"
|
||||
)
|
||||
|
||||
// this tests the new room.attendants call
|
||||
// basically the same test as endpoints_test
|
||||
func TestRoomAttendants(t *testing.T) {
|
||||
testInit(t)
|
||||
r := require.New(t)
|
||||
a := assert.New(t)
|
||||
|
||||
testPath := filepath.Join("testrun", t.Name())
|
||||
os.RemoveAll(testPath)
|
||||
|
||||
ctx := context.Background()
|
||||
ctx, cancel := context.WithCancel(ctx)
|
||||
t.Cleanup(cancel)
|
||||
|
||||
// create the roomsrv
|
||||
ts := makeNamedTestBot(t, "server", ctx, nil)
|
||||
ctx = ts.ctx
|
||||
|
||||
// three new clients, connected to server automaticaly
|
||||
alf := ts.makeTestClient("alf")
|
||||
bre := ts.makeTestClient("bre")
|
||||
carl := ts.makeTestClient("carl")
|
||||
|
||||
// start with carl
|
||||
// ===============
|
||||
var ok bool
|
||||
err := carl.Async(ctx, &ok, muxrpc.TypeJSON, muxrpc.Method{"room", "announce"})
|
||||
r.NoError(err)
|
||||
a.True(ok, "announce should be fine")
|
||||
|
||||
carlsSource, err := carl.Source(ctx, muxrpc.TypeJSON, muxrpc.Method{"room", "attendants"})
|
||||
r.NoError(err)
|
||||
t.Log("carl opened attendants")
|
||||
|
||||
// first message should be initial state
|
||||
a.True(carlsSource.Next(ctx))
|
||||
var initState server.AttendantsInitialState
|
||||
decodeJSONsrc(t, carlsSource, &initState)
|
||||
a.Equal("state", initState.Type)
|
||||
a.Len(initState.IDs, 1)
|
||||
a.True(initState.IDs[0].Equal(&carl.feed))
|
||||
|
||||
announcementsForCarl := make(announcements)
|
||||
go logAttendantsStream(ts, carlsSource, "carl", announcementsForCarl)
|
||||
time.Sleep(1 * time.Second) // give some time to process new events
|
||||
|
||||
a.Len(announcementsForCarl, 0, "none yet")
|
||||
|
||||
// let alf join the room
|
||||
// =====================
|
||||
err = alf.Async(ctx, &ok, muxrpc.TypeJSON, muxrpc.Method{"room", "announce"})
|
||||
r.NoError(err)
|
||||
a.True(ok, "announce should be fine")
|
||||
|
||||
alfsSource, err := alf.Source(ctx, muxrpc.TypeJSON, muxrpc.Method{"room", "attendants"})
|
||||
r.NoError(err)
|
||||
|
||||
// first message should be initial state
|
||||
a.True(alfsSource.Next(ctx))
|
||||
decodeJSONsrc(t, alfsSource, &initState)
|
||||
a.Equal("state", initState.Type)
|
||||
a.Len(initState.IDs, 2)
|
||||
assertListContains(t, initState.IDs, carl.feed)
|
||||
assertListContains(t, initState.IDs, alf.feed)
|
||||
|
||||
announcementsForAlf := make(announcements)
|
||||
go logAttendantsStream(ts, alfsSource, "alf", announcementsForAlf)
|
||||
time.Sleep(1 * time.Second) // give some time to process new events
|
||||
|
||||
// assert what alf saw
|
||||
var seen bool
|
||||
a.Len(announcementsForAlf, 0, "none yet")
|
||||
|
||||
// assert what carl saw
|
||||
_, seen = announcementsForCarl[alf.feed.Ref()]
|
||||
a.True(seen, "carl saw alf")
|
||||
|
||||
// let bre join the room
|
||||
err = bre.Async(ctx, &ok, muxrpc.TypeJSON, muxrpc.Method{"room", "announce"})
|
||||
r.NoError(err)
|
||||
a.True(ok, "announce should be fine")
|
||||
|
||||
bresSource, err := bre.Source(ctx, muxrpc.TypeJSON, muxrpc.Method{"room", "attendants"})
|
||||
r.NoError(err)
|
||||
|
||||
// first message should be initial state
|
||||
a.True(bresSource.Next(ctx))
|
||||
decodeJSONsrc(t, bresSource, &initState)
|
||||
a.Equal("state", initState.Type)
|
||||
a.Len(initState.IDs, 3)
|
||||
assertListContains(t, initState.IDs, alf.feed)
|
||||
assertListContains(t, initState.IDs, bre.feed)
|
||||
assertListContains(t, initState.IDs, carl.feed)
|
||||
|
||||
announcementsForBre := make(announcements)
|
||||
go logAttendantsStream(ts, bresSource, "bre", announcementsForBre)
|
||||
|
||||
time.Sleep(1 * time.Second) // give some time to process new events
|
||||
|
||||
a.Len(announcementsForBre, 0, "none yet")
|
||||
|
||||
// the two present people saw her
|
||||
_, seen = announcementsForAlf[bre.feed.Ref()]
|
||||
a.True(seen, "alf saw bre")
|
||||
|
||||
_, seen = announcementsForCarl[bre.feed.Ref()]
|
||||
a.True(seen, "carl saw alf")
|
||||
|
||||
// shutdown alf first
|
||||
alf.Terminate()
|
||||
|
||||
time.Sleep(1 * time.Second) // give some time to process new events
|
||||
|
||||
// bre and arl should have removed him
|
||||
|
||||
_, seen = announcementsForBre[alf.feed.Ref()]
|
||||
a.False(seen, "alf should be gone for bre")
|
||||
|
||||
_, seen = announcementsForCarl[alf.feed.Ref()]
|
||||
a.False(seen, "alf should be gone for carl")
|
||||
|
||||
// terminate server and the clients
|
||||
ts.srv.Shutdown()
|
||||
bre.Terminate()
|
||||
carl.Terminate()
|
||||
ts.srv.Close()
|
||||
|
||||
// wait for all muxrpc serve()s to exit
|
||||
r.NoError(ts.serveGroup.Wait())
|
||||
cancel()
|
||||
|
||||
}
|
||||
|
||||
func assertListContains(t *testing.T, lst []refs.FeedRef, who refs.FeedRef) {
|
||||
var found = false
|
||||
for _, feed := range lst {
|
||||
if feed.Equal(&who) {
|
||||
found = true
|
||||
}
|
||||
}
|
||||
if !found {
|
||||
t.Errorf("did not find %s in list of %d", who.ShortRef(), len(lst))
|
||||
}
|
||||
}
|
||||
|
||||
func decodeJSONsrc(t *testing.T, src *muxrpc.ByteSource, val interface{}) {
|
||||
err := src.Reader(func(rd io.Reader) error {
|
||||
return json.NewDecoder(rd).Decode(val)
|
||||
})
|
||||
if err != nil {
|
||||
t.Fatal(err)
|
||||
}
|
||||
}
|
||||
|
||||
// consume endpoint messaes and put each peer on the passed map
|
||||
func logAttendantsStream(ts *testSession, src *muxrpc.ByteSource, who string, a announcements) {
|
||||
var update server.AttendantsUpdate
|
||||
|
||||
for src.Next(ts.ctx) {
|
||||
body, err := src.Bytes()
|
||||
if err != nil {
|
||||
panic(err)
|
||||
}
|
||||
// ts.t.Log(who, "got body:", string(body))
|
||||
|
||||
err = json.Unmarshal(body, &update)
|
||||
if err != nil {
|
||||
panic(err)
|
||||
}
|
||||
ts.t.Log(who, "got an update:", update.Type, update.ID.ShortRef())
|
||||
|
||||
switch update.Type {
|
||||
case "joined":
|
||||
a[update.ID.Ref()] = struct{}{}
|
||||
case "left":
|
||||
delete(a, update.ID.Ref())
|
||||
default:
|
||||
ts.t.Fatalf("%s: unexpected update type: %v", who, update.Type)
|
||||
}
|
||||
}
|
||||
|
||||
if err := src.Err(); err != nil {
|
||||
ts.t.Log(who, "source errored: ", err)
|
||||
return
|
||||
}
|
||||
|
||||
ts.t.Log(who, "stream closed")
|
||||
}
|
|
@ -1,3 +1,5 @@
|
|||
// SPDX-License-Identifier: MIT
|
||||
|
||||
package go_test
|
||||
|
||||
import (
|
||||
|
@ -19,6 +21,8 @@ type announcements map[string]struct{}
|
|||
|
||||
// we will let three clients (alf, bre, crl) join and see that the endpoint output is as expected
|
||||
func TestEndpointClients(t *testing.T) {
|
||||
testInit(t)
|
||||
|
||||
r := require.New(t)
|
||||
a := assert.New(t)
|
||||
|
||||
|
@ -40,12 +44,12 @@ func TestEndpointClients(t *testing.T) {
|
|||
|
||||
// let carl join the room
|
||||
// carl wont announce to emulate manyverse
|
||||
carlEndpointsSerc, err := carl.Source(ctx, muxrpc.TypeJSON, muxrpc.Method{"tunnel", "endpoints"})
|
||||
carlEndpointsSrc, err := carl.Source(ctx, muxrpc.TypeJSON, muxrpc.Method{"tunnel", "endpoints"})
|
||||
r.NoError(err)
|
||||
t.Log("carl opened endpoints")
|
||||
|
||||
announcementsForCarl := make(announcements)
|
||||
go logEndpointsStream(ts, carlEndpointsSerc, "carl", announcementsForCarl)
|
||||
go logEndpointsStream(ts, carlEndpointsSrc, "carl", announcementsForCarl)
|
||||
time.Sleep(1 * time.Second) // give some time to process new events
|
||||
|
||||
_, seen := announcementsForCarl[carl.feed.Ref()]
|
||||
|
|
|
@ -1,3 +1,5 @@
|
|||
// SPDX-License-Identifier: MIT
|
||||
|
||||
package go_test
|
||||
|
||||
import (
|
||||
|
|
|
@ -50,13 +50,13 @@ func (s *Server) initHandlers() {
|
|||
mux.RegisterAsync(muxrpc.Method{"manifest"}, manifest)
|
||||
mux.RegisterAsync(muxrpc.Method{"whoami"}, whoami)
|
||||
|
||||
// register tunnel.connect etc twice (as tunnel.* and room.*)
|
||||
var method = muxrpc.Method{"tunnel"}
|
||||
tunnelHandler.Register(mux, method)
|
||||
// register old room v1 commands
|
||||
tunnelHandler.RegisterTunnel(mux)
|
||||
|
||||
method = muxrpc.Method{"room"}
|
||||
tunnelHandler.Register(mux, method)
|
||||
// register new room v2 commands
|
||||
tunnelHandler.RegisterRoom(mux)
|
||||
|
||||
var method = muxrpc.Method{"room"}
|
||||
mux.RegisterAsync(append(method, "registerAlias"), typemux.AsyncFunc(aliasHandler.Register))
|
||||
mux.RegisterAsync(append(method, "revokeAlias"), typemux.AsyncFunc(aliasHandler.Revoke))
|
||||
|
||||
|
|
|
@ -43,8 +43,8 @@ const manifest manifestHandler = `
|
|||
"announce": "sync",
|
||||
"leave": "sync",
|
||||
"connect": "duplex",
|
||||
"endpoints": "source",
|
||||
"isRoom": "async",
|
||||
"attendants": "source",
|
||||
"metadata": "async",
|
||||
"ping": "sync"
|
||||
},
|
||||
|
||||
|
|
|
@ -1,6 +1,7 @@
|
|||
package roomstate
|
||||
|
||||
import (
|
||||
"fmt"
|
||||
"sort"
|
||||
"sync"
|
||||
|
||||
|
@ -14,8 +15,11 @@ import (
|
|||
type Manager struct {
|
||||
logger kitlog.Logger
|
||||
|
||||
updater broadcasts.RoomChangeSink
|
||||
broadcaster *broadcasts.RoomChangeBroadcast
|
||||
endpointsUpdater broadcasts.EndpointsEmitter
|
||||
endpointsbroadcaster *broadcasts.EndpointsBroadcast
|
||||
|
||||
attendantsUpdater broadcasts.AttendantsEmitter
|
||||
attendantsbroadcaster *broadcasts.AttendantsBroadcast
|
||||
|
||||
roomMu *sync.Mutex
|
||||
room roomStateMap
|
||||
|
@ -24,7 +28,8 @@ type Manager struct {
|
|||
func NewManager(log kitlog.Logger) *Manager {
|
||||
var m Manager
|
||||
m.logger = log
|
||||
m.updater, m.broadcaster = broadcasts.NewRoomChanger()
|
||||
m.endpointsUpdater, m.endpointsbroadcaster = broadcasts.NewEndpointsEmitter()
|
||||
m.attendantsUpdater, m.attendantsbroadcaster = broadcasts.NewAttendantsEmitter()
|
||||
m.roomMu = new(sync.Mutex)
|
||||
m.room = make(roomStateMap)
|
||||
|
||||
|
@ -44,25 +49,46 @@ func (rsm roomStateMap) AsList() []string {
|
|||
return memberList
|
||||
}
|
||||
|
||||
// Register listens to changes to the room
|
||||
func (m *Manager) Register(sink broadcasts.RoomChangeSink) {
|
||||
m.broadcaster.Register(sink)
|
||||
func (m *Manager) RegisterLegacyEndpoints(sink broadcasts.EndpointsEmitter) {
|
||||
m.endpointsbroadcaster.Register(sink)
|
||||
}
|
||||
|
||||
// List just returns a list of feed references
|
||||
func (m *Manager) RegisterAttendantsUpdates(sink broadcasts.AttendantsEmitter) {
|
||||
m.attendantsbroadcaster.Register(sink)
|
||||
}
|
||||
|
||||
// List just returns a list of feed references as strings
|
||||
func (m *Manager) List() []string {
|
||||
m.roomMu.Lock()
|
||||
defer m.roomMu.Unlock()
|
||||
return m.room.AsList()
|
||||
}
|
||||
|
||||
func (m *Manager) ListAsRefs() []refs.FeedRef {
|
||||
m.roomMu.Lock()
|
||||
lst := m.room.AsList()
|
||||
m.roomMu.Unlock()
|
||||
|
||||
rlst := make([]refs.FeedRef, len(lst))
|
||||
for i, s := range lst {
|
||||
fr, err := refs.ParseFeedRef(s)
|
||||
if err != nil {
|
||||
panic(fmt.Errorf("invalid feed ref in room state: %d: %s", i, err))
|
||||
}
|
||||
rlst[i] = *fr
|
||||
}
|
||||
return rlst
|
||||
}
|
||||
|
||||
// AddEndpoint adds the endpoint to the room
|
||||
func (m *Manager) AddEndpoint(who refs.FeedRef, edp muxrpc.Endpoint) {
|
||||
m.roomMu.Lock()
|
||||
// add ref to to the room map
|
||||
m.room[who.Ref()] = edp
|
||||
// update all the connected tunnel.endpoints calls
|
||||
m.updater.Update(m.room.AsList())
|
||||
m.endpointsUpdater.Update(m.room.AsList())
|
||||
// update all the connected room.attendants calls
|
||||
m.attendantsUpdater.Joined(who)
|
||||
m.roomMu.Unlock()
|
||||
}
|
||||
|
||||
|
@ -72,7 +98,9 @@ func (m *Manager) Remove(who refs.FeedRef) {
|
|||
// remove ref from lobby
|
||||
delete(m.room, who.Ref())
|
||||
// update all the connected tunnel.endpoints calls
|
||||
m.updater.Update(m.room.AsList())
|
||||
m.endpointsUpdater.Update(m.room.AsList())
|
||||
// update all the connected room.attendants calls
|
||||
m.attendantsUpdater.Left(who)
|
||||
m.roomMu.Unlock()
|
||||
}
|
||||
|
||||
|
@ -88,7 +116,8 @@ func (m *Manager) AlreadyAdded(who refs.FeedRef, edp muxrpc.Endpoint) bool {
|
|||
m.room[who.Ref()] = edp
|
||||
|
||||
// update everyone
|
||||
m.updater.Update(m.room.AsList())
|
||||
m.endpointsUpdater.Update(m.room.AsList())
|
||||
m.attendantsUpdater.Joined(who)
|
||||
}
|
||||
|
||||
m.roomMu.Unlock()
|
||||
|
|
Loading…
Reference in New Issue