// Copyright 2017 The etcd Authors // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. package etcdserver import ( "bytes" "context" "encoding/json" "fmt" "io" "net/http" "strings" "time" pb "go.etcd.io/etcd/api/v3/etcdserverpb" "go.etcd.io/etcd/api/v3/v3rpc/rpctypes" "go.etcd.io/etcd/client/pkg/v3/types" "go.etcd.io/etcd/server/v3/storage/mvcc" "go.uber.org/zap" ) // CheckInitialHashKV compares initial hash values with its peers // before serving any peer/client traffic. Only mismatch when hashes // are different at requested revision, with same compact revision. func (s *EtcdServer) CheckInitialHashKV() error { if !s.Cfg.InitialCorruptCheck { return nil } lg := s.Logger() lg.Info( "starting initial corruption check", zap.String("local-member-id", s.MemberId().String()), zap.Duration("timeout", s.Cfg.ReqTimeout()), ) h, rev, crev, err := s.kv.HashByRev(0) if err != nil { return fmt.Errorf("%s failed to fetch hash (%v)", s.MemberId(), err) } peers := s.getPeerHashKVs(rev) mismatch := 0 for _, p := range peers { if p.resp != nil { peerID := types.ID(p.resp.Header.MemberId) fields := []zap.Field{ zap.String("local-member-id", s.MemberId().String()), zap.Int64("local-member-revision", rev), zap.Int64("local-member-compact-revision", crev), zap.Uint32("local-member-hash", h), zap.String("remote-peer-id", peerID.String()), zap.Strings("remote-peer-endpoints", p.eps), zap.Int64("remote-peer-revision", p.resp.Header.Revision), zap.Int64("remote-peer-compact-revision", p.resp.CompactRevision), zap.Uint32("remote-peer-hash", p.resp.Hash), } if h != p.resp.Hash { if crev == p.resp.CompactRevision { lg.Warn("found different hash values from remote peer", fields...) mismatch++ } else { lg.Warn("found different compact revision values from remote peer", fields...) } } continue } if p.err != nil { switch p.err { case rpctypes.ErrFutureRev: lg.Warn( "cannot fetch hash from slow remote peer", zap.String("local-member-id", s.MemberId().String()), zap.Int64("local-member-revision", rev), zap.Int64("local-member-compact-revision", crev), zap.Uint32("local-member-hash", h), zap.String("remote-peer-id", p.id.String()), zap.Strings("remote-peer-endpoints", p.eps), zap.Error(err), ) case rpctypes.ErrCompacted: lg.Warn( "cannot fetch hash from remote peer; local member is behind", zap.String("local-member-id", s.MemberId().String()), zap.Int64("local-member-revision", rev), zap.Int64("local-member-compact-revision", crev), zap.Uint32("local-member-hash", h), zap.String("remote-peer-id", p.id.String()), zap.Strings("remote-peer-endpoints", p.eps), zap.Error(err), ) } } } if mismatch > 0 { return fmt.Errorf("%s found data inconsistency with peers", s.MemberId()) } lg.Info( "initial corruption checking passed; no corruption", zap.String("local-member-id", s.MemberId().String()), ) return nil } func (s *EtcdServer) monitorKVHash() { t := s.Cfg.CorruptCheckTime if t == 0 { return } lg := s.Logger() lg.Info( "enabled corruption checking", zap.String("local-member-id", s.MemberId().String()), zap.Duration("interval", t), ) for { select { case <-s.stopping: return case <-time.After(t): } if !s.isLeader() { continue } if err := s.checkHashKV(); err != nil { lg.Warn("failed to check hash KV", zap.Error(err)) } } } func (s *EtcdServer) checkHashKV() error { lg := s.Logger() h, rev, crev, err := s.kv.HashByRev(0) if err != nil { return err } peers := s.getPeerHashKVs(rev) ctx, cancel := context.WithTimeout(context.Background(), s.Cfg.ReqTimeout()) err = s.linearizableReadNotify(ctx) cancel() if err != nil { return err } h2, rev2, crev2, err := s.kv.HashByRev(0) if err != nil { return err } alarmed := false mismatch := func(id uint64) { if alarmed { return } alarmed = true a := &pb.AlarmRequest{ MemberID: id, Action: pb.AlarmRequest_ACTIVATE, Alarm: pb.AlarmType_CORRUPT, } s.GoAttach(func() { s.raftRequest(s.ctx, pb.InternalRaftRequest{Alarm: a}) }) } if h2 != h && rev2 == rev && crev == crev2 { lg.Warn( "found hash mismatch", zap.Int64("revision-1", rev), zap.Int64("compact-revision-1", crev), zap.Uint32("hash-1", h), zap.Int64("revision-2", rev2), zap.Int64("compact-revision-2", crev2), zap.Uint32("hash-2", h2), ) mismatch(uint64(s.MemberId())) } checkedCount := 0 for _, p := range peers { if p.resp == nil { continue } checkedCount++ id := p.resp.Header.MemberId // leader expects follower's latest revision less than or equal to leader's if p.resp.Header.Revision > rev2 { lg.Warn( "revision from follower must be less than or equal to leader's", zap.Int64("leader-revision", rev2), zap.Int64("follower-revision", p.resp.Header.Revision), zap.String("follower-peer-id", types.ID(id).String()), ) mismatch(id) } // leader expects follower's latest compact revision less than or equal to leader's if p.resp.CompactRevision > crev2 { lg.Warn( "compact revision from follower must be less than or equal to leader's", zap.Int64("leader-compact-revision", crev2), zap.Int64("follower-compact-revision", p.resp.CompactRevision), zap.String("follower-peer-id", types.ID(id).String()), ) mismatch(id) } // follower's compact revision is leader's old one, then hashes must match if p.resp.CompactRevision == crev && p.resp.Hash != h { lg.Warn( "same compact revision then hashes must match", zap.Int64("leader-compact-revision", crev2), zap.Uint32("leader-hash", h), zap.Int64("follower-compact-revision", p.resp.CompactRevision), zap.Uint32("follower-hash", p.resp.Hash), zap.String("follower-peer-id", types.ID(id).String()), ) mismatch(id) } } lg.Info("finished peer corruption check", zap.Int("number-of-peers-checked", checkedCount)) return nil } type peerInfo struct { id types.ID eps []string } type peerHashKVResp struct { peerInfo resp *pb.HashKVResponse err error } func (s *EtcdServer) getPeerHashKVs(rev int64) []*peerHashKVResp { // TODO: handle the case when "s.cluster.Members" have not // been populated (e.g. no snapshot to load from disk) members := s.cluster.Members() peers := make([]peerInfo, 0, len(members)) for _, m := range members { if m.ID == s.MemberId() { continue } peers = append(peers, peerInfo{id: m.ID, eps: m.PeerURLs}) } lg := s.Logger() var resps []*peerHashKVResp for _, p := range peers { if len(p.eps) == 0 { continue } respsLen := len(resps) var lastErr error for _, ep := range p.eps { ctx, cancel := context.WithTimeout(context.Background(), s.Cfg.ReqTimeout()) resp, lastErr := s.getPeerHashKVHTTP(ctx, ep, rev) cancel() if lastErr == nil { resps = append(resps, &peerHashKVResp{peerInfo: p, resp: resp, err: nil}) break } lg.Warn( "failed hash kv request", zap.String("local-member-id", s.MemberId().String()), zap.Int64("requested-revision", rev), zap.String("remote-peer-endpoint", ep), zap.Error(lastErr), ) } // failed to get hashKV from all endpoints of this peer if respsLen == len(resps) { resps = append(resps, &peerHashKVResp{peerInfo: p, resp: nil, err: lastErr}) } } return resps } const PeerHashKVPath = "/members/hashkv" type hashKVHandler struct { lg *zap.Logger server *EtcdServer } func (s *EtcdServer) HashKVHandler() http.Handler { return &hashKVHandler{lg: s.Logger(), server: s} } func (h *hashKVHandler) ServeHTTP(w http.ResponseWriter, r *http.Request) { if r.Method != http.MethodGet { w.Header().Set("Allow", http.MethodGet) http.Error(w, "Method Not Allowed", http.StatusMethodNotAllowed) return } if r.URL.Path != PeerHashKVPath { http.Error(w, "bad path", http.StatusBadRequest) return } defer r.Body.Close() b, err := io.ReadAll(r.Body) if err != nil { http.Error(w, "error reading body", http.StatusBadRequest) return } req := &pb.HashKVRequest{} if err := json.Unmarshal(b, req); err != nil { h.lg.Warn("failed to unmarshal request", zap.Error(err)) http.Error(w, "error unmarshalling request", http.StatusBadRequest) return } hash, rev, compactRev, err := h.server.KV().HashByRev(req.Revision) if err != nil { h.lg.Warn( "failed to get hashKV", zap.Int64("requested-revision", req.Revision), zap.Error(err), ) http.Error(w, err.Error(), http.StatusBadRequest) return } resp := &pb.HashKVResponse{Header: &pb.ResponseHeader{Revision: rev}, Hash: hash, CompactRevision: compactRev} respBytes, err := json.Marshal(resp) if err != nil { h.lg.Warn("failed to marshal hashKV response", zap.Error(err)) http.Error(w, err.Error(), http.StatusInternalServerError) return } w.Header().Set("X-Etcd-Cluster-ID", h.server.Cluster().ID().String()) w.Header().Set("Content-Type", "application/json") w.Write(respBytes) } // getPeerHashKVHTTP fetch hash of kv store at the given rev via http call to the given url func (s *EtcdServer) getPeerHashKVHTTP(ctx context.Context, url string, rev int64) (*pb.HashKVResponse, error) { cc := &http.Client{Transport: s.peerRt} hashReq := &pb.HashKVRequest{Revision: rev} hashReqBytes, err := json.Marshal(hashReq) if err != nil { return nil, err } requestUrl := url + PeerHashKVPath req, err := http.NewRequest(http.MethodGet, requestUrl, bytes.NewReader(hashReqBytes)) if err != nil { return nil, err } req = req.WithContext(ctx) req.Header.Set("Content-Type", "application/json") req.Cancel = ctx.Done() resp, err := cc.Do(req) if err != nil { return nil, err } defer resp.Body.Close() b, err := io.ReadAll(resp.Body) if err != nil { return nil, err } if resp.StatusCode == http.StatusBadRequest { if strings.Contains(string(b), mvcc.ErrCompacted.Error()) { return nil, rpctypes.ErrCompacted } if strings.Contains(string(b), mvcc.ErrFutureRev.Error()) { return nil, rpctypes.ErrFutureRev } } if resp.StatusCode != http.StatusOK { return nil, fmt.Errorf("unknown error: %s", string(b)) } hashResp := &pb.HashKVResponse{} if err := json.Unmarshal(b, hashResp); err != nil { return nil, err } return hashResp, nil }