服务器之家

服务器之家 > 正文

logrus hook输出日志到本地磁盘的操作

时间:2021-02-02 01:02     来源/作者:焗个面包

logrus是go的一个日志框架,它最让人激动的应该是hook机制,可以在初始化时为logrus添加hook,logrus可以实现各种扩展功能,可以将日志输出到elasticsearch和activemq等中间件去,甚至可以输出到你的email和叮叮中去,不要问为为什么可以发现可以输入到叮叮中去,都是泪,手动笑哭!

言归正传,这里就简单的通过hook机制将文件输出到本地磁盘。

首先

go get github.com/sirupsen/logrus

然后

logrus和go lib里面一样有6个等级,可以直接调用

?
1
2
3
4
5
6
logrus.Debug("Useful debugging information.")
logrus.Info("Something noteworthy happened!")
logrus.Warn("You should probably take a look at this.")
logrus.Error("Something failed but I'm not quitting.")
logrus.Fatal("Bye.")  //log之后会调用os.Exit(1)
logrus.Panic("I'm bailing.")  //log之后会panic()

项目例子结构

logrus hook输出日志到本地磁盘的操作

main.go

?
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
package main
 
import (
 "fmt"
 "github.com/sirupsen/logrus"
 "logT/logS"
)
func main() {
  //创建一个hook,将日志存储路径输入进去
 hook := logS.NewHook("d:/log/golog.log")
 //加载hook之前打印日志
 logrus.WithField("file", "d:/log/golog.log").Info("New logrus hook err.")
 logrus.AddHook(hook)
 //加载hook之后打印日志
 logrus.WithFields(logrus.Fields{
 "animal": "walrus",
 }).Info("A walrus appears")
}

hook.go

不要看下面三个go文件代码很长,其实大多数都是固定代码,也就NewHook函数自己扩展定义就好

package logS

?
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
import (
 "fmt"
 "github.com/sirupsen/logrus"
 "os"
 "strings"
)
 
// Hook 写文件的Logrus Hook
type Hook struct {
 W LoggerInterface
}
 
func NewHook(file string) (f *Hook) {
 w := NewFileWriter()
 config := fmt.Sprintf(`{"filename":"%s","maxdays":7}`, file)
 err := w.Init(config)
 if err != nil {
 return nil
 }
 
 return &Hook{w}
}
 
// Fire 实现Hook的Fire接口
func (hook *Hook) Fire(entry *logrus.Entry) (err error) {
 message, err := getMessage(entry)
 if err != nil {
 fmt.Fprintf(os.Stderr, "Unable to read entry, %v", err)
 return err
 }
 switch entry.Level {
 case logrus.PanicLevel:
 fallthrough
 case logrus.FatalLevel:
 fallthrough
 case logrus.ErrorLevel:
 return hook.W.WriteMsg(fmt.Sprintf("[ERROR] %s", message), LevelError)
 case logrus.WarnLevel:
 return hook.W.WriteMsg(fmt.Sprintf("[WARN] %s", message), LevelWarn)
 case logrus.InfoLevel:
 return hook.W.WriteMsg(fmt.Sprintf("[INFO] %s", message), LevelInfo)
 case logrus.DebugLevel:
 return hook.W.WriteMsg(fmt.Sprintf("[DEBUG] %s", message), LevelDebug)
 default:
 return nil
 }
}
 
// Levels 实现Hook的Levels接口
func (hook *Hook) Levels() []logrus.Level {
 return []logrus.Level{
 logrus.PanicLevel,
 logrus.FatalLevel,
 logrus.ErrorLevel,
 logrus.WarnLevel,
 logrus.InfoLevel,
 logrus.DebugLevel,
 }
}
 
func getMessage(entry *logrus.Entry) (message string, err error) {
 message = message + fmt.Sprintf("%s ", entry.Message)
 file, lineNumber := GetCallerIgnoringLogMulti(2)
 if file != "" {
 sep := fmt.Sprintf("%s/src/", os.Getenv("GOPATH"))
 fileName := strings.Split(file, sep)
 if len(fileName) >= 2 {
  file = fileName[1]
 }
 }
 message = fmt.Sprintf("%s:%d ", file, lineNumber) + message
 
 for k, v := range entry.Data {
 message = message + fmt.Sprintf("%v:%v ", k, v)
 }
 return
}

caller.go

?
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
package logS
 
import (
 "runtime"
 "strings"
)
 
func GetCaller(callDepth int, suffixesToIgnore ...string) (file string, line int) {
 // bump by 1 to ignore the getCaller (this) stackframe
 callDepth++
outer:
 for {
 var ok bool
 _, file, line, ok = runtime.Caller(callDepth)
 if !ok {
  file = "???"
  line = 0
  break
 }
 
 for _, s := range suffixesToIgnore {
  if strings.HasSuffix(file, s) {
  callDepth++
  continue outer
  }
 }
 break
 }
 return
}
 
// GetCallerIgnoringLogMulti TODO
func GetCallerIgnoringLogMulti(callDepth int) (string, int) {
 // the +1 is to ignore this (getCallerIgnoringLogMulti) frame
 return GetCaller(callDepth+1, "logrus/hooks.go", "logrus/entry.go", "logrus/logger.go", "logrus/exported.go", "asm_amd64.s")
}

file.go

?
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
package logS
 
import (
 "encoding/json"
 "errors"
 "fmt"
 "io/ioutil"
 "log"
 "os"
 "path/filepath"
 "strings"
 "sync"
 "time"
)
 
// RFC5424 log message levels.
const (
 LevelError = iota
 LevelWarn
 LevelInfo
 LevelDebug
)
 
// LoggerInterface Logger接口
type LoggerInterface interface {
 Init(config string) error
 WriteMsg(msg string, level int) error
 Destroy()
 Flush()
}
 
// LogWriter implements LoggerInterface.
// It writes messages by lines limit, file size limit, or time frequency.
type LogWriter struct {
 *log.Logger
 mw *MuxWriter
 // The opened file
 Filename string `json:"filename"`
 
 Maxlines     int `json:"maxlines"`
 maxlinesCurlines int
 
 // Rotate at size
 Maxsize    int `json:"maxsize"`
 maxsizeCursize int
 
 // Rotate daily
 Daily     bool `json:"daily"`
 Maxdays    int64 `json:"maxdays"`
 dailyOpendate int
 
 Rotate bool `json:"rotate"`
 
 startLock sync.Mutex // Only one log can write to the file
 
 Level int `json:"level"`
}
 
// MuxWriter an *os.File writer with locker.
type MuxWriter struct {
 sync.Mutex
 fd *os.File
}
 
// write to os.File.
func (l *MuxWriter) Write(b []byte) (int, error) {
 l.Lock()
 defer l.Unlock()
 return l.fd.Write(b)
}
 
// SetFd set os.File in writer.
func (l *MuxWriter) SetFd(fd *os.File) {
 if l.fd != nil {
 _ = l.fd.Close()
 }
 l.fd = fd
}
 
// NewFileWriter create a FileLogWriter returning as LoggerInterface.
func NewFileWriter() LoggerInterface {
 w := &LogWriter{
 Filename: "",
 Maxlines: 1000000,
 Maxsize: 1 << 28, //256 MB
 Daily:  true,
 Maxdays: 7,
 Rotate:  true,
 Level:  LevelDebug,
 }
 // use MuxWriter instead direct use os.File for lock write when rotate
 w.mw = new(MuxWriter)
 // set MuxWriter as Logger's io.Writer
 w.Logger = log.New(w.mw, "", log.Ldate|log.Ltime)
 return w
}
 
// Init file logger with json config.
// jsonconfig like:
// {
// "filename":"logs/sample.log",
// "maxlines":10000,
// "maxsize":1<<30,
// "daily":true,
// "maxdays":15,
// "rotate":true
// }
func (w *LogWriter) Init(jsonconfig string) error {
 err := json.Unmarshal([]byte(jsonconfig), w)
 if err != nil {
 return err
 }
 if len(w.Filename) == 0 {
 return errors.New("jsonconfig must have filename")
 }
 err = w.startLogger()
 return err
}
 
// start file logger. create log file and set to locker-inside file writer.
func (w *LogWriter) startLogger() error {
 fd, err := w.createLogFile()
 if err != nil {
 return err
 }
 w.mw.SetFd(fd)
 err = w.initFd()
 if err != nil {
 return err
 }
 return nil
}
 
func (w *LogWriter) docheck(size int) {
 w.startLock.Lock()
 defer w.startLock.Unlock()
 if w.Rotate && ((w.Maxlines > 0 && w.maxlinesCurlines >= w.Maxlines) ||
 (w.Maxsize > 0 && w.maxsizeCursize >= w.Maxsize) ||
 (w.Daily && time.Now().Day() != w.dailyOpendate)) {
 if err := w.DoRotate(); err != nil {
  fmt.Fprintf(os.Stderr, "FileLogWriter(%q): %s\n", w.Filename, err)
  return
 }
 }
 w.maxlinesCurlines++
 w.maxsizeCursize += size
}
 
// WriteMsg write logger message into file.
func (w *LogWriter) WriteMsg(msg string, level int) error {
 if level > w.Level {
 return nil
 }
 n := 24 + len(msg) // 24 stand for the length "2013/06/23 21:00:22 [T] "
 w.docheck(n)
 w.Logger.Print(msg)
 return nil
}
 
func (w *LogWriter) createLogFile() (*os.File, error) {
 // Open the log file
 fd, err := os.OpenFile(w.Filename, os.O_WRONLY|os.O_APPEND|os.O_CREATE, 0660)
 return fd, err
}
 
func (w *LogWriter) initFd() error {
 fd := w.mw.fd
 finfo, err := fd.Stat()
 if err != nil {
 return fmt.Errorf("get stat err: %s", err)
 }
 w.maxsizeCursize = int(finfo.Size())
 w.dailyOpendate = time.Now().Day()
 if finfo.Size() > 0 {
 content, err := ioutil.ReadFile(w.Filename)
 if err != nil {
  return err
 }
 w.maxlinesCurlines = len(strings.Split(string(content), "\n"))
 } else {
 w.maxlinesCurlines = 0
 }
 return nil
}
 
// DoRotate means it need to write file in new file.
// new file name like xx.log.2013-01-01.2
func (w *LogWriter) DoRotate() error {
 _, err := os.Lstat(w.Filename)
 if err == nil { // file exists
 // Find the next available number
 num := 1
 fname := ""
 for ; err == nil && num <= 999; num++ {
  fname = w.Filename + fmt.Sprintf(".%s.%03d", time.Now().Format("2006-01-02"), num)
  _, err = os.Lstat(fname)
 }
 // return error if the last file checked still existed
 if err == nil {
  return fmt.Errorf("Rotate: Cannot find free log number to rename %s", w.Filename)
 }
 
 // block Logger's io.Writer
 w.mw.Lock()
 defer w.mw.Unlock()
 
 fd := w.mw.fd
 _ = fd.Close()
 
 // close fd before rename
 // Rename the file to its newfound home
 err = os.Rename(w.Filename, fname)
 if err != nil {
  return fmt.Errorf("Rotate: %s", err)
 }
 
 // re-start logger
 err = w.startLogger()
 if err != nil {
  return fmt.Errorf("Rotate StartLogger: %s", err)
 }
 
 go w.deleteOldLog()
 }
 
 return nil
}
 
func (w *LogWriter) deleteOldLog() {
 dir := filepath.Dir(w.Filename)
 _ = filepath.Walk(dir, func(path string, info os.FileInfo, err error) (returnErr error) {
 defer func() {
  if r := recover(); r != nil {
  returnErr = fmt.Errorf("Unable to delete old log '%s', error: %+v", path, r)
  fmt.Println(returnErr)
  }
 }()
 
 if !info.IsDir() && info.ModTime().Unix() < (time.Now().Unix()-60*60*24*w.Maxdays) {
  if strings.HasPrefix(filepath.Base(path), filepath.Base(w.Filename)) {
  _ = os.Remove(path)
  }
 }
 return
 })
}
 
// Destroy destroy file logger, close file writer.
func (w *LogWriter) Destroy() {
 _ = w.mw.fd.Close()
}
 
// Flush file logger.
// there are no buffering messages in file logger in memory.
// flush file means sync file from disk.
func (w *LogWriter) Flush() {
 _ = w.mw.fd.Sync()
}

补充知识:golang logrus自定义hook:日志切片hook、邮件警报hook、kafkahook

logrus Hook 分析

logrus hook 接口定义很简单。如下

?
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
package logrus
 
// A hook to be fired when logging on the logging levels returned from
// `Levels()` on your implementation of the interface. Note that this is not
// fired in a goroutine or a channel with workers, you should handle such
// functionality yourself if your call is non-blocking and you don't wish for
// the logging calls for levels returned from `Levels()` to block.
type Hook interface {
 Levels() []Level
 Fire(*Entry) error
}
 
// Internal type for storing the hooks on a logger instance.
type LevelHooks map[Level][]Hook
 
// Add a hook to an instance of logger. This is called with
// `log.Hooks.Add(new(MyHook))` where `MyHook` implements the `Hook` interface.
func (hooks LevelHooks) Add(hook Hook) {
 for _, level := range hook.Levels() {
 hooks[level] = append(hooks[level], hook)
 }
}
 
// Fire all the hooks for the passed level. Used by `entry.log` to fire
// appropriate hooks for a log entry.
func (hooks LevelHooks) Fire(level Level, entry *Entry) error {
 for _, hook := range hooks[level] {
 if err := hook.Fire(entry); err != nil {
  return err
 }
 }
 return nil
}

只需实现 该结构的接口。

?
1
2
3
4
type Hook interface {
 Levels() []Level
 Fire(*Entry) error
}

就会被logrus框架遍历调用已注册的 hook 的 Fire 方法

获取日志实例

?
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
// log_hook.go
package logger
 
import (
 "fmt"
 "github.com/sirupsen/logrus"
 "library/util/constant"
 "os"
)
 
//自实现 logrus hook
func getLogger(module string) *logrus.Logger {
 //实例化
 logger := logrus.New()
 //设置输出
 logger.Out = os.Stdout
 //设置日志级别
 logger.SetLevel(logrus.DebugLevel)
 //设置日志格式
 //自定writer就行, hook 交给 lfshook
 logger.AddHook(newLogrusHook(constant.GetLogPath(), module))
 
 logger.SetFormatter(&logrus.JSONFormatter{
 TimestampFormat:"2006-01-02 15:04:05",
 })
 return logger
}
 
//确保每次调用使用的文件都是唯一的。
func GetNewFieldLoggerContext(module,appField string) *logrus.Entry {
 logger:= getLogger(module)
 return logger.WithFields(logrus.Fields{
 "app": appField,
 })
}
 
//订阅 警告日志
func SubscribeLog(entry *logrus.Entry, subMap SubscribeMap) {
 logger := entry.Logger
 logger.AddHook(newSubScribeHook(subMap))
 fmt.Println("日志订阅成功")
}

constant.GetLogPath() 可以替换为自己的日志文件输出目录地址,比如我的mac上则是:/usr/local/log ,直接替换即可。

日志切片hook

代码

?
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
// writer.go
package logger
 
import (
 "fmt"
 "github.com/pkg/errors"
 "io"
 "library/util"
 "os"
 "path/filepath"
 "sync"
 "time"
)
 
type LogWriter struct {
 logDir       string //日志根目录地址。
 module       string //模块 名
  curFileName   string //当前被指定的filename
 curBaseFileName   string //在使用中的file
 turnCateDuration  time.Duration
 mutex      sync.RWMutex
 outFh      *os.File
}
 
func (w *LogWriter) Write(p []byte) (n int, err error) {
 w.mutex.Lock()
 defer w.mutex.Unlock()
 if out, err:= w.getWriter(); err!=nil {
 return 0, errors.New("failed to fetch target io.Writer")
 }else{
 return out.Write(p)
 }
}
 
func (w *LogWriter) getFileName() string {
 base := time.Now().Truncate(w.turnCateDuration)
 return fmt.Sprintf("%s/%s/%s_%s", w.logDir, base.Format("2006-01-02"), w.module, base.Format("15"))
}
 
func (w *LogWriter) getWriter()(io.Writer, error) {
 fileName := w.curBaseFileName
 //判断是否有新的文件名
 //会出现新的文件名
 baseFileName := w.getFileName()
 if baseFileName != fileName {
 fileName = baseFileName
 }
 
 dirname := filepath.Dir(fileName)
 if err := os.MkdirAll(dirname, 0755); err != nil {
 return nil, errors.Wrapf(err, "failed to create directory %s", dirname)
 }
 
 fileHandler, err := os.OpenFile(fileName, os.O_CREATE|os.O_APPEND|os.O_WRONLY, 0644)
 if err != nil {
 return nil, errors.Errorf("failed to open file %s", err)
 }
 w.outFh.Close()
 w.outFh = fileHandler
 w.curBaseFileName = fileName
 w.curFileName = fileName
 
 return fileHandler, nil
}
 
func New(logPath, module string, duration time.Duration) *LogWriter {
 return &LogWriter{
 logDir: logPath,
 module: module,
 turnCateDuration:duration,
 curFileName: "",
 curBaseFileName: "",
 }
}
?
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
// hook.go
package logger
 
import (
 "github.com/rifflock/lfshook"
 "github.com/sirupsen/logrus"
 "time"
)
func newLogrusHook(logPath, moduel string) logrus.Hook {
 logrus.SetLevel(logrus.WarnLevel)
 
 writer := New(logPath, moduel, time.Hour * 2)
 
 lfsHook := lfshook.NewHook(lfshook.WriterMap{
 logrus.DebugLevel: writer,
 logrus.InfoLevel: writer,
 logrus.WarnLevel: writer,
 logrus.ErrorLevel: writer,
 logrus.FatalLevel: writer,
 logrus.PanicLevel: writer,
 }, &logrus.TextFormatter{DisableColors: true})
 
 // writer 生成新的log文件类型 writer 在通过new hook函数 消费 fire 函数
 // writer 是实现了writer 接口的库,在日志调用write是做预处理
 return lfsHook
}

测试代码

?
1
2
3
4
func TestGetLogger(t *testing.T) {
 lg := GetNewFieldLoggerContext("test","d")
 lg.Logger.Info("????")
}

解析

logger实例持有了 自定义的 io.writer 结构体,在消费Fire函数时,会调用Write方法,此时通过Truncate时间切片函数逻辑判断需要写入的文件。或创建新的文件。

注: 文章提供的代码是按天切分文件夹的,文件夹内模块日志再按2小时切分。可自行替换成按模块切分。

邮件警报hook

代码

?
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
// subscribeHook.go
package logger
 
import (
 "fmt"
 "github.com/sirupsen/logrus"
 "library/email"
 "strings"
)
 
type SubscribeMap map[logrus.Level][]*email.Receiver
type SubscribeHook struct {
 subMap SubscribeMap
}
//此处可以自实现hook 目前使用三方hook
func(h *SubscribeHook)Levels() []logrus.Level{
 return logrus.AllLevels
}
 
func(h *SubscribeHook)Fire(entry *logrus.Entry) error{
 for level, receivers := range h.subMap {
 //命中 准备消费
 if level == entry.Level {
  if len(receivers) > 0 {
  email.SendEmail(receivers, fmt.Sprintf("%s:[系统日志警报]", entry.Level.String()),
   fmt.Sprintf("错误内容: %s",entry.Message))
  }
 }
 }
 return nil
}
func NewSubscribeMap(level logrus.Level, receiverStr string) SubscribeMap{
 subMap := SubscribeMap{}
 addressList := strings.Split(receiverStr,";")
 var receivers []*email.Receiver
 for _, address := range addressList {
 receivers = append(receivers, &email.Receiver{Email: address})
 }
 subMap[level] = receivers
 return subMap
}
func newSubScribeHook(subMap SubscribeMap) *SubscribeHook {
 return &SubscribeHook{subMap}
?
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
// email.go
package email
 
import (
 "fmt"
 "gopkg.in/gomail.v2"
 "regexp"
 "strconv"
)
 
type Sender struct {
 User   string
 Password string
 Host   string
 Port   int
 MailTo  []string
 Subject  string
 Content  string
}
 
type Receiver struct {
 Email  string
}
 
func (r *Receiver) Check() bool {
 pattern := `\w+([-+.]\w+)*@\w+([-.]\w+)*\.\w+([-.]\w+)*` //匹配电子邮箱
 reg := regexp.MustCompile(pattern)
 return reg.MatchString(r.Email)
}
 
func (s *Sender) clean (){
 
}
 
//检查 邮箱正确性
func (s *Sender)NewReceiver(email string) *Receiver {
 rec := &Receiver{Email:email}
 if rec.Check() {
 m.MailTo = []string{email}
 return rec
 }else{
 fmt.Printf("email check fail 【%s】\n", email)
 return nil
 }
}
func (s *Sender)NewReceivers(receivers []*Receiver) {
 for _, rec := range receivers {
 if rec.Check() {
  m.MailTo = append(m.MailTo, rec.Email)
 }else{
  fmt.Printf("email check fail 【%s】\n", rec.Email)
 }
 }
}
// 163邮箱 password 为开启smtp后给的秘钥
var m = Sender{User:"6666666@163.com", Password:"666666666", Host: "smtp.163.com", Port: 465}
 
func SendEmail(receivers []*Receiver,subject, content string){
 m.NewReceivers(receivers)
 m.Subject = subject
 m.Content = content
 
 e := gomail.NewMessage()
 e.SetHeader("From", e.FormatAddress(m.User, "hengsheng"))
 e.SetHeader("To", m.MailTo...)  //发送给多个用户
 e.SetHeader("Subject", m.Subject) //设置邮件主题
 e.SetBody("text/html", m.Content)  //设置邮件正文
 d := gomail.NewDialer(m.Host, m.Port, m.User, m.Password)
 err := d.DialAndSend(e)
 if err != nil {
 fmt.Printf("error 邮件发送错误! %s \n", err.Error())
 }
}

使用

同理在writer时 如果是错误日志则发送邮件。

?
1
2
3
4
5
o.logger = logger.GetNewFieldLoggerContext("test", "666")
if subscribeSocket {
 logger.SubscribeLog(o.Logger, logger.NewSubscribeMap(logrus.ErrorLevel, "a@163.com;b@163.com"))
 }
 // o 为实际结构体实例

kafkahook

?
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
// kafka hook
package logger
 
import (
 "github.com/sirupsen/logrus"
 "library/kafka"
 "library/util/constant"
)
 
type KafKaHook struct {
 kafkaProducer  *kafka.KafkaProducer
}
 
 
func(h *KafKaHook)Levels() []logrus.Level{
 return logrus.AllLevels
}
 
func(h *KafKaHook)Fire(entry *logrus.Entry) error{
 h.kafkaProducer.SendMsgSync(entry.Message)
 return nil
}
 
func newKafkaHook() *KafKaHook{
 producer := kafka.NewKafkaProducer(constant.KafkaLogElkTopic,true)
 return &KafKaHook{kafkaProducer: producer}
}

使用时logger.AddHook(newKafkaHook()) 即可

kafka模块

生产者

?
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
// kafkaProducer.go
package kafka
 
import (
 "errors"
 "fmt"
 "github.com/Shopify/sarama"
 "library/util/constant"
 "log"
 "time"
)
 
func GetKafkaAddress()[]string{
 return "127.0.0.1:9092"
}
 
//同步消息模式
func SyncProducer(topic, message string) error {
 config := sarama.NewConfig()
 config.Producer.Return.Successes = true
 config.Producer.Timeout = 5 * time.Second
 p, err := sarama.NewSyncProducer(GetKafkaAddress(), config)
 if err != nil {
 return errors.New(fmt.Sprintf("sarama.NewSyncProducer err, message=%s \n", err))
 }
 defer p.Close()
 msg := &sarama.ProducerMessage{
 Topic: topic,
 Value: sarama.ByteEncoder(message),
 }
 part, offset, err := p.SendMessage(msg)
 if err != nil {
 return errors.New(fmt.Sprintf("send sdsds err=%s \n", err))
 } else {
 fmt.Printf("发送成功,partition=%d, offset=%d \n", part, offset)
 return nil
 }
}
 
//async 异步生产者
type KafkaProducer struct {
 topic    string
 asyncProducer  *sarama.AsyncProducer
 syncProducer  *sarama.SyncProducer
 sync    bool
}
 
func NewKafkaProducer(topic string, sync bool) *KafkaProducer {
 k := &KafkaProducer{
 topic:   topic,
 sync:   sync,
 }
 if sync {
 k.initSync()
 }else{
 k.initAsync()
 }
 return k
}
 
func (k *KafkaProducer) initAsync() bool {
 if k.sync {
 fmt.Printf("sync producer cant call async func !\n")
 return false
 }
 config := sarama.NewConfig()
 //等待服务器所有副本都保存成功后的响应
 config.Producer.RequiredAcks = sarama.WaitForAll
 //随机向partition发送消息
 config.Producer.Partitioner = sarama.NewRandomPartitioner
 //是否等待成功和失败后的响应,只有上面的RequireAcks设置不是NoReponse这里才有用.
 config.Producer.Return.Successes = true
 config.Producer.Return.Errors = true
 //设置使用的kafka版本,如果低于V0_10_0_0版本,消息中的timestrap没有作用.需要消费和生产同时配置
 //注意,版本设置不对的话,kafka会返回很奇怪的错误,并且无法成功发送消息
 config.Version = sarama.V0_10_0_1
 
 producer, e := sarama.NewAsyncProducer(GetKafkaAddress(), config)
 if e != nil {
 fmt.Println(e)
 return false
 }
 k.asyncProducer = &producer
 defer producer.AsyncClose()
 pd := *k.asyncProducer
 go func() {
 for{
  select {
  case <-pd.Successes():
  //fmt.Println("offset: ", suc.Offset, "timestamp: ", suc.Timestamp.String(), "partitions: ", suc.Partition)
  case fail := <-pd.Errors():
  fmt.Printf("err: %s \n", fail.Err.Error())
  }
 }
 }()
 
 return true
}
 
func (k *KafkaProducer) initSync() bool {
 if !k.sync {
 fmt.Println("async producer cant call sync func !")
 return false
 }
 
 config := sarama.NewConfig()
 config.Producer.Return.Successes = true
 config.Producer.Timeout = 5 * time.Second
 p, err := sarama.NewSyncProducer(GetKafkaAddress(), config)
 k.syncProducer = &p
 if err != nil {
 log.Printf("sarama.NewSyncProducer err, message=%s \n", err)
 return false
 }
 return true
}
 
func (k *KafkaProducer) SendMsgAsync(sendStr string) {
 
 msg := &sarama.ProducerMessage{
 Topic: k.topic,
 }
 
 //将字符串转化为字节数组
 msg.Value = sarama.ByteEncoder(sendStr)
 //fmt.Println(value)
 
 //使用通道发送
 pd := *k.asyncProducer
 pd.Input() <- msg
}
 
func (k *KafkaProducer) SendMsgSync(sendStr string) bool {
 msg := &sarama.ProducerMessage{
 Topic: k.topic,
 Value: sarama.ByteEncoder(sendStr),
 }
 pd := *k.syncProducer
 part, offset, err := pd.SendMessage(msg)
 if err != nil {
 fmt.Printf("发送失败 send message(%s) err=%s \n", sendStr, err)
 return false
 } else {
 fmt.Printf("发送成功 partition=%d, offset=%d \n", part, offset)
 return true
 }
}

调用 SendMsgSync 或 SendMsgAsync 生产消息,注意初始化时的参数要保证一致!

消费者组

?
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
// kafkaConsumerGroup.go
 
package kafka
 
import (
 "context"
 "fmt"
 "github.com/Shopify/sarama"
 "log"
 "sync"
)
 
func NewKafkaConsumerGroup(topics []string, group string, businessCall func(message *sarama.ConsumerMessage) bool) *KafkaConsumerGroup {
 k := &KafkaConsumerGroup{
 brokers:  GetKafkaAddress(),
 topics:  topics,
 group:       group,
 channelBufferSize: 2,
 ready:       make(chan bool),
 version:  "1.1.1",
 handler:  businessCall,
 }
 k.Init()
 return k
}
 
// 消费者组(consumer group): 相同的group.id的消费者将视为同一个消费者组,
// 每个消费者都需要设置一个组id, 每条消息只能被 consumer group 中的一个
// Consumer 消费,但可以被多个 consumer group 消费
type KafkaConsumerGroup struct {
 //代理(broker): 一台kafka服务器称之为一个broker
 brokers   []string
 //主题(topic): 消息的一种逻辑分组,用于对消息分门别类,每一类消息称之为一个主题,相同主题的消息放在一个队列中
 topics    []string
 version   string
 ready       chan bool
 group       string
 channelBufferSize  int
 //业务调用
 handler     func(message *sarama.ConsumerMessage) bool
}
 
func (k *KafkaConsumerGroup)Init() func() {
 
 version,err := sarama.ParseKafkaVersion(k.version)
 if err!=nil{
 fmt.Printf("Error parsing Kafka version: %v", err)
 }
 cfg := sarama.NewConfig()
 cfg.Version = version
 // 分区分配策略
 cfg.Consumer.Group.Rebalance.Strategy = sarama.BalanceStrategyRange
 // 未找到组消费位移的时候从哪边开始消费
 cfg.Consumer.Offsets.Initial = -2
 // channel长度
 cfg.ChannelBufferSize = k.channelBufferSize
 ctx, cancel := context.WithCancel(context.Background())
 client, err := sarama.NewConsumerGroup(k.brokers, k.group, cfg)
 if err != nil {
  fmt.Printf("Error creating consumer group client: %v", err)
 }
 
 wg := &sync.WaitGroup{}
 wg.Add(1)
 go func() {
  defer func() {
  wg.Done()
  //util.HandlePanic("client.Consume panic", log.StandardLogger())
  }()
  for {
  if err := client.Consume(ctx, k.topics, k); err != nil {
   log.Printf("Error from consumer: %v", err)
  }
  // check if context was cancelled, signaling that the consumer should stop
  if ctx.Err() != nil {
   log.Println(ctx.Err())
   return
  }
  k.ready = make(chan bool)
  }
 }()
 
 <-k.ready
 fmt.Printf("Sarama consumer up and running!... \n")
 // 保证在系统退出时,通道里面的消息被消费
 return func() {
  cancel()
  wg.Wait()
  if err = client.Close(); err != nil {
  fmt.Printf("Error closing client: %v \n", err)
  }
 }
 
}
 
// Setup is run at the beginning of a new session, before ConsumeClaim
func (k *KafkaConsumerGroup) Setup(sarama.ConsumerGroupSession) error {
 // Mark the consumer as ready
 close(k.ready)
 return nil
}
 
// Cleanup is run at the end of a session, once all ConsumeClaim goroutines have exited
func (k *KafkaConsumerGroup) Cleanup(sarama.ConsumerGroupSession) error {
 return nil
}
 
// ConsumeClaim must start a consumer loop of ConsumerGroupClaim's Messages().
func (k *KafkaConsumerGroup) ConsumeClaim(session sarama.ConsumerGroupSession, claim sarama.ConsumerGroupClaim) error {
 
 // NOTE:
 // Do not move the code below to a goroutine.
 // The `ConsumeClaim` itself is called within a goroutine, see:
 // https://github.com/Shopify/sarama/blob/master/consumer_group.go#L27-L29
 // 具体消费消息
 for message := range claim.Messages() {
 //msg := string(message.Value)
 //k.logger.Infof("卡夫卡: %s", msg)
 
 if ok:= k.handler(message); ok {
  // 更新位移
  session.MarkMessage(message, "")
 }
 //run.Run(msg)
 }
 return nil
}

测试代码

?
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
func TestKafkaConsumerGroup_Init(t *testing.T) {
 //pd := NewKafkaProducer("test-fail",true)
 //pd.InitSync()
 k := NewKafkaConsumerGroup([]string{constant.KafkaALiSdkTopic}, "group-2", func(message *sarama.ConsumerMessage) bool {
 fmt.Println(string(message.Value))
 //如果失败的处理逻辑
 //if ok := pd.SendMsgSync("666666"); ok {
 // return true
 //}
 return false
 
 })
 consumerDone := k.Init()
 
 sigterm := make(chan os.Signal, 1)
 signal.Notify(sigterm, syscall.SIGINT, syscall.SIGTERM)
 select {
 case <-sigterm:
 fmt.Println("terminating: via signal")
 }
 consumerDone()
}

这里有一些补偿逻辑在里面。

以上就是logrus相关hook。

好了,这篇logrus hook输出日志到本地磁盘的操作就是小编分享给大家的全部内容了,希望能给大家一个参考,也希望大家多多支持服务器之家。

原文链接:https://noshoes.blog.csdn.net/article/details/82909121

标签:

相关文章

热门资讯

2020微信伤感网名听哭了 让对方看到心疼的伤感网名大全
2020微信伤感网名听哭了 让对方看到心疼的伤感网名大全 2019-12-26
Intellij idea2020永久破解,亲测可用!!!
Intellij idea2020永久破解,亲测可用!!! 2020-07-29
背刺什么意思 网络词语背刺是什么梗
背刺什么意思 网络词语背刺是什么梗 2020-05-22
苹果12mini价格表官网报价 iPhone12mini全版本价格汇总
苹果12mini价格表官网报价 iPhone12mini全版本价格汇总 2020-11-13
yue是什么意思 网络流行语yue了是什么梗
yue是什么意思 网络流行语yue了是什么梗 2020-10-11
返回顶部