本文目的是在于对netty有一个大体印象,便于深刻了解netty前有一些基本概念java
sequenceDiagram Channel->>Buffer: 读取 Buffer->>程序:读取 程序->>Buffer:写入 Buffer->>Channel: 写入
- 在NIO中,全部的数据都是从Channel中读取并缓存到Buffer中,用户本身的代码再从Buffer中读取
- 要想写数据,则必须先写入Buffer中,而后在把Buffer写入到Channnel中。
Channel 一个链接的抽象。全部的读写数据,最终都是经过Channel流通的。主要有NioSocketChannel、NioServerSocketChannelbootstrap
netty本身实现的一个缓存Buf,相比于JDK的ByteBuffer有一下优势缓存
- 长度能够动态拓展(在写入的时候,判断到capacity不够,会从新开辟一段大容量的buf,而后把以前buf中的数据拷贝到新的buf中)
- 操做简单。JDK的ByteBuffer只有一个位置指针position。netty的ByteBuf则有readIndex(度位置)、writeIndex(写位置)。在对buf操做时候,不须要flip()和rewind();操做简单。
ChannelHandler 从Channel中读取到数据后,就须要把数据交给ChannelHandler进行处理app
记录当前ChannelHandler的环境上下文,大体有如下信息。 每个ChannelHandler都会有一个ChannelHandlerContext与之对应(一对一关系)框架
- 记录当前ChannelHandler对象
- 标识是inbound仍是outbound
- 所属的ChannelPipeline
- 前一个ChannelHandlerContext、后一个ChannelHandlerContext。 这样造成一个处理链,相似于SpringMVC中的拦截器链。
netty是事件驱动的。在获取到不一样的事件后(数据),会作不一样的业务逻辑处理,这时候有的可能须要多个Handler协做完成,有的Handler可能对当前的事件不作关心,有的可能处理完了,不想要后面的Handler处理了。<br/> 这时候如何对事件进行传播处理,这时候就须要用到ChannelPipeline了。ChannelPipeline中保存了头部的ChannelHandlerContext(进来的类型事件会从头开始)和尾部的ChannelHandlerContext(出去的类型事件会从尾部开始),他们是一个串行连接。socket
![]()
![]()
Channel和Channel之间都是经过ChannelHandlerContext联系起来的。ide
具体内容这篇博客讲的很详细oop
package io.netty.example.discard; import io.netty.bootstrap.ServerBootstrap; import io.netty.channel.ChannelFuture; import io.netty.channel.ChannelInitializer; import io.netty.channel.ChannelOption; import io.netty.channel.EventLoopGroup; import io.netty.channel.nio.NioEventLoopGroup; import io.netty.channel.socket.SocketChannel; import io.netty.channel.socket.nio.NioServerSocketChannel; /** * Discards any incoming data. */ public class DiscardServer { private int port; public DiscardServer(int port) { this.port = port; } public void run() throws Exception { EventLoopGroup bossGroup = new NioEventLoopGroup(); // (1) EventLoopGroup workerGroup = new NioEventLoopGroup(); try { ServerBootstrap b = new ServerBootstrap(); // (2) b.group(bossGroup, workerGroup) .channel(NioServerSocketChannel.class) // (3) .childHandler(new ChannelInitializer<SocketChannel>() { // (4) @Override public void initChannel(SocketChannel ch) throws Exception { ch.pipeline().addLast(new DiscardServerHandler()); } }) .option(ChannelOption.SO_BACKLOG, 128) // (5) .childOption(ChannelOption.SO_KEEPALIVE, true); // (6) // Bind and start to accept incoming connections. ChannelFuture f = b.bind(port).sync(); // (7) // Wait until the server socket is closed. // In this example, this does not happen, but you can do that to gracefully // shut down your server. f.channel().closeFuture().sync(); } finally { workerGroup.shutdownGracefully(); bossGroup.shutdownGracefully(); } } public static void main(String[] args) throws Exception { int port; if (args.length > 0) { port = Integer.parseInt(args[0]); } else { port = 8080; } new DiscardServer(port).run(); } }
package io.netty.example.discard; import io.netty.buffer.ByteBuf; import io.netty.channel.ChannelHandlerContext; import io.netty.channel.ChannelInboundHandlerAdapter; /** * Handles a server-side channel. */ public class DiscardServerHandler extends ChannelInboundHandlerAdapter { // (1) @Override public void channelRead(ChannelHandlerContext ctx, Object msg) { // (2) try { // Do something with msg } finally { ReferenceCountUtil.release(msg); } } @Override public void exceptionCaught(ChannelHandlerContext ctx, Throwable cause) { // (4) // Close the connection when an exception is raised. cause.printStackTrace(); ctx.close(); } }
package io.netty.example.time; public class TimeServerHandler extends ChannelInboundHandlerAdapter { @Override public void channelActive(final ChannelHandlerContext ctx) { // (1) final ByteBuf time = ctx.alloc().buffer(4); // (2) time.writeInt((int) (System.currentTimeMillis() / 1000L + 2208988800L)); final ChannelFuture f = ctx.writeAndFlush(time); // (3) f.addListener(new ChannelFutureListener() { @Override public void operationComplete(ChannelFuture future) { assert f == future; ctx.close(); } }); // (4) } @Override public void exceptionCaught(ChannelHandlerContext ctx, Throwable cause) { cause.printStackTrace(); ctx.close(); } }
package io.netty.example.time; public class TimeClient { public static void main(String[] args) throws Exception { String host = args[0]; int port = Integer.parseInt(args[1]); EventLoopGroup workerGroup = new NioEventLoopGroup(); try { Bootstrap b = new Bootstrap(); // (1) b.group(workerGroup); // (2) b.channel(NioSocketChannel.class); // (3) b.option(ChannelOption.SO_KEEPALIVE, true); // (4) b.handler(new ChannelInitializer<SocketChannel>() { @Override public void initChannel(SocketChannel ch) throws Exception { ch.pipeline().addLast(new TimeClientHandler()); } }); // Start the client. ChannelFuture f = b.connect(host, port).sync(); // (5) // Wait until the connection is closed. f.channel().closeFuture().sync(); } finally { workerGroup.shutdownGracefully(); } } }
package io.netty.example.time; import java.util.Date; public class TimeClientHandler extends ChannelInboundHandlerAdapter { @Override public void channelRead(ChannelHandlerContext ctx, Object msg) { ByteBuf m = (ByteBuf) msg; // (1) try { long currentTimeMillis = (m.readUnsignedInt() - 2208988800L) * 1000L; System.out.println(new Date(currentTimeMillis)); ctx.close(); } finally { m.release(); } } @Override public void exceptionCaught(ChannelHandlerContext ctx, Throwable cause) { cause.printStackTrace(); ctx.close(); } }