public static void logMethodCall(Object object, String methodName, Object[] arguments) { String className = getClassName(object); String logname = "methodCalls." + className + "." + methodName; InternalLogger objLog = InternalLoggerFactory.getInstance(logname); if (!objLog.isTraceEnabled()) return; StringBuilder msg = new StringBuilder(methodName); msg.append("("); if (arguments != null) { for (int i = 0; i < arguments.length;) { msg.append(normalizedValue(arguments[i])); if (++i < arguments.length) { msg.append(","); } } } msg.append(")"); objLog.log(InternalLogLevel.TRACE, className, msg.toString(), "called from MetaClass.invokeMethod"); }
/** * Init timeouts and the connection registry and start the netty IO server synchronously */ @Override public void init(Container container) { super.init(container); try { // Configure netty InternalLoggerFactory.setDefaultFactory(new Slf4JLoggerFactory() { @Override public InternalLogger newInstance(String name) { return new NettyInternalLogger(name); } }); ResourceLeakDetector.setLevel(CoreConstants.NettyConstants.RESOURCE_LEAK_DETECTION); // Start server startServer(); } catch (InterruptedException e) { throw new StartupException("Could not start netty server", e); } }
/** * Configure netty and initialize related Components. * Afterwards call {@link #initClient()} method to start the netty IO client asynchronously. */ @Override public void init(Container container) { super.init(container); // Configure netty InternalLoggerFactory.setDefaultFactory(new Slf4JLoggerFactory() { @Override public InternalLogger newInstance(String name) { return new NettyInternalLogger(name); } }); ResourceLeakDetector.setLevel(CoreConstants.NettyConstants.RESOURCE_LEAK_DETECTION); // And try to connect isActive = true; initClient(); // register BroadcastListener IntentFilter filter = new IntentFilter(); filter.addAction(WifiManager.SUPPLICANT_CONNECTION_CHANGE_ACTION); filter.addAction(WifiManager.NETWORK_STATE_CHANGED_ACTION); filter.addAction(WifiManager.WIFI_STATE_CHANGED_ACTION); filter.addAction(ConnectivityManager.CONNECTIVITY_ACTION); requireComponent(ContainerService.KEY_CONTEXT).registerReceiver(broadcastReceiver, filter); }
/** * Main entry point * @param args None for now */ public static void main(String[] args) { log.info("TSDBLite booting...."); ExtendedThreadManager.install(); InternalLoggerFactory .setDefaultFactory(Slf4JLoggerFactory.INSTANCE); final String jmxmpIface = ConfigurationHelper.getSystemThenEnvProperty(Constants.CONF_JMXMP_IFACE, Constants.DEFAULT_JMXMP_IFACE); final int jmxmpPort = ConfigurationHelper.getIntSystemThenEnvProperty(Constants.CONF_JMXMP_PORT, Constants.DEFAULT_JMXMP_PORT); JMXHelper.fireUpJMXMPServer(jmxmpIface, jmxmpPort, JMXHelper.getHeliosMBeanServer()); server = Server.getInstance(); final Thread mainThread = Thread.currentThread(); StdInCommandHandler.getInstance().registerCommand("stop", new Runnable(){ @Override public void run() { if(server!=null) { log.info("Stopping TSDBLite Server....."); server.stop(); log.info("TSDBLite Server Stopped. Bye."); mainThread.interrupt(); } } }); try { Thread.currentThread().join(); } catch (Exception x) {/* No Op */} }
public void retain() { if (counter.incrementAndGet() == 1) { if (instanceCounter.getAndIncrement() > 0) { InternalLogger instance = InternalLoggerFactory.getInstance(getClass()); instance.info("Initialized PauseDetectorWrapper more than once."); } pauseDetector = new SimplePauseDetector(TimeUnit.MILLISECONDS.toNanos(10), TimeUnit.MILLISECONDS.toNanos(10), 3); Runtime.getRuntime().addShutdownHook(new Thread("ShutdownHook for SimplePauseDetector") { @Override public void run() { if (pauseDetector != null) { pauseDetector.shutdown(); } } }); } }
public void init(Container container, ServerBundleConfiguration config) { logger.info("Initializing the container"); // Override the supplied one ServerConfiguration configuration = container.getConfiguration().getServerConfiguration(); AbstractHttpConnector connector = null; InternalLoggerFactory.setDefaultFactory(new Slf4JLoggerFactory()); logger.info("Loading the http connectors"); for (ConnectorConfiguration connectorConfig : configuration.getConnectorConfigurations()) { if (connectorConfig.getScheme() == Scheme.https) { connector = createHttpsConnector(connectorConfig, container.getRouter()); } else { connector = createHttpConnector(connectorConfig, container.getRouter()); } connector.registerListener(container.getMessageObserver()); connector.initialize(); connectors.add(connector); } }
public void start() { Configuration config = Configuration.INSTANCE; InternalLoggerFactory.setDefaultFactory(Slf4JLoggerFactory.INSTANCE); bossGroup = new NioEventLoopGroup(1); workerGroup = new NioEventLoopGroup(); try { ServerBootstrap bootstrap = new ServerBootstrap(); bootstrap.group(bossGroup, workerGroup) .channel(NioServerSocketChannel.class) .childHandler(new ChannelInitializer<SocketChannel>() { @Override protected void initChannel(SocketChannel socketChannel) throws Exception { socketChannel.pipeline() .addLast("logging", new LoggingHandler(LogLevel.DEBUG)) .addLast(new SocksInitRequestDecoder()) .addLast(new SocksMessageEncoder()) .addLast(new Socks5Handler()) .addLast(Status.TRAFFIC_HANDLER); } }); log.info("\tStartup {}-{}-client [{}{}]", Constants.APP_NAME, Constants.APP_VERSION, config.getMode(), config.getMode().equals("socks5") ? "" : ":" + config.getProtocol()); new Thread(() -> new UdpServer().start()).start(); ChannelFuture future = bootstrap.bind(config.getLocalHost(), config.getLocalPort()).sync(); future.addListener(future1 -> log.info("\tTCP listening at {}:{}...", config.getLocalHost(), config.getLocalPort())); future.channel().closeFuture().sync(); } catch (Exception e) { log.error("\tSocket bind failure ({})", e.getMessage()); } finally { log.info("\tShutting down"); bossGroup.shutdownGracefully(); workerGroup.shutdownGracefully(); } }
public void start() { Configuration config = Configuration.INSTANCE; InternalLoggerFactory.setDefaultFactory(Slf4JLoggerFactory.INSTANCE); EventLoopGroup bossGroup = new NioEventLoopGroup(1); EventLoopGroup workerGroup = new NioEventLoopGroup(); try { ServerBootstrap bootstrap = new ServerBootstrap(); bootstrap.group(bossGroup, workerGroup) .channel(NioServerSocketChannel.class) .childHandler(new ChannelInitializer<SocketChannel>() { protected void initChannel(SocketChannel socketChannel) throws Exception { socketChannel.pipeline() .addLast("logging", new LoggingHandler(LogLevel.DEBUG)) .addLast(new XConnectHandler()); if (config.getReadLimit() != 0 || config.getWriteLimit() != 0) { socketChannel.pipeline().addLast( new GlobalTrafficShapingHandler(Executors.newScheduledThreadPool(1), config.getWriteLimit(), config.getReadLimit()) ); } } }); log.info("\tStartup {}-{}-server [{}]", Constants.APP_NAME, Constants.APP_VERSION, config.getProtocol()); new Thread(() -> new UdpServer().start()).start(); ChannelFuture future = bootstrap.bind(config.getHost(), config.getPort()).sync(); future.addListener(future1 -> log.info("\tTCP listening at {}:{}...", config.getHost(), config.getPort())); future.channel().closeFuture().sync(); } catch (Exception e) { log.error("\tSocket bind failure ({})", e.getMessage()); } finally { log.info("\tShutting down and recycling..."); bossGroup.shutdownGracefully(); workerGroup.shutdownGracefully(); Configuration.shutdownRelays(); } System.exit(0); }
public static void setAhessianLogger(final Logger log) { InternalLoggerFactory.setDefaultFactory(new InternalLoggerFactory() { @Override public InternalLogger newInstance(String name) { return (InternalLogger) new JdkLogger(log, "ahessian-jmx"); } }); }
/** * Starts the server. * @throws InterruptedException */ public static void start() { InternalLoggerFactory.setDefaultFactory(new Slf4JLoggerFactory()); EventLoopGroup group = new NioEventLoopGroup(); try { Bootstrap boot = new Bootstrap(); boot.group(group) .channel(NioSocketChannel.class) .handler(new ChannelInitializer<SocketChannel>() { @Override public void initChannel(SocketChannel ch) throws Exception { ch.pipeline().addLast(new SortClientInitializer()); } }); LOG.info("Client connecting to {}:{}", ClientMain.SERVER_ADDRESS, ClientMain.SERVER_PORT); // Start the client. ChannelFuture f = boot.connect(new InetSocketAddress(ClientMain.SERVER_ADDRESS, ClientMain.SERVER_PORT)).sync(); // Wait until the connection is closed. f.channel().closeFuture().sync(); } catch (Exception e) { e.printStackTrace(); } finally { // The connection is closed automatically on shutdown. group.shutdownGracefully(); LOG.info("Client Exit."); } }
/** * Starts the server. * @throws InterruptedException */ public static void start() throws InterruptedException { InternalLoggerFactory.setDefaultFactory(new Slf4JLoggerFactory()); // configure the server EventLoopGroup bossGroup = new NioEventLoopGroup(1); EventLoopGroup workerGroup = new NioEventLoopGroup(); try { ServerBootstrap boot = new ServerBootstrap(); boot.group(bossGroup, workerGroup) .channel(NioServerSocketChannel.class) .childHandler(new ChannelInitializer<SocketChannel>() { @Override public void initChannel(SocketChannel sChannel) throws Exception { sChannel.pipeline().addLast(new SortServerInitializer()); } }) .option(ChannelOption.SO_BACKLOG, 128) .childOption(ChannelOption.SO_KEEPALIVE, true) .childOption(ChannelOption.TCP_NODELAY, true); // Start the server. ChannelFuture f = boot.bind(new InetSocketAddress(ServerMain.SERVER_ADDRESS, ServerMain.SERVER_PORT)).sync(); LOG.info("Server started at {}:{}, JobId: {}", ServerMain.SERVER_ADDRESS, ServerMain.SERVER_PORT, ServerMain.JOB_ID); // Wait until the server socket is closed. f.channel().closeFuture().sync(); } catch (Exception e) { e.printStackTrace(); } finally { // Shut down all event loops to terminate all threads. bossGroup.shutdownGracefully(); workerGroup.shutdownGracefully(); // Wait until all threads are terminated. bossGroup.terminationFuture().sync(); workerGroup.terminationFuture().sync(); LOG.info("Server Exit."); } }
/** * Creates a new instance of the PravegaConnectionListener class. * * @param ssl Whether to use SSL. * @param host The name of the host to listen to. * @param port The port to listen on. * @param streamSegmentStore The SegmentStore to delegate all requests to. * @param statsRecorder (Optional) A StatsRecorder for Metrics. */ public PravegaConnectionListener(boolean ssl, String host, int port, StreamSegmentStore streamSegmentStore, SegmentStatsRecorder statsRecorder) { this.ssl = ssl; this.host = Exceptions.checkNotNullOrEmpty(host, "host"); this.port = port; this.store = Preconditions.checkNotNull(streamSegmentStore, "streamSegmentStore"); this.statsRecorder = statsRecorder; InternalLoggerFactory.setDefaultFactory(Slf4JLoggerFactory.INSTANCE); }
@Before public void setup() throws Exception { originalLevel = ResourceLeakDetector.getLevel(); ResourceLeakDetector.setLevel(Level.PARANOID); InternalLoggerFactory.setDefaultFactory(Slf4JLoggerFactory.INSTANCE); this.serviceBuilder = ServiceBuilder.newInMemoryBuilder(ServiceBuilderConfig.getDefaultConfig()); this.serviceBuilder.initialize(); }
public SpdyFrameLogger(InternalLogLevel level) { if (level == null) { throw new NullPointerException("level"); } logger = InternalLoggerFactory.getInstance(getClass()); this.level = level; }
@Bean @Resource(name = "channelInitializer") public ServerBootstrap serverBootstrapFactory(ChannelInitializer<SocketChannel> channelInitializer) { // 配置服务器 EventLoopGroup bossGroup = new NioEventLoopGroup(); EventLoopGroup workerGroup = new NioEventLoopGroup(); InternalLoggerFactory.setDefaultFactory(new Slf4JLoggerFactory()); ServerBootstrap serverBootstrap = new ServerBootstrap(); serverBootstrap.group(bossGroup, workerGroup).channel(NioServerSocketChannel.class) .handler(new LoggingHandler(LogLevel.INFO)).childHandler(channelInitializer) .option(ChannelOption.SO_BACKLOG, 128).childOption(ChannelOption.SO_KEEPALIVE, true); return serverBootstrap; }
/** * Creates a new instance whose logger name is the fully qualified class * name of the instance. * * @param level the log level */ public LoggingHandler(LogLevel level) { if (level == null) { throw new NullPointerException("level"); } logger = InternalLoggerFactory.getInstance(getClass()); this.level = level; internalLevel = level.toInternalLevel(); }
/** * Creates a new instance with the specified logger name. * * @param level the log level */ public LoggingHandler(Class<?> clazz, LogLevel level) { if (clazz == null) { throw new NullPointerException("clazz"); } if (level == null) { throw new NullPointerException("level"); } logger = InternalLoggerFactory.getInstance(clazz); this.level = level; internalLevel = level.toInternalLevel(); }
/** * Creates a new instance with the specified logger name. * * @param level the log level */ public LoggingHandler(String name, LogLevel level) { if (name == null) { throw new NullPointerException("name"); } if (level == null) { throw new NullPointerException("level"); } logger = InternalLoggerFactory.getInstance(name); this.level = level; internalLevel = level.toInternalLevel(); }
@After public void tearDown() { final Logger logger = (Logger) LoggerFactory.getLogger(getClass()); final StatusManager sm = rootLogger.getLoggerContext().getStatusManager(); int count = 0; for (Status s : sm.getCopyOfStatusList()) { final int level = s.getEffectiveLevel(); if (level == Status.INFO) { continue; } if (s.getMessage().contains(InternalLoggerFactory.class.getName())) { // Skip the warnings related with Netty. continue; } count++; switch (level) { case Status.WARN: if (s.getThrowable() != null) { logger.warn(s.getMessage(), s.getThrowable()); } else { logger.warn(s.getMessage()); } break; case Status.ERROR: if (s.getThrowable() != null) { logger.warn(s.getMessage(), s.getThrowable()); } else { logger.warn(s.getMessage()); } break; } } if (count > 0) { fail("Appender raised an exception."); } }
/** * Creates a new instance whose logger name is the fully qualified class * name of the instance. * * @param level the log level */ public LoggingHandler(LogLevel level) { if (level == null) { throw new NullPointerException("level"); } logger = InternalLoggerFactory.getInstance(getClass()); this.level = level; internalLevel = InternalLogLevel.DEBUG; }
/** * Creates a new instance with the specified logger name. * * @param clazz the class type to generate the logger for * @param level the log level */ public LoggingHandler(Class<?> clazz, LogLevel level) { if (clazz == null) { throw new NullPointerException("clazz"); } if (level == null) { throw new NullPointerException("level"); } logger = InternalLoggerFactory.getInstance(clazz); this.level = level; internalLevel = InternalLogLevel.DEBUG; }
/** * Creates a new instance with the specified logger name. * * @param name the name of the class to use for the logger * @param level the log level */ public LoggingHandler(String name, LogLevel level) { if (name == null) { throw new NullPointerException("name"); } if (level == null) { throw new NullPointerException("level"); } logger = InternalLoggerFactory.getInstance(name); this.level = level; internalLevel = InternalLogLevel.DEBUG; }
/** * 系统参数配置 * @throws Exception */ public void initSystem() throws Exception { PropertyConfigurator.configure("Log4j.properties"); InternalLoggerFactory.setDefaultFactory(new Slf4JLoggerFactory()); log.info(System.getProperty("file.encoding")); System.setProperty("io.netty.recycler.maxCapacity.default", PropertyUtil.getProperty("io.netty.recycler.maxCapacity.default")); System.setProperty("io.netty.leakDetectionLevel", "paranoid"); DbHelper.init(); }
public static void main(String[] args) throws Exception { Properties props = PEFileUtils.loadPropertiesFile(LoadBalancer.class, PEConstants.CONFIG_FILE_NAME); if (args.length == 2 && "-port".equalsIgnoreCase(args[0])) props.setProperty(PORT_PROPERTY, args[1]); else if (args.length > 0) throw new Exception("Usage: LoadBalancer [-port <port>]"); InternalLoggerFactory.setDefaultFactory(new Log4JLoggerFactory()); LoadBalancer loadBalancer = new LoadBalancer(props); loadBalancer.run(); }
/** * Creates a new instance with the specified logger name. * * @param clazz the class type to generate the logger for. * @param level the log level. */ public LoggingHandler(Class<?> clazz, LogLevel level) { if (clazz == null) { throw new NullPointerException("clazz"); } if (level == null) { throw new NullPointerException("level"); } logger = InternalLoggerFactory.getInstance(clazz); this.level = level; internalLevel = level.toInternalLevel(); }
/** * Creates a new instance with the specified logger name. * * @param name the name of the class to use for the logger. * @param level the log level. */ public LoggingHandler(String name, LogLevel level) { if (name == null) { throw new NullPointerException("name"); } if (level == null) { throw new NullPointerException("level"); } logger = InternalLoggerFactory.getInstance(name); this.level = level; internalLevel = level.toInternalLevel(); }
@Override public void init(BundleContext context, DependencyManager manager) throws Exception { // set the Netty log factory InternalLoggerFactory.setDefaultFactory(new Slf4JLoggerFactory()); // create all OSGi managers createManagers(manager); // listen for the HubManager to be published hubManagerTracker = new ServiceTracker(context, HubManager.class.getName(), null) { @Override public Object addingService(ServiceReference ref) { startHubManager((HubManager)context.getService(ref)); return null; } }; hubManagerTracker.open(); // wait for ConfigurationAdmin to become available to start advertising presence presenceTracker = new ServiceTracker(context, ConfigurationManager.class.getName(), null) { @Override public Object addingService(ServiceReference serviceRef) { ServiceReference ref = context.getServiceReference(ConfigurationManager.class.getName()); if (ref != null) { // start advertisements return context.getService(ref); } else { return null; } } @Override public void removedService(ServiceReference ref, Object service) { super.removedService(ref, service); } }; presenceTracker.open(); }
@Override public void acceptOptions(List<String> args, File gameDir, File assetsDir, String profile) { NailedLauncher.initialize(gameDir); System.setOut(new PrintStream(new LoggerOutputStream(LogManager.getLogger("SYSOUT"), Level.INFO), true)); System.setErr(new PrintStream(new LoggerOutputStream(LogManager.getLogger("SYSERR"), Level.WARN), true)); InternalLoggerFactory.getInstance("INITLOGGER"); }
/** * The main method. * * @param args * Arguments for the command line program. Must contain addresses for announce, multicast and listen, a * classpath for a .blocks file and either --create or --deploy to determine whether to only create or * also deploy a distribution. */ public static void main(String args[]) { InternalLoggerFactory.setDefaultFactory(new Log4j2LoggerFactory()); if (args.length > 0) { LOGGER.debug("argument 0 is \"{}\"", args[0]); if (args[0].equals("--help") || args[0].equals("-h")) { System.out.println("Parameters: [--help| $pathToConfig]"); System.out.println("\t--help: print this message"); System.out.println("\t$pathToConfig the path to the used config file."); System.exit(0); } else { parseArguments(args); } } else { exitFalseInput(); } ResourceSet newSet = new ResourceSetImpl(); newSet.setPackageRegistry(new EPackageRegistryImpl()); newSet.getPackageRegistry().put(ModelPackage.eNS_URI, ModelPackage.eINSTANCE); FunctionBlockLoader blockLoader = new FunctionBlockLoader(path); blockLoader.loadBlocks(); functionBlocks = blockLoader.getBlocks(); serverManager = new TCPUDPServerManager(); serverManager.startServer(new SimpleAddressBasedServerConfig(moduleID, Collections.singletonList(listen), Collections.singletonList(multicast), Collections.singletonList(announce), announceInterval)); ModuleRegistrator moduleRegistrator = new ModuleRegistrator(); serverManager.getModuleManager().addListener(moduleRegistrator); CommandLoop loop = new CommandLoop(functionBlocks, blockLoader.getApplicationName(), serverManager); loop.loop(createOrDeploy); exit(); }
public static void main(String[] args) throws Exception { // must do this first for logback InternalLoggerFactory.setDefaultFactory( new Slf4JLoggerFactory( )); // args[0] should be the path of the transficc.props properties file String propsFilePath = args[0]; Properties props = new Properties( ); // load up the props for socket and DB config File pfile = new File( propsFilePath); logger.info( "Loading config from: " + propsFilePath); InputStream istrm = new FileInputStream( propsFilePath); props.load( istrm); String sport = props.getProperty( "SocketServerPort", "8080"); logger.info( "Binding to port: " + sport); int port = Integer.parseInt( sport); // Create the transficc client, and the Q it uses to send events including // market data back to the thread that pushes updates on the web sockets. // Also an executor to provide a thread to run the TFClient. We keep a ref // to the transficc service so we can pass it to the web socket handler. LinkedBlockingQueue<JSON.Message> outQ = new LinkedBlockingQueue<JSON.Message>( ); TFClient tfc = new TFClient( props, outQ); TransficcService tfs = tfc.GetService( ); // need an executor for the thread that will intermittently send data to the client ThreadFactoryBuilder builder = new ThreadFactoryBuilder( ); builder.setDaemon( true); builder.setNameFormat( "transficc-%d"); ExecutorService transficcExecutor = Executors.newSingleThreadExecutor( builder.build( )); FutureTask<String> transficcTask = new FutureTask<String>( tfc); transficcExecutor.execute( transficcTask); // Now we can create the pusher object and thread, which consumes the event // on the outQ. Those events originate with TFClient callbacks from transficc, // and also incoming websock events like subscription requests. WebSockPusher pusher = new WebSockPusher( props, outQ, tfs); ExecutorService pusherExecutor = Executors.newSingleThreadExecutor( builder.build( )); FutureTask<String> pusherTask = new FutureTask<String>( pusher); pusherExecutor.execute( pusherTask); // Compose the server components... EventLoopGroup bossGroup = new NioEventLoopGroup( 1); EventLoopGroup workerGroup = new NioEventLoopGroup( ); try { ServerBootstrap b = new ServerBootstrap( ); LoggingHandler lh = new LoggingHandler(LogLevel.INFO); ChannelInitializer ci = new ChannelInitializer<SocketChannel>( ) { @Override public void initChannel(SocketChannel ch) throws Exception { ChannelPipeline p = ch.pipeline(); p.addLast("encoder", new HttpResponseEncoder()); p.addLast("decoder", new HttpRequestDecoder()); p.addLast("aggregator", new HttpObjectAggregator(65536)); p.addLast("handler", new WebSocketHandler( props, outQ)); } }; b.group( bossGroup, workerGroup).channel( NioServerSocketChannel.class).handler( lh).childHandler( ci); // Fire up the server... ChannelFuture f = b.bind( port).sync( ); logger.info( "Server started"); // Wait until the server socket is closed. f.channel( ).closeFuture( ).sync( ); } finally { logger.info( "Server shutdown started"); // Shut down all event loops to terminate all threads. bossGroup.shutdownGracefully( ); workerGroup.shutdownGracefully(); logger.info( "Server shutdown completed"); } }
public ChannelFuture run() throws Exception { InternalLoggerFactory.setDefaultFactory(Log4JLoggerFactory.INSTANCE); Channel channle = getDefaultServerBootstrap().childHandler(new TcpChannelInitializer()).bind(port).sync().channel(); channels.add(channle); logger.info("MQTT server is started at port " + port + '.'); ChannelFuture future = getDefaultServerBootstrap().childHandler(new HttpChannelInitializer()).bind(httpPort); Channel httpChannel = future.sync().channel(); channels.add(httpChannel); logger.info("MQTT websocket server is started at port " + httpPort + '.'); return future; }
public static InternalLogger getLogger(Class<?> c) { return InternalLoggerFactory.getInstance(c); }