当前位置: 首页 > news >正文

网站网页的书签怎么做手机网站设计建设

网站网页的书签怎么做,手机网站设计建设,建设项目环境影响登记表网站,做微信网站公司名称研究背景 1、之所以选用kafka是因为量起来的话单台logstash的抗压能力比较差 2、为了解决整个链路查询的问题#xff0c;多个Feign传层的话#xff0c;可以按照一个ID进行穿层#xff0c;所以采用logback的MDC进行对唯一标识存储并且在Feign的调用链放在Header里#xff…研究背景 1、之所以选用kafka是因为量起来的话单台logstash的抗压能力比较差 2、为了解决整个链路查询的问题多个Feign传层的话可以按照一个ID进行穿层所以采用logback的MDC进行对唯一标识存储并且在Feign的调用链放在Header里这里命名为TID 下载地址 ZKKafka https://mirrors.bfsu.edu.cn/apache/kafka/2.7.0/kafka_2.13-2.7.0.tgz https://mirrors.bfsu.edu.cn/apache/zookeeper/zookeeper-3.7.0/apache-zookeeper-3.7.0-bin.tar.gz ELK https://artifacts.elastic.co/downloads/kibana/kibana-7.12.0-windows-x86_64.zip https://artifacts.elastic.co/downloads/elasticsearch/elasticsearch-7.12.0-windows-x86_64.zip https://artifacts.elastic.co/downloads/logstash/logstash-7.12.0-windows-x86_64.zip 在拦截器里增加相对应的拦截代码 Component Slf4j public class ContextInterceptor implements HandlerInterceptor {RequestContext context RequestContext.getCurrentContext();context.reset();log.debug(traceId: MDC.get(traceId));String requestId MDC.get(traceId);requestId StringUtils.isEmpty(requestId) ? request.getHeader(RequestContext.REQUEST_ID) : requestId;requestId StringUtils.isEmpty(requestId) ? request.getParameter(RequestContext.REQUEST_ID) : requestId;requestId StringUtils.isEmpty(requestId) ? UUIDUtil.uuid() : requestId;MDC.put(TID, requestId);} 配置日志配置文件logback-spring.xml ?xml version1.0 encodingUTF-8? configuration!-- springProfile用于指定当前激活的环境如果spring.profile.active的值是哪个就会激活对应节点下的配置 --springProfile namelocal!-- configuration to be enabled when the staging profile is active --springProperty scopecontext namemodule sourcespring.application.namedefaultValueundefinded/!-- 该节点会读取Environment中配置的值在这里我们读取application.yml中的值 --springProperty scopecontext namebootstrapServers sourcespring.kafka.bootstrap-serversdefaultValue127.0.0.1:9092/appender nameSTDOUT classch.qos.logback.core.ConsoleAppender!-- encoders are assigned the typech.qos.logback.classic.encoder.PatternLayoutEncoder by default --encoderpattern%boldYellow(${module})|%d|%highlight(%-5level)|%X{TID}|%cyan(%logger{15}) - %msg %n/pattern/encoder/appender!-- kafka的appender配置 --appender namekafka classcom.github.danielwegener.logback.kafka.KafkaAppenderencoderpattern${module}|%d|%-5level|%X{TID}|%logger{15} - %msg/pattern/encodertopictest/topickeyingStrategy classcom.github.danielwegener.logback.kafka.keying.NoKeyKeyingStrategy/deliveryStrategy classcom.github.danielwegener.logback.kafka.delivery.AsynchronousDeliveryStrategy/!-- Optional parameter to use a fixed partition --!-- partition0/partition --!-- Optional parameter to include log timestamps into the kafka message --!-- appendTimestamptrue/appendTimestamp --!-- each producerConfig translates to regular kafka-client config (format: keyvalue) --!-- producer configs are documented here: https://kafka.apache.org/documentation.html#newproducerconfigs --!-- bootstrap.servers is the only mandatory producerConfig --producerConfigbootstrap.servers${bootstrapServers}/producerConfig!-- 如果kafka不可用则输出到控制台 --appender-ref refSTDOUT//appender!-- 指定项目中的logger --!--logger nameorg.springframework.test levelINFO appender-ref refkafka //logger--logger namecom.springcloudsite levelINFO appender-ref refkafka //loggerroot levelinfoappender-ref refSTDOUT //root/springProfile /configuration 正则配置说明     pattern为正则表达 %boldYellow(${module}) 黄色的模块名称 %d 日期时间 %highlight(%-5level)高亮的日志级别如info error trace登 %X{TID} : traceID 追踪使用的ID %cyan(%logger{15}) :简写类名路径 %msg %n :具体日志信息 打印出来的效果如下 配置zkkafka 1. 安装JDK 1.1 安装文件http://www.oracle.com/technetwork/java/javase/downloads/index.html 下载JDK 1.2 安装完成后需要添加以下的环境变量右键点击“我的电脑” - 高级系统设置 - 环境变量  JAVA_HOME: C:\Program Files\Java\jdk1.8.0_171 (jdk的安装路径) Path: 在现有的值后面添加; %JAVA_HOME%\bin 1.3 打开cmd运行 java -version 查看当前系统Java的版本 2. 安装ZOOKEEPER Kafka的运行依赖于Zookeeper所以在运行Kafka之前我们需要安装并运行Zookeeper 2.1 下载安装文件 http://zookeeper.apache.org/releases.html 2.2 解压文件  2.3 打开zookeeper-3.4.13\conf把zoo_sample.cfg重命名成zoo.cfg 2.4 从文本编辑器里打开zoo.cfg 2.5 把dataDir的值改成“./zookeeper-3.4.13/data” 2.6 添加如下系统变量 ZOOKEEPER_HOME: C:\Users\localadmin\CODE\zookeeper-3.4.13 (zookeeper目录)Path: 在现有的值后面添加 ;%ZOOKEEPER_HOME%\bin; 2.7 运行Zookeeper: 打开cmd然后执行 zkserver cmd 窗口不要关闭 3. 安装并运行KAFKA 3.1 下载安装文件 http://kafka.apache.org/downloads.html 3.2 解压文件 3.3 打开kafka_2.11-2.0.0\config 3.4 从文本编辑器里打开 server.properties 3.5 把 log.dirs的值改成 “./logs” 3.6 打开cmd 3.7 进入kafka文件目录: cd C:\Users\localadmin\CODE\kafka_2.11-2.0.0(kafka目录) 3.8 输入并执行:  .\bin\windows\kafka-server-start.bat .\config\server.properties cmd 窗口不要关闭 4. 创建TOPICS 4.1 打开cmd 并进入cd C:\Users\localadmin\CODE\kafka_2.11-2.0.0\bin\windows 4.2 创建一个topic kafka-topics.bat --create --zookeeper localhost:2181 --replication-factor 1 --partitions 1 --topic test 5. 打开一个PRODUCER cd C:\Users\localadmin\CODE\kafka_2.11-2.0.0\bin\windows kafka-console-producer.bat --broker-list localhost:9092 --topic test 6. 打开一个CONSUMER: cd C:\Users\localadmin\CODE\kafka_2.11-2.0.0\bin\windows kafka-console-consumer.bat --bootstrap-server localhost:9092 --topic test --from-beginning 7. 测试: 配置ELK kibana.yml # Kibana is served by a back end server. This setting specifies the port to use. server.port: 5601# Specifies the address to which the Kibana server will bind. IP addresses and host names are both valid values. # The default is localhost, which usually means remote machines will not be able to connect. # To allow connections from remote users, set this parameter to a non-loopback address. server.host: localhost# Enables you to specify a path to mount Kibana at if you are running behind a proxy. # Use the server.rewriteBasePath setting to tell Kibana if it should remove the basePath # from requests it receives, and to prevent a deprecation warning at startup. # This setting cannot end in a slash. #server.basePath: # Specifies whether Kibana should rewrite requests that are prefixed with # server.basePath or require that they are rewritten by your reverse proxy. # This setting was effectively always false before Kibana 6.3 and will # default to true starting in Kibana 7.0. #server.rewriteBasePath: false# Specifies the public URL at which Kibana is available for end users. If # server.basePath is configured this URL should end with the same basePath. #server.publicBaseUrl: # The maximum payload size in bytes for incoming server requests. #server.maxPayloadBytes: 1048576# The Kibana servers name. This is used for display purposes. #server.name: your-hostname# The URLs of the Elasticsearch instances to use for all your queries. elasticsearch.hosts: [http://localhost:9200]# Kibana uses an index in Elasticsearch to store saved searches, visualizations and # dashboards. Kibana creates a new index if the index doesnt already exist. #kibana.index: .kibana# The default application to load. #kibana.defaultAppId: home# If your Elasticsearch is protected with basic authentication, these settings provide # the username and password that the Kibana server uses to perform maintenance on the Kibana # index at startup. Your Kibana users still need to authenticate with Elasticsearch, which # is proxied through the Kibana server. #elasticsearch.username: kibana_system #elasticsearch.password: pass# Enables SSL and paths to the PEM-format SSL certificate and SSL key files, respectively. # These settings enable SSL for outgoing requests from the Kibana server to the browser. #server.ssl.enabled: false #server.ssl.certificate: /path/to/your/server.crt #server.ssl.key: /path/to/your/server.key# Optional settings that provide the paths to the PEM-format SSL certificate and key files. # These files are used to verify the identity of Kibana to Elasticsearch and are required when # xpack.security.http.ssl.client_authentication in Elasticsearch is set to required. #elasticsearch.ssl.certificate: /path/to/your/client.crt #elasticsearch.ssl.key: /path/to/your/client.key# Optional setting that enables you to specify a path to the PEM file for the certificate # authority for your Elasticsearch instance. #elasticsearch.ssl.certificateAuthorities: [ /path/to/your/CA.pem ]# To disregard the validity of SSL certificates, change this settings value to none. #elasticsearch.ssl.verificationMode: full# Time in milliseconds to wait for Elasticsearch to respond to pings. Defaults to the value of # the elasticsearch.requestTimeout setting. #elasticsearch.pingTimeout: 1500# Time in milliseconds to wait for responses from the back end or Elasticsearch. This value # must be a positive integer. elasticsearch.requestTimeout: 30000# List of Kibana client-side headers to send to Elasticsearch. To send *no* client-side # headers, set this value to [] (an empty list). #elasticsearch.requestHeadersWhitelist: [ authorization ]# Header names and values that are sent to Elasticsearch. Any custom headers cannot be overwritten # by client-side headers, regardless of the elasticsearch.requestHeadersWhitelist configuration. #elasticsearch.customHeaders: {}# Time in milliseconds for Elasticsearch to wait for responses from shards. Set to 0 to disable. #elasticsearch.shardTimeout: 30000# Logs queries sent to Elasticsearch. Requires logging.verbose set to true. #elasticsearch.logQueries: false# Specifies the path where Kibana creates the process ID file. #pid.file: /run/kibana/kibana.pid# Enables you to specify a file where Kibana stores log output. #logging.dest: stdout# Set the value of this setting to true to suppress all logging output. #logging.silent: false# Set the value of this setting to true to suppress all logging output other than error messages. #logging.quiet: false# Set the value of this setting to true to log all events, including system usage information # and all requests. #logging.verbose: false# Set the interval in milliseconds to sample system and process performance # metrics. Minimum is 100ms. Defaults to 5000. #ops.interval: 5000# Specifies locale to be used for all localizable strings, dates and number formats. # Supported languages are the following: English - en , by default , Chinese - zh-CN . #i18n.locale: en然后到对应bin目录下启动直接点击 kibana.bat启动即可或者在CMD命令启动 之后是启动效果 配置elasticsearch.yml # Elasticsearch Configuration # # NOTE: Elasticsearch comes with reasonable defaults for most settings. # Before you set out to tweak and tune the configuration, make sure you # understand what are you trying to accomplish and the consequences. # # The primary way of configuring a node is via this file. This template lists # the most important settings you may want to configure for a production cluster. # # Please consult the documentation for further information on configuration options: # https://www.elastic.co/guide/en/elasticsearch/reference/index.html # # ---------------------------------- Cluster ----------------------------------- # # Use a descriptive name for your cluster: # #cluster.name: my-application # # ------------------------------------ Node ------------------------------------ # # Use a descriptive name for the node: # #node.name: node-1 # # Add custom attributes to the node: # #node.attr.rack: r1 # # ----------------------------------- Paths ------------------------------------ # # Path to directory where to store the data (separate multiple locations by comma): # #path.data: /path/to/data # # Path to log files: # #path.logs: /path/to/logs # # ----------------------------------- Memory ----------------------------------- # # Lock the memory on startup: # #bootstrap.memory_lock: true # # Make sure that the heap size is set to about half the memory available # on the system and that the owner of the process is allowed to use this # limit. # # Elasticsearch performs poorly when the system is swapping the memory. # # ---------------------------------- Network ----------------------------------- # # By default Elasticsearch is only accessible on localhost. Set a different # address here to expose this node on the network: # cluster.name: docker-cluster node.name: node-1 node.master: true network.host: 0.0.0.0#xpack.license.self_generated.type: trial #xpack.security.enabled: true #xpack.monitoring.collection.enabled: true # # By default Elasticsearch listens for HTTP traffic on the first free port it # finds starting at 9200. Set a specific HTTP port here: # http.port: 9200 # # For more information, consult the network module documentation. # # --------------------------------- Discovery ---------------------------------- # # Pass an initial list of hosts to perform discovery when this node is started: # The default list of hosts is [127.0.0.1, [::1]] # #discovery.seed_hosts: [host1, host2] # # Bootstrap the cluster using an initial set of master-eligible nodes: # cluster.initial_master_nodes: [node-1] # # For more information, consult the discovery and cluster formation module documentation. # # ---------------------------------- Various ----------------------------------- # # Require explicit names when deleting indices: # #action.destructive_requires_name: true启动bin目录下的elasticsearch.bat 以下是启动效果 配置logstash.conf # Sample Logstash configuration for creating a simple # Beats - Logstash - Elasticsearch pipeline.input {kafka {bootstrap_servers localhost:9092topics [test]group_id test} }filter { mutate {split { message | }}if [message][0] {mutate { add_field {apiname %{[message][0]}}}}if [message][1] {mutate { add_field {current_time %{[message][1]}}}} if [message][2] {mutate { add_field {current_level %{[message][2]}}}} if [message][3] {mutate { add_field {traceid %{[message][3]}}}}}output {elasticsearch {hosts [http://localhost:9200]#index local-purchase-order | %{YYYY-MM-dd}index logstash-%{YYYY-MM-dd}#template_name logstash#template_overwrite true#index %{[metadata][beat]}-%{[metadata][version]}-%{YYYY.MM.dd}#user elastic#password changeme}stdout{codec rubydebug} } 配置logstash.yml #/usr/share/logstash/config/logstash.yml #jvm.options log4j2.properties logstash-sample.conf logstash.yml pipelines.yml startup.options http.host: 0.0.0.0 # [ http://elasticsearch:9200 ] xpack.monitoring.elasticsearch.hosts: ${ELASTICSEARCH_URL} 启动使用命令 可以进到bin下 D:\app\elk\logstash\bin 输入命令logstash -f D:\app\elk\logstash\config\logstash.conf 最后打开地址 http://localhost:9600/ http://localhost:9200/ http://localhost:5601/ 分别验证结果
http://www.huolong8.cn/news/89591/

相关文章:

  • 制作网站公司图片不用下载就能看的网站的浏览器
  • 有哪些平面设计网站制作网页创建站点的步骤
  • 微网站需要备案吗wordpress 需要多大数据库
  • 网站制作价格服务wordpress免费资讯主题
  • dede 电商网站模板设计本家装
  • 网站建设 爱诚科技网站的推广是怎么做的
  • 服务好的网站制作做微商加入什么移动电商网站
  • 企业网站备个人google英文网站
  • 烟台定制网站建设价格网站模板制作流程
  • 怎样用网站做淘宝客代网站建设
  • 帝国cms二手网站模板网站子站怎么建设
  • 河南网站建设哪个公司做得好做名片赞机器人电脑网站是多少
  • 胶州做网站php编程语言
  • 民治营销型网站设计哪家好搭建单位网站
  • 网站交互用什么做抚州seo排名
  • 服装html网站模板下载天元建设集团有限公司第二公司
  • 阿里云网站建设里云官网模版今天军事新闻最新消息
  • 江西网站建设价位北京微信网站设计报价
  • 成都网址建设医疗网站建设及优化
  • 有什么网站可以做六级题目嘛肥城网站建设方案
  • 上海网站seo优化wordpress自定义代码在哪里设置
  • 网站建设优化建站台州优化官方网站
  • 哈尔滨住房建设发展集团网站深圳龙华区邮政编码
  • 品牌宝网站认证网页设计心得体会正文
  • 网站建设公司crm系统个人兼职网站制作
  • 网站免费模版代码南京seo公司哪家
  • 费县网站建设百度一下知道首页
  • 网站建设公司有前途吗企业信息查询公示系统
  • 营销型网站头部布局的元素黑客钓鱼网站的制作
  • 个人网站可以做论坛吗?psd转wordpress主题