With 7.10, the PCF logging integration goes GA, adding the ability to send logs from PCF at scale via Filebeat.
Filebeat ships with modules for observability and security data sources that simplify the collection, parsing, and visualization of common log formats down to a single command. They achieve this by combining automatic default paths based on your operating system, with Elasticsearch Ingest Node pipeline definitions, and with Kibana dashboards. Plus, a few Filebeat modules ship with preconfigured machine learning jobs.
Containerizing everything or running in a cloud environment? Container monitoring and cloud monitoring with the Elastic Stack is simple. Deploy Filebeat in a Kubernetes, Docker, or cloud deployment and get all of the log streams — complete with their pod, container, node, VM, host, and other metadata for automatic correlation. Plus, Beats Autodiscover features detect new containers and adaptively monitor them with the appropriate Filebeat modules.
Filebeat uses a backpressure-sensitive protocol when sending data to Logstash or Elasticsearch to account for higher volumes of data. If Logstash is busy crunching data, it lets Filebeat know to slow down its read. Once the congestion is resolved, Filebeat will build back up to its original pace and keep on shippin'.
Filebeat is part of the Elastic Stack, meaning it works seamlessly with Logstash, Elasticsearch, and Kibana. Whether you want to transform or enrich your logs and files with Logstash, fiddle with some analytics in Elasticsearch, or build and share dashboards in Kibana, Filebeat makes it easy to ship your data to where it matters most.