Apache Pekko Connectors Documentation Parsing Lines
Type to start searching
    apache/pekko-connectors
    apache/pekko-connectors
    • Overview
      • Versions
      • Contributing
      • External Components
      • Self-contained examples
      • Other documentation
      • Integration Patterns
      • Release Notes
      • License Report
    • Data Transformations
      • Parsing Lines
      • JSON
      • Compressing/decompressing
      • Comma-Separated Values - CSV
      • RecordIO Framing
      • Text and charsets
      • Extensible Markup Language - XML
    • AMQP
      • Artifacts
      • Connecting to server
      • Sending messages
      • Receiving messages
      • Using Pub/Sub
      • Using rabbitmq as an RPC mechanism
      • Acknowledging messages downstream
    • Apache Cassandra
      • Artifacts
      • Sessions
      • Reading from Cassandra
      • Writing to Cassandra
      • Custom Session creation
    • Apache Geode
      • Artifacts
      • Setup
      • Writing to Geode
      • Reading from Geode
      • Geode basic commands
    • Apache Kafka
      • Separate repository
    • Apache Kudu
      • Artifacts
      • Configuration
      • Writing to Kudu in a Flow
      • Writing to Kudu with a Sink
    • Apache Solr
      • Artifacts
      • Set up a Solr client
      • Reading from Solr
      • Writing to Solr
      • Update documents
      • Delete documents by ids
      • Delete documents by query
    • Avro Parquet
      • Artifacts
      • Source Initiation
      • Sink Initiation
      • Flow Initiation
      • Running the example code
    • AWS EventBridge
      • Artifacts
      • Setup
      • Publish messages to AWS EventBridge Event Bus
      • Integration testing
    • AWS DynamoDB
      • Artifacts
      • Setup
      • Sending requests and receiving responses
      • Error Retries and Exponential Backoff
      • Shared AWS client configuration
    • AWS Kinesis and Firehose
      • Artifacts
      • Kinesis Data Streams
      • AWS KCL Scheduler Source & checkpointer
    • AWS Lambda
      • Artifacts
      • Setup
      • Sending messages
      • AwsLambdaFlow configuration
    • AWS S3
      • Artifacts
      • Configuration
      • Store a file in S3
      • Download a file from S3
      • Access object metadata without downloading object from S3
      • List bucket contents
      • List bucket contents and common prefixes
      • Copy upload (multi part)
      • Apply S3 settings to a part of the stream
      • Bucket management
      • Running the example code
    • AWS SNS
      • Artifacts
      • Setup
      • Publish messages to an SNS topic
    • AWS SQS
      • Artifacts
      • Setup
      • Read from an SQS queue
      • Publish messages to an SQS queue
      • Updating message statuses
      • Integration testing
    • Azure Storage Queue
      • Artifacts
      • Init Azure Storage API
      • Queuing a message
      • Processing and deleting messages
    • Couchbase
      • Artifacts
      • Overview
      • Reading from Couchbase in Apache Pekko Streams
      • Writing to Couchbase in Apache Pekko Streams
      • Using CouchbaseSession directly
    • Elasticsearch
      • Artifacts
      • Elasticsearch parameters
      • Elasticsearch as Source and Sink
      • Elasticsearch as Flow
    • File
      • Artifacts
      • Writing to and reading from files
      • Tailing a file into a stream
      • Creating directories
      • Listing directory contents
      • Listening to changes in a directory
      • Rotating the file to stream into
      • ZIP Archive
      • TAR Archive
    • FTP
      • Artifacts
      • Configuring the connection settings
      • Traversing a remote FTP folder recursively
      • Retrieving files
      • Writing files
      • Removing files
      • Moving files
      • Creating directory
    • Google Common
      • Artifacts
      • Configuration
      • Credentials
      • Accessing settings
      • Apply custom settings to a part of the stream
      • Interop with Google Java client libraries
      • Accessing other Google APIs
    • Google Cloud BigQuery
      • Artifacts
      • Configuration
      • Imports
      • Setup data classes
      • Run a query
      • Load data into BigQuery
      • Managing datasets and tables
      • Apply custom settings to a part of the stream
      • Make raw API requests
    • Google Cloud BigQuery Storage
      • Artifacts
      • Build setup
      • Configuration
      • Reading
      • Running the test code
    • Google Cloud Pub/Sub
      • Artifacts
      • Usage
      • Running the examples
    • Google Cloud Pub/Sub gRPC
      • Artifacts
      • Binary compatibility
      • Build setup
      • Configuration
      • Publishing
      • Subscribing
      • Running the test code
    • Google Cloud Storage
      • Artifacts
      • Configuration
      • Store a file in Google Cloud Storage
      • Download a file from Google Cloud Storage
      • Access object metadata without downloading object from Google Cloud Storage
      • List bucket contents
      • Rewrite (multi part)
      • Apply Google Cloud Storage settings to a part of the stream
      • Bucket management
      • Running the example code
    • Google FCM
      • Artifacts
      • Settings
      • Sending notifications
      • Scala only
    • gRPC
      • Apache Pekko gRPC
    • Hadoop Distributed File System - HDFS
      • Artifacts
      • Specifying a Hadoop Version
      • Set up client
      • Writing
      • Configuration
      • Reading
      • Running the example code
    • HBase
      • Artifacts
      • Converters
      • Settings
      • Source
      • Flow
      • Sink
      • HBase administration commands
    • Huawei Push Kit
      • Artifacts
      • Settings
      • Sending notifications
      • Scala only
    • HTTP
      • Apache Pekko HTTP
    • IBM Bluemix Cloud Object Storage
      • Connection limitations
    • InfluxDB
      • Influxdata, the makers of InfluxDB now offer an Apache Pekko Streams-aware client library in https://github.com/influxdata/influxdb-client-java/tree/master/client-scala
      • Artifacts
      • Set up InfluxDB client
      • InfluxDB as Source and Sink
    • IronMQ
      • Artifacts
      • Consumer
      • Producer
    • JMS
      • Artifacts
      • Producer
      • Consumer
      • Browse
      • Using IBM MQ
    • MongoDB
      • Artifacts
      • Initialization
      • Source
      • Flow and Sink
    • MQTT
      • Artifacts
      • Settings
      • Reading from MQTT
      • Publishing to MQTT
      • Publish and subscribe in a single flow
      • Using flow with Acknowledge on message sent
      • Capturing MQTT client logging
      • Running the example code
    • MQTT v5
      • Artifacts
      • Settings
      • Reading from MQTT
      • Publishing to MQTT
      • Publish and subscribe in a single flow
      • Using flow with Acknowledge on message sent
      • Capturing MQTT client logging
      • Running the example code
    • MQTT Streaming
      • Artifacts
      • Flow through a client session
      • Flow through a server session
    • Opensearch
      • Artifacts
      • Opensearch parameters
      • Opensearch as Source and Sink
      • Opensearch as Flow
    • OrientDB
      • Artifacts
      • Database connection
      • Reading ODocument from OrientDB
      • Typed messages
      • Source configuration
      • Writing to OrientDB
    • Pravega
      • Artifacts
      • Concepts
      • Configuration
      • Support
    • Server-sent Events (SSE)
      • Artifacts
      • Usage
    • Slick (JDBC)
      • Artifacts
      • Initialization
      • Starting a Database Session
      • Closing a Database Session
      • Using a Slick Source
      • Using a Slick Flow or Sink
      • Flow
      • Flow with pass-through
    • Spring Web
      • Artifacts
      • Usage
      • Shameless plug: Apache Pekko HTTP
    • TCP
      • Apache Pekko TCP
    • UDP
      • Artifacts
      • Sending
      • Receiving
      • Running the example code
    • Unix Domain Socket
      • Artifacts
      • Binding to a file
      • Connecting to a file
    • link Apache Software Foundation
    • link License
    • link Security
    • link Privacy
    • link Events
    • link Donate
    • link Thanks

    Parsing Lines

    Most Apache Pekko Connectors sources stream ByteStringByteString elements which normally won’t contain data line by line. To split these at line separators use FramingFraming as described in the Apache Pekko Stream documentation.

    Apache Pekko documentation

    Edit this page
    1.2.0-M0+59-20323efa*
    Previous Data Transformations
    Next JSON
    Copyright © 2011-2022 Lightbend, Inc. Apache Pekko, Pekko, and its feather logo are trademarks of The Apache Software Foundation.