All Classes and Interfaces

Class
Description
The Abstract class of Arrow Aggregate Operator for Pandas AggregateFunction.
Abstract class extends BeamStateHandler, which implements the common handle logic.
Base class for all Python DataStream operators executed in embedded Python environment.
Abstract class for all stream operators to execute Python functions in embedded Python environment.
Base class for all stream operators to execute Python Stateless Functions in embedded Python environment.
Base class for all Python DataStream operators executed in Python process.
AbstractExternalOneInputPythonFunctionOperator is responsible for launching beam runner which will start a python harness to execute user defined python function.
Abstract class for all stream operators to execute Python functions in external environment.
AbstractExternalTwoInputPythonFunctionOperator is responsible for launching beam runner which will start a python harness to execute two-input user defined python function.
AbstractOneInputEmbeddedPythonFunctionOperator is responsible for run Python DataStream operators with one input in Embedded Python environment.
Base class for all one input stream operators to execute Python functions.
The base class of python environment manager which is used to create the PythonEnvironment object used to execute Python functions.
Python lease resource which includes environment variables and working directory of execution python environment.
Base class for all stream operators to execute Python functions.
Base class for all stream operators to execute Python ScalarFunctions.
Base class for all stream operators to execute Python Stateless Functions.
The Abstract class of Stream Arrow Python AggregateFunction Operator for RANGE clause bounded Over Window Aggregation.
The Abstract class of Stream Arrow Python AggregateFunction Operator for ROWS clause bounded Over Window Aggregation.
The Abstract class of Stream Arrow Python AggregateFunction Operator for Over Window Aggregation.
AbstractTwoInputEmbeddedPythonFunctionOperator is responsible for run Python DataStream operators with two input user defined python function in Embedded Python environment.
Creates arrays of objects.
A TypeSerializer for ArrayData.
TypeSerializerSnapshot for ArrayDataSerializer.
ArrowFieldWriter for Array.
ArrayWriter for ArrayData input.
ArrayWriter for RowData input.
Arrow column vector for Array.
Arrow column vector for BigInt.
Arrow column vector for Binary.
Arrow column vector for Boolean.
Arrow column vector for Date.
Arrow column vector for DecimalData.
Arrow column vector for Double.
Base class for arrow field writer which is used to convert a field to an Arrow format.
Arrow column vector for Float.
Arrow column vector for Int.
Arrow column vector for Map.
Arrow column vector for Null.
Arrow Python ScalarFunction operator.
ArrowReader which read the underlying Arrow format data as RowData.
Arrow column vector for Row.
The base class ArrowSerializer which will serialize/deserialize RowType data to/from arrow bytes.
Arrow column vector for Int.
Deprecated.
This class is based on the SourceFunction API, which is due to be removed.
A ScanTableSource for serialized arrow record batch data.
Factory for creating configured instances of ArrowTableSource..
Table options for the ArrowTableSource.
Arrow column vector for Time.
Arrow column vector for Timestamp.
Arrow column vector for TinyInt.
Utilities for Arrow.
Arrow column vector for VarBinary.
Arrow column vector for VarChar.
Writer which serializes the Flink rows to Arrow format.
Deserialization schema from Avro bytes to Row.
Serialization schema that serializes Row into Avro bytes.
The Batch Arrow Python AggregateFunction Operator for Group Aggregation.
The Batch Arrow Python AggregateFunction Operator for Group Window Aggregation.
The Batch Arrow Python AggregateFunction Operator for Over Window Aggregation.
BeamBagStateHandler handles operations on ListState, which backs Beam bag states.
BeamDataStreamPythonFunctionRunner is responsible for starting a beam python harness to execute user defined python function.
A BeamStateStore that returns keyed states based on BeamFnApi.StateRequest.
BeamMapStateHandler handles operations on a MapState.
A BeamStateStore that returns operator states based on BeamFnApi.StateRequest.
A BeamPythonFunctionRunner used to execute Python functions.
Interface for doing actual operations on Flink state based on BeamFnApi.StateRequest.
The handler for Beam state requests sent from Python side, which does actual operations on Flink state.
Interface for getting the underlying state based on Beam state request (keyed state or operator state).
A BeamTablePythonFunctionRunner used to execute Python functions in Table API.
We create the BigDecSerializer instead of using the BigDecSerializer of flink-core module for performance reasons in Python deserialization.
Serializer configuration snapshot for compatibility and format evolution.
ArrowFieldWriter for BigInt.
BigIntWriter for ArrayData input.
BigIntWriter for RowData input.
ArrowFieldWriter for Binary.
BinaryWriter for ArrayData input.
BinaryWriter for RowData input.
ArrowFieldWriter for Boolean.
BooleanWriter for ArrayData input.
BooleanWriter for RowData input.
Creates byte arrays (byte[]).
A utility class for converting byte[][] to String and String to byte[][].
A wrapper of the byte array.
The serializer of ByteArrayWrapper.
Serializer configuration snapshot for compatibility and format evolution.
Executor which will perform chaining optimization before generating the StreamGraph.
Utilities to clean up the leaking classes.
constants.
Deserialization schema from CSV to Flink types.
A builder for creating a CsvRowDeserializationSchema.
Serialization schema that serializes an object of Flink types into a CSV bytes.
A builder for creating a CsvRowSerializationSchema.
TimestampAssigner which extracts timestamp from the second field of the input element.
DataStreamPythonFunction maintains the serialized python function which will be used in BeamDataStreamPythonFunctionRunner.
DataStreamPythonFunctionInfo holds a PythonFunction and its function type.
Interface for Python DataStream operators.
Takes int instead of long as the serialized value.
Serializer configuration snapshot for compatibility and format evolution.
DateWriter for ArrayData input.
DateWriter for RowData input.
We create the DecimalSerializer instead of using the DecimalSerializer of flink-table-runtime for performance reasons in Python deserialization.
TypeSerializerSnapshot for DecimalDataSerializer.
ArrowFieldWriter for Decimal.
DecimalWriter for ArrayData input.
DecimalWriter for RowData input.
A JobBundleFactory for which the implementation can specify a custom EnvironmentFactory for environment management.
A container for EnvironmentFactory and its corresponding Grpc servers.
Holder for an SdkHarnessClient along with its associated state and data servers.
For those Transformation that don't have an operator entity, DelegateOperatorTransformation provides a SimpleOperatorFactory containing a DelegateOperatorTransformation.DelegateOperator , which can hold special configurations during transformation preprocessing for Python jobs, and later be queried at translation stage.
ArrowFieldWriter for Double.
DoubleWriter for ArrayData input.
DoubleWriter for RowData input.
The EmbeddedPythonBatchCoBroadcastProcessOperator is responsible for executing the Python CoBroadcastProcess Function under BATCH mode, EmbeddedPythonCoProcessOperator is used under STREAMING mode.
The EmbeddedPythonBatchKeyedCoBroadcastProcessOperator is responsible for executing the Python CoBroadcastProcess function under BATCH mode, EmbeddedPythonKeyedCoProcessOperator is used under STREAMING mode.
EmbeddedPythonCoProcessOperator is responsible for executing Python CoProcessFunction in embedded Python environment.
A PythonEnvironment for executing UDFs in embedded environment.
The base class of python environment manager which is used to create the PythonEnvironment object.
EmbeddedPythonKeyedCoProcessOperator is responsible for executing user defined python KeyedCoProcessFunction in embedded Python environment.
EmbeddedPythonKeyedProcessOperator is responsible for executing user defined python KeyedProcessFunction in embedded Python environment.
EmbeddedPythonProcessOperator is responsible for executing Python ProcessFunction in embedded Python environment.
The Python ScalarFunction operator in embedded Python environment.
The Python TableFunction operator in embedded Python environment.
EmbeddedPythonWindowOperator<K,IN,OUT,W extends org.apache.flink.table.runtime.operators.window.Window>
EmbeddedPythonWindowOperator is responsible for executing user defined python ProcessWindowFunction in embedded Python environment.
The ExternalPythonBatchCoBroadcastProcessOperator is responsible for executing the Python CoBroadcastProcess Function under BATCH mode, ExternalPythonCoProcessOperator is used under STREAMING mode.
The ExternalPythonBatchKeyedCoBroadcastProcessOperator is responsible for executing the Python CoBroadcastProcess function under BATCH mode, ExternalPythonKeyedCoProcessOperator is used under STREAMING mode.
The ExternalPythonCoProcessOperator is responsible for executing the Python CoProcess Function.
ExternalPythonKeyedCoProcessOperator is responsible for launching beam runner which will start a python harness to execute user defined python CoProcess function.
ExternalPythonKeyedProcessOperator is responsible for launching beam runner which will start a python harness to execute user defined python function.
ExternalPythonProcessOperator is responsible for launching beam runner which will start a python harness to execute user defined python ProcessFunction.
 
A representation of the coder
Protobuf type org.apache.flink.fn_execution.v1.CoderInfoDescriptor.ArrowType
Protobuf type org.apache.flink.fn_execution.v1.CoderInfoDescriptor.ArrowType
 
A representation of the coder
 
for Table & SQL
for Table & SQL
 
Protobuf enum org.apache.flink.fn_execution.v1.CoderInfoDescriptor.Mode
only used in batch over window the data consists of [window data][arrow data]
only used in batch over window the data consists of [window data][arrow data]
 
for DataStream
for DataStream
 
Protobuf type org.apache.flink.fn_execution.v1.CoderInfoDescriptor.RowType
Protobuf type org.apache.flink.fn_execution.v1.CoderInfoDescriptor.RowType
 
 
Protobuf type org.apache.flink.fn_execution.v1.GroupWindow
Protobuf type org.apache.flink.fn_execution.v1.GroupWindow
Protobuf enum org.apache.flink.fn_execution.v1.GroupWindow.WindowProperty
Protobuf enum org.apache.flink.fn_execution.v1.GroupWindow.WindowType
 
Protobuf type org.apache.flink.fn_execution.v1.Input
Protobuf type org.apache.flink.fn_execution.v1.Input
 
 
Protobuf type org.apache.flink.fn_execution.v1.JobParameter
Protobuf type org.apache.flink.fn_execution.v1.JobParameter
 
Used to describe the info of over window in pandas batch over window aggregation
Used to describe the info of over window in pandas batch over window aggregation
Protobuf enum org.apache.flink.fn_execution.v1.OverWindow.WindowType
 
A representation of the data schema.
Protobuf type org.apache.flink.fn_execution.v1.Schema.BinaryInfo
Protobuf type org.apache.flink.fn_execution.v1.Schema.BinaryInfo
 
A representation of the data schema.
Protobuf type org.apache.flink.fn_execution.v1.Schema.CharInfo
Protobuf type org.apache.flink.fn_execution.v1.Schema.CharInfo
 
Protobuf type org.apache.flink.fn_execution.v1.Schema.DecimalInfo
Protobuf type org.apache.flink.fn_execution.v1.Schema.DecimalInfo
 
Protobuf type org.apache.flink.fn_execution.v1.Schema.Field
Protobuf type org.apache.flink.fn_execution.v1.Schema.Field
 
Protobuf type org.apache.flink.fn_execution.v1.Schema.FieldType
Protobuf type org.apache.flink.fn_execution.v1.Schema.FieldType
 
 
Protobuf type org.apache.flink.fn_execution.v1.Schema.LocalZonedTimestampInfo
Protobuf type org.apache.flink.fn_execution.v1.Schema.LocalZonedTimestampInfo
 
Protobuf type org.apache.flink.fn_execution.v1.Schema.MapInfo
Protobuf type org.apache.flink.fn_execution.v1.Schema.MapInfo
 
Protobuf type org.apache.flink.fn_execution.v1.Schema.TimeInfo
Protobuf type org.apache.flink.fn_execution.v1.Schema.TimeInfo
 
Protobuf type org.apache.flink.fn_execution.v1.Schema.TimestampInfo
Protobuf type org.apache.flink.fn_execution.v1.Schema.TimestampInfo
 
Protobuf enum org.apache.flink.fn_execution.v1.Schema.TypeName
Protobuf type org.apache.flink.fn_execution.v1.Schema.VarBinaryInfo
Protobuf type org.apache.flink.fn_execution.v1.Schema.VarBinaryInfo
 
Protobuf type org.apache.flink.fn_execution.v1.Schema.VarCharInfo
Protobuf type org.apache.flink.fn_execution.v1.Schema.VarCharInfo
 
Protobuf type org.apache.flink.fn_execution.v1.Schema.ZonedTimestampInfo
Protobuf type org.apache.flink.fn_execution.v1.Schema.ZonedTimestampInfo
 
 
A representation of State
A representation of State
Protobuf type org.apache.flink.fn_execution.v1.StateDescriptor.StateTTLConfig
Protobuf type org.apache.flink.fn_execution.v1.StateDescriptor.StateTTLConfig
TTL cleanup strategies.
TTL cleanup strategies.
Protobuf enum org.apache.flink.fn_execution.v1.StateDescriptor.StateTTLConfig.CleanupStrategies.EmptyCleanupStrategy
Configuration of cleanup strategy while taking the full snapshot.
Configuration of cleanup strategy while taking the full snapshot.
 
Protobuf type org.apache.flink.fn_execution.v1.StateDescriptor.StateTTLConfig.CleanupStrategies.MapStrategiesEntry
Protobuf type org.apache.flink.fn_execution.v1.StateDescriptor.StateTTLConfig.CleanupStrategies.MapStrategiesEntry
 
 
Configuration of cleanup strategy using custom compaction filter in RocksDB.
Configuration of cleanup strategy using custom compaction filter in RocksDB.
 
Fixed strategies ordinals in strategies config field.
 
This option configures whether expired user value can be returned or not.
This option configures time scale to use for ttl.
This option value configures when to update last access timestamp which prolongs state TTL.
 
 
A representation of the data type information in DataStream.
Protobuf type org.apache.flink.fn_execution.v1.TypeInfo.AvroTypeInfo
Protobuf type org.apache.flink.fn_execution.v1.TypeInfo.AvroTypeInfo
 
A representation of the data type information in DataStream.
Protobuf type org.apache.flink.fn_execution.v1.TypeInfo.MapTypeInfo
Protobuf type org.apache.flink.fn_execution.v1.TypeInfo.MapTypeInfo
 
Protobuf type org.apache.flink.fn_execution.v1.TypeInfo.RowTypeInfo
Protobuf type org.apache.flink.fn_execution.v1.TypeInfo.RowTypeInfo
Protobuf type org.apache.flink.fn_execution.v1.TypeInfo.RowTypeInfo.Field
Protobuf type org.apache.flink.fn_execution.v1.TypeInfo.RowTypeInfo.Field
 
 
Protobuf type org.apache.flink.fn_execution.v1.TypeInfo.TupleTypeInfo
Protobuf type org.apache.flink.fn_execution.v1.TypeInfo.TupleTypeInfo
 
 
Protobuf enum org.apache.flink.fn_execution.v1.TypeInfo.TypeName
 
Protobuf type org.apache.flink.fn_execution.v1.UserDefinedAggregateFunction
Protobuf type org.apache.flink.fn_execution.v1.UserDefinedAggregateFunction
Protobuf type org.apache.flink.fn_execution.v1.UserDefinedAggregateFunction.DataViewSpec
Protobuf type org.apache.flink.fn_execution.v1.UserDefinedAggregateFunction.DataViewSpec
 
Protobuf type org.apache.flink.fn_execution.v1.UserDefinedAggregateFunction.DataViewSpec.ListView
Protobuf type org.apache.flink.fn_execution.v1.UserDefinedAggregateFunction.DataViewSpec.ListView
 
Protobuf type org.apache.flink.fn_execution.v1.UserDefinedAggregateFunction.DataViewSpec.MapView
Protobuf type org.apache.flink.fn_execution.v1.UserDefinedAggregateFunction.DataViewSpec.MapView
 
 
 
A list of the user-defined aggregate functions to be executed in a group aggregate operation.
A list of the user-defined aggregate functions to be executed in a group aggregate operation.
 
User defined DataStream function definition.
User defined DataStream function definition.
Protobuf enum org.apache.flink.fn_execution.v1.UserDefinedDataStreamFunction.FunctionType
Protobuf type org.apache.flink.fn_execution.v1.UserDefinedDataStreamFunction.RuntimeContext
Protobuf type org.apache.flink.fn_execution.v1.UserDefinedDataStreamFunction.RuntimeContext
 
 
User-defined function definition.
User-defined function definition.
 
A list of user-defined functions to be executed in a batch.
A list of user-defined functions to be executed in a batch.
 
Helper class for forwarding Python metrics to Java accumulators and metrics.
Flink Gauge for DistributionResult.
Flink Gauge for GaugeResult.
ArrowFieldWriter for Float.
FloatWriter for ArrayData input.
FloatWriter for RowData input.
An implementation of the Beam Fn State service.
Helper class to create a HashMap taking Numeric data as key or value from Python side.
Partitioner that partitions by id.
IntWriter for ArrayData input.
IntWriter for RowData input.
The type of the Python map state iterate request.
Deserialization schema from JSON to Flink types.
Serialization schema that serializes an object of Flink types into a JSON bytes.
KeyByKeySelector is responsible for extracting the first field of the input row as key.
A TypeSerializer for MapData.
TypeSerializerSnapshot for MapDataSerializer.
MapWriter for ArrayData input.
MapWriter for RowData input.
Flink Gauge for Python Distribution.
Flink Gauge for Python Gauge.
The PartitionCustomKeySelector will return the first field of the input row value.
A PickledByteArrayTypeInfo indicates that the data of this type is a generated primitive byte array by pickle.
A PythonEnvironment for executing UDFs in Process.
The ProcessPythonEnvironmentManager is used to prepare the working dir of python UDF worker and create ProcessPythonEnvironment object of Beam Fn API.
Utilities used to construct protobuf objects or construct objects from protobuf objects.
Utility class that contains helper methods to create a TableSource from a file which contains Python objects.
A Transformation representing a Python Co-Broadcast-Process operation, which will be translated into different operations by PythonBroadcastStateTransformationTranslator.
Configurations for the Python job which are used at run time.
A Util class to handle the configurations of Python jobs.
Utility class for using DataStream connectors in Python.
The serializable InvocationHandler as the proxy for first column selector.
A ProcessFunction that convert Row to RowData.
A SerializationSchema for Row that only serialize the second column using a wrapped SerializationSchema for PythonConnectorUtils.SecondColumnSerializationSchema.
Utilities for using CSV format in PyFlink.
PythonDependencyInfo contains the information of third-party dependencies.
Utility class for Python dependency management.
A main class used to launch Python applications.
Table source factory for PythonDynamicTableSource.
Options for PythonDynamicTableSource.
Implementation of ScanTableSource for python elements table.
The base interface of python environment for executing UDFs.
The base interface of python environment manager which is used to create the PythonEnvironment object used to execute Python functions.
Utils used to prepare the python environment.
The factory which creates the PythonFunction objects from given module name and object name.
The cache key.
Default implementation of PythonFunctionFactory.
The base interface of runner which is responsible for the execution of Python functions.
The Py4j Gateway Server provides RPC service for user's python process.
A simple watch dog interface.
A Transformation representing a Python Keyed-Co-Broadcast-Process operation, which will be translated into different operations by PythonKeyedBroadcastStateTransformationTranslator.
A util class which attempts to chain all available Python functions.
Utilities used by Python operators.
Configuration options for the Python API.
The class for command line options that refer to a Python program or JAR program with Python command line options.
The Python ScalarFunction operator.
The set of resources that can be shared by all the Python operators in a slot.
Command line parser for Python shell.
The Python AggregateFunction operator.
The Python TableAggregateFunction operator.
PythonStreamGroupWindowAggregateOperator<K,W extends org.apache.flink.table.runtime.operators.window.Window>
The Python Group Window AggregateFunction operator.
The Python TableFunction operator.
Python utilities.
A util class for converting the given TypeInformation to other objects.
Utilities for converting Flink logical types, such as convert it to the related TypeSerializer or ProtoType.
The element in the Object Array will be converted to the corresponding Data through element DataConverter.
The element in the Object Array will be converted to the corresponding Data through element DataConverter.
Python Long will be converted to Long in PemJa, so we need ByteDataConverter to convert Java Long to internal Byte.
Python Long will be converted to Long in PemJa, so we need ByteDataConverter to convert Java Long to internal Byte.
Data Converter that converts the data to the format data which can be used in PemJa.
Data Converter that converts the data to the java format data which can be used in PemJa.
Python Float will be converted to Double in PemJa, so we need FloatDataConverter to convert Java Double to internal Float.
Python Float will be converted to Double in PemJa, so we need FloatDataConverter to convert Java Double to internal Float.
Identity data converter.
Identity data converter.
Python Long will be converted to Long in PemJa, so we need IntDataConverter to convert Java Long to internal Integer.
Python Long will be converted to Long in PemJa, so we need IntDataConverter to convert Java Long to internal Integer.
The element in the List will be converted to the corresponding Data through element DataConverter.
Converter That convert the logicalType to the related Prototype.
The key/value in the Map will be converted to the corresponding Data through key/value DataConverter.
The key/value in the Map will be converted to the corresponding Data through key/value DataConverter.
Row Data will be converted to the Object Array [RowKind(as Long Object), Field Values(as Object Array)].
RowData will be converted to the Object Array [RowKind(as Long Object), Field Values(as Object Array)].
RowData Data will be converted to the Object Array [RowKind(as Long Object), Field Values(as Object Array)].
Python Long will be converted to Long in PemJa, so we need ShortDataConverter to convert Java Long to internal Short.
Python Long will be converted to Long in PemJa, so we need ShortDataConverter to convert Java Long to internal Short.
Python datetime.time will be converted to Time in PemJa, so we need TimeDataConverter to convert Java Double to internal Integer.
Tuple Data will be converted to the Object Array.
Get DataConverter according to the given typeInformation.
Get coder proto according to the given type information.
Get serializers according to the given typeInformation.
MapFunction which removes the timestamp field from the input element.
A TypeSerializer for RowData.
TypeSerializerSnapshot for RowDataSerializer.
RowWriter for ArrayData input.
RowWriter for RowData input.
Output collector for Python UDF runner.
A gRPC server factory.
Creates a gRPC Server using the default server factory.
Factory that constructs client-accessible URLs from a local server address and port.
A holder for shared resource singletons.
Defines a resource, and the way to create and destroy instances of it.
ArrowFieldWriter for SmallInt.
SmallIntWriter for ArrayData input.
SmallIntWriter for RowData input.
StreamArrowPythonGroupWindowAggregateFunctionOperator<K,W extends org.apache.flink.table.runtime.operators.window.Window>
The Stream Arrow Python AggregateFunction Operator for Group Window Aggregation.
The Stream Arrow Python AggregateFunction Operator for ROWS clause proc-time bounded OVER window.
The Stream Arrow Python AggregateFunction Operator for ROWS clause proc-time bounded OVER window.
The Stream Arrow Python AggregateFunction Operator for RANGE clause event-time bounded OVER window.
The Stream Arrow Python AggregateFunction Operator for RANGE clause event-time bounded OVER window.
The collector is used to convert a RowData to a StreamRecord.
We create the StringSerializer instead of using the StringSerializer of flink-core module because the StringSerializer of flink-core module serialize every Char of String in serialize method and deserialize the Char to build the String.
Serializer configuration snapshot for compatibility and format evolution.
Utilities to handle triggered timer.
Handles the interaction with the Python worker for registering and deleting timers.
TimerRegistrationAction used to register Timer.
Utilities for timer.
Uses int instead of long as the serialized value.
Serializer configuration snapshot for compatibility and format evolution.
Uses similar serialization/deserialization of SqlTimestampSerializer to serialize Timestamp.
TypeSerializerSnapshot for TimestampSerializer.
ArrowFieldWriter for Timestamp.
TimestampWriter for ArrayData input.
TimestampWriter for RowData input.
TimeWriter for ArrayData input.
TimeWriter for RowData input.
ArrowFieldWriter for TinyInt.
TinyIntWriter for ArrayData input.
TinyIntWriter for RowData input.
ArrowFieldWriter for VarBinary.
VarBinaryWriter for ArrayData input.
VarBinaryWriter for RowData input.
ArrowFieldWriter for VarChar.
VarCharWriter for ArrayData input.
VarCharWriter for RowData input.