-
Notifications
You must be signed in to change notification settings - Fork 441
Language Extensibility
There are two main options for language extensibility in Azure Functions.
- Custom handler - Recommended for most scenarios when you want to write Azure Functions in any language. Communicates with the Azure Functions host over HTTP. To learn more, see Azure Functions custom handlers documentation.
- gRPC language worker - Recommended if you want to build a custom Azure Functions language worker. Communicates with the Azure Functions host using gRPC.
The rest of this article describes the gRPC language worker.
In order to bring Language Extensibility to functions, we've split the functions runtime into the host (which manages function events), and language worker processes (which runs user functions for a given language). These two pieces communicate using gRPC
as a messaging layer.
- https://github.com/Azure/azure-functions-nodejs-worker
- https://github.com/Azure/azure-functions-java-worker
- https://github.com/Azure/azure-functions-python-worker
We are using Microsoft.Extensions.Configuration
in order to handle the worker configuration. Here's some sample configuration for node.
{
languageWorkers: {
node: {
workerDirectoy: <path to parent folder that has the worker>
arguments: <debugging port> // --inspect=5858
... other worker specific settings
}
}
}
This allows the above json configuration, as well as configuration via:
- Command line,
/languageWorkers:node:arguments= --inspect=5858
- Environment variable,
languageWorkers:node:workerDirectory = "d:/git/worker"
For gRPC
, we currently have one call which creates a duplex stream of StreamingMessages
, which are the protobuf messages defined below. All messages between host and worker flow across this bi-directional stream.
The host runs the gRPC
server, and the language worker processes act as clients.
service FunctionRpc {
rpc EventStream (stream StreamingMessage) returns (stream StreamingMessage) {}
}
Every message that crosses the channel is a StreamingMessage
, which contains a request id and a content
union that contains one sub-message.
This contract lives here: https://github.com/Azure/azure-functions-language-worker-protobuf/blob/dev/src/proto/FunctionRpc.proto
message StreamingMessage {
string request_id = 1;
oneof content {
// Worker signals to host that it has been started
StartStream start_stream = 20;
// Host sends capabilities/init data to worker
WorkerInitRequest worker_init_request = 17;
// Worker responds after initializing with its capabilities & status
WorkerInitResponse worker_init_response = 16;
// Host sends required metadata to worker to load function
FunctionLoadRequest function_load_request = 8;
// Worker responds after loading with the load result
FunctionLoadResponse function_load_response = 9;
// Host sends invocation information (function id, binding data, parameters) to worker
InvocationRequest invocation_request = 4;
// Worker sends response to host
InvocationResponse invocation_response = 5;
// Structured log from the worker based off of the ILogger interface
RpcLog rpc_log = 2;
}
}
At it's core, the language worker needs to provide event handlers for streaming messages. Here's an example in node:
eventStream.on('data', (msg) => {
let event = <string>msg.content; // find the type of streaming message
let eventHandler = (<any>this)[event]; // find the handler on the class for the streaming message type
if (eventHandler) {
eventHandler.apply(this, [msg.requestId, msg[event]]); // invoke the handler with the id and message
} else {
console.error(`Worker ${workerId} had no handler for message '${event}'`)
}
});
...
// example of a streaming message handler for workerInitRequest
public workerInitRequest(requestId: string, msg: rpc.WorkerInitRequest) {
// nothing to be done, so write the workerInitResponse back to the stream
this._eventStream.write({
requestId: requestId,
workerInitResponse: {
result: {
status: Status.Success
}
}
});
}
More message handler implementations here: https://github.com/Azure/azure-functions-nodejs-worker/blob/dev/src/WorkerChannel.ts
- Start Host
- Start
gRPC
server - For each function, Host creates a Worker (via
IWorkerProvider
) if necessary. One Worker process per language. -
StartStream
- Worker connects to HostgRPC
server -
WorkerInitRequest
,WorkerInitResponse
- Host and Worker exchange version info & capabilities -
FunctionLoadRequest
,FunctionLoadResponse
- Host sends Worker function metadata, Worker loads function -
InvocationRequest
,InvocationResponse
- Host sends function id, binding data, parameters, Worker response with function result
More details on on the design here: https://github.com/Azure/azure-functions-powershell-worker/blob/dev/docs/designs/PowerShell-AzF-Overall-Design.md
- Configuration Settings
- function.json
- host.json
- host.json (v2)
- Http Functions
- Function Runtime Versioning
- Official Functions developers guide
- Host Health Monitor
- Managing Connections
- Renaming a Function
- Retrieving information about the currently running function
- Site Extension Resolution
- Linux Consumption Regions
- Using LinuxFxVersion for Linux Function apps
- Out-of-proc Cancellation Tokens
- Assembly Resolution in Azure Functions
- ILogger
- Precompiled functions
- Official Functions C# developer reference
- Contributor Onboarding
- Development Process
- Deploying the Functions runtime as a private site extension
- Authoring & Testing Language Extensions
- Bindings in out-of-proc
- Language Extensibility
- Worker Capabilities
- Investigating and reporting issues with timer triggered functions not firing
- Sharing Your Function App name privately
- Azure Functions CLI release notes [moved here]
- Function App Zipped Deployment [deprecated]