[TOC]
A number of IPC messages sent (primarily between the browser and renderer
processes) are still defined using Chrome's old IPC system in //ipc
. This
system uses
base::Pickle
as the basis for message serialization and is supported by a number if IPC_*
preprocessor macros defined in //ipc
and used around the source tree.
There is an ongoing, distributed effort to get these messages converted to Mojo interface messages. Messages that still need to be converted are tracked in two spreadsheets:
- Chrome IPC to Mojo migration for non-web platform messages
- Mojoifying Platform Features for web platform messages
This document is concerned primarily with rote conversion of legacy IPC messages to Mojo interface messages. If you are considering more holistic refactoring and better isolation of an entire subsystem of the browser, you may consider servicifying the feature instead of merely converting its IPCs.
See other Mojo & Services documentation for introductory guides, API references, and more.
Each Content child process has a single IPC::Channel
implementation going
between it and the browser process, and this is used as the sole two-way FIFO
to send legacy IPC messages between the processes.
There are two fundamental types of legacy IPC messages: control messages,
defined via IPC_MESSAGE_CONTROLn
macros (where n
is some small integer) and
routed messages defined via IPC_MESSAGE_ROUTEDn
macros.
Control messages generally go between a browser-side process host (e.g.,
RenderProcessHost
or GpuProcessHost
) and the child-side ChildThreadImpl
subclass. All of these classes implement IPC::Sender
and thus have a Send
method for sending a control message to their remote counterpart, and they
implement IPC::Listener
to receive incoming control messages via
OnMessageReceived
.
Routed messages are relegated to routes which have arbitrary meaning
determined by their use within a given process. For example, renderers use
routes to isolate messages scoped to individual render frames, and so such
routed messages will travel between a RenderFrameHostImpl
and its
corresponding RenderFrameImpl
, both of which also implement IPC::Sender
and
IPC::Listener
.
Routed messages in the old IPC system always carry a routing ID to identify
to the receiving endpoint which routed object (e.g. which RenderFrameImpl
or RenderViewImpl
or whatever) the message is targeting. Each endpoint is thus
required to do some additional book-keeping to track what each routing ID means.
Mojo interfaces obviate the need for routing IDs, as new "routes" can be established by simply creating a new interface pipe and passing one endpoint to something which knows how to bind it.
When thinking about an IPC message conversion to Mojo, it's important to
consider whether the message is a control message or a routed message, as this
determines where you might find an existing Mojo interface to carry your
message, or where you will want to add a new end-to-end Mojo interface for that
purpose. This can mean the difference between a single per-process interface
going between each RenderProcessHostImpl
and its corresponding
RenderThreadImpl
, vs a per-frame interface going between each
RenderFrameHostImpl
and its corresponding RenderFrameImpl
.
One very important consideration when doing IPC conversions is the relative ordering of IPC-driven operations. With the old IPC system, because every message between two processes is globally ordered, it is quite easy for parts of the system to (intentionally or often unintentionally) rely on strict ordering guarantees.
For example, imagine a WebContentsObserver
in the browser processes observes
a frame navigation and immediately sends an IPC message to the frame to
configure some new behavior. The implementation may be inadvertently relying on
this message arriving before some other tangentially related message sent to
the same frame shortly after the same navigation event.
Mojo does not (and in fact cannot) make any strict ordering guarantees between separate message pipes, as message pipes may be freely moved across process boundaries and thus cannot necessarily share a common FIFO at all times.
If the two messages described above were moved to separate Mojo interfaces on separate message pipes, renderer behavior could break as the first message may arrive after the second message.
The best solution to this problem is to rethink the IPC surface and/or implementation on either side to eliminate ordering dependencies between two interfaces that otherwise seem to be logically distinct. Failing that, Mojo's solution to this problem is to support associated interfaces. In a nutshell, these allow multiple distinct interfaces to be multiplexed over a shared message pipe.
The previous section mentions associated interfaces as a general-purpose solution for establishing a mutual FIFO between multiple logical Mojo interfaces by having them share a single message pipe.
In Chrome, the IPC::Channel
which carries all legacy IPC messages between
two processes is itself a Mojo message pipe. We provide a mechanism for
associating arbitrary Mojo interfaces with this pipe, which means messages can
be converted to Mojo while preserving strict FIFO with respect to other legacy
IPC messages. Such interfaces are designated in Chrome parlance as
Channel-associated interfaces.
*** aside NOTE: Channel-associated interface acquisition is not constrained by the Service Manager in any way, so security reviewers need to be careful to inspect new additions and uses of such interfaces.
Usage of Channel-associated interfaces should be rare but is considered a reasonable intermediate solution for incremental IPC conversions where it would be too risky or noisy to convert a large IPC surface all at once, but it would also be impossible to split the IPC surface between legacy IPC and a dedicated Mojo interface pipe without introducing timing bugs.
At this point in Chrome's development, practical usage of Channel-associated
interfaces is restricted to the IPC::Channel
between the browser process and
a renderer process, as this is the most complex IPC surface with the most
implicit ordering dependencies. A few simple APIs exist to support this.
RenderProcessHostImpl
owns an IPC::Channel
to its corresponding
RenderThreadImpl
in the render process. This object has a
GetRemoteAssociatedInterfaces
method which can be used to pass arbitrary
associated interface requests:
mojo::PendingAssociatedRemote<magic::mojom::GoatTeleporter> teleporter;
channel_->GetRemoteAssociatedInterfaces()->GetInterface(teleporter.BindNewEndpointAndPassReceiver());
// These messages are all guaranteed to arrive in the same order they were sent.
channel_->Send(new FooMsg_SomeLegacyIPC);
teleporter->TeleportAllGoats();
channel_->Send(new FooMsg_AnotherLegacyIPC);
Likewise, ChildThreadImpl
has an IPC::Channel
that can be used in the same
way to send such messages back to the browser.
To receive and bind incoming Channel-associated interface requests, the above
objects also implement IPC::Listener::OnAssociatedInterfaceRequest
.
For supplementation of routed messages, both RenderFrameHostImpl
and
RenderFrameImpl
define a GetRemoteAssociatedInterfaces
method which works
like the one on IPC::Channel
, and both objects also implement
IPC::Listener::OnAssociatedInterfaceRequest
for processing incoming associated
interface requests specific to their own frame.
There are some example conversion CLs which use Channel-associated interfaces here and here.
There are a few questions you should ask before embarking upon any IPC message conversion journey, and there are many potential approaches to consider. The right one depends on context.
Note that this section assumes the message is traveling between the browser process and a renderer process. Other cases are rare and developers may wish to consult [email protected] before proceeding with them. Otherwise, apply the following basic algorithm to decide how to proceed:
- General note: If the message is a reply to some other message (typically these take a "request ID" argument), see the note about message replies at the bottom of this section.
- Consider whether or not the message makes sense as part of the IPC surface of
a new or existing service somewhere in
//services
or//chrome/services
, etc. This is less and less likely to be the case as time goes on, as many remaining IPC conversions are quite narrowly dealing with specific browser/renderer details rather than the browser's supporting subsystems. If defining a new service, you may wish to consult some of the other Mojo & Services documentation first. - If the message is an
IPC_MESSAGE_CONTROL
message:- If there are likely to be strict ordering requirements between this
message and other legacy IPC or Channel-associated interface messages,
consider using a new or existing
Channel-associated interface between
RenderProcessHostImpl
andRenderThreadImpl
. - If the message is sent from a renderer to the browser:
- If an existing interface is bound by
RenderProcessHostImpl
and requested throughRenderThread
's Connector and seems to be a good fit for the message, add the equivalent Mojo message to that interface. - If no such interface exists, consider adding one for this message and any related messages.
- If an existing interface is bound by
- If the message is sent from the browser to a renderer:
- If an existing interface is bound by
RenderThreadImpl
and requested through aBrowserContext
Connector referencing a specificRenderProcessHost
identity, and the interface seems to be a good fit for the message, add the equivalent Mojo message to that interface. - If no such interface exists, consider adding one for this message and any related messages.
- If an existing interface is bound by
- If there are likely to be strict ordering requirements between this
message and other legacy IPC or Channel-associated interface messages,
consider using a new or existing
Channel-associated interface between
- If the message is an
IPC_MESSAGE_ROUTED
message:- Determine what the routing endpoints are. If they are
RenderFrameHostImpl
andRenderFrameImpl
:- If there are likely to be strict ordering requirements between this
message and other legacy IPC or Channel-associated interface messages,
consider using a new or existing
Channel-associated interface between
RenderFrameHostImpl
andRenderFrameImpl
. - If the message is sent from a renderer to the browser:
- If an existing interface is bound by
RenderFrameHostImpl
and acquired viaRenderFrame::GetBrowserInterfaceBroker
and the interface seems to be a good fit for this message, add the equivalent Mojo message to that interface. - If no such interface exists, consider adding one and registering it
with
RenderFrameHostImpl
'sBrowserInterfaceBroker
. See the simple example in the "Intro to Mojo & Services" document.
- If an existing interface is bound by
- If the message is sent from the browser to a renderer, consider
adding a Mojo equivalent to the
content.mojom.Frame
interface defined here.
- If there are likely to be strict ordering requirements between this
message and other legacy IPC or Channel-associated interface messages,
consider using a new or existing
Channel-associated interface between
- If the routing endpoints are not frame objects (for example, they may
be
RenderView
/RenderViewHost
objects), this is a special case which does not yet have an easy conversion approach readily available. Contact [email protected] to propose or discuss options.
- Determine what the routing endpoints are. If they are
*** aside NOTE: If you are converting a sync IPC, see the section on Synchronous Calls in the Mojo documentation.
If the message is a reply, meaning it has a "request ID" which correlates it to a prior message in the opposite direction, consider converting the request message following the algorithm above. Unlike with legacy IPC, Mojo messages support replies as a first-class concept. So for example if you have:
IPC_CONTROL_MESSAGE2(FooHostMsg_DoTheThing,
int /* request_id */,
std::string /* name */);
IPC_CONTROL_MESSAGE2(FooMsg_DidTheThing,
int /* request_id */,
bool /* success */);
You should consider defining an interface Foo
which is bound in
RenderProcessHostImpl
and acquired from RenderThreadImpl
, with the following
mojom definition:
interface Foo {
DoTheThing(string name) => (bool success);
};
See Receiving responses for more information.
Occasionally it is useful to do partial IPC conversions, where you want to
convert a message to a Mojo interface method but you don't want to necessarily
convert every structure passed by the message. In this case, you can leverage
Mojo's
type-mapping
system to repurpose existing IPC::ParamTraits
.
*** aside
NOTE: Although in some cases IPC::ParamTraits<T>
specializations are
defined manually in library code, the IPC_STRUCT*
macro helpers also define
IPC::ParamTraits<T>
specializations under the hood. All advice in this section
pertains to both kinds of definitions.
If a mojom struct is declared without a struct body and is tagged with
[Native]
, and a corresponding typemap is provided for the struct, the emitted
C++ bindings will -- as if by magic -- replace the mojom type with the
typemapped C++ type and will internally use the existing IPC::ParamTraits<T>
specialization for that type in order to serialize and deserialize the struct.
For example, given the
resource_messages.h
header
which defines an IPC mapping for content::ResourceRequest
:
IPC_STRUCT_TRAITS_BEGIN(content::ResourceRequest)
IPC_STRUCT_TRAITS_MEMBER(method)
IPC_STRUCT_TRAITS_MEMBER(url)
// ...
IPC_STRUCT_TRAITS_END()
and the
resource_request.h
header
which actually defines the content::ResourceRequest
type:
namespace content {
struct CONTENT_EXPORT ResourceRequest {
// ...
};
} // namespace content
we can declare a corresponding "native" mojom struct:
module content.mojom;
[Native]
struct URLRequest;
and add a typemap like
url_request.typemap
to define how to map between them:
mojom = "//content/public/common/url_loader.mojom"
public_headers = [ "//content/common/resource_request.h" ]
traits_headers = [ "//content/common/resource_messages.h" ]
...
type_mappings = [ "content.mojom.URLRequest=content::ResourceRequest" ]
Note specifically that public_headers includes the definition of the native C++ type, and traits_headers includes the definition of the legacy IPC traits.
As a result of all this, other mojom files can now reference
content.mojom.URLRequest
as a type for method parameters and other struct
fields, and the generated C++ bindings will represent those values exclusively
as content::ResourceRequest
objects.
This same basic approach can be used to leverage existing IPC_ENUM_TRAITS
for
invocations for [Native]
mojom enum aliases.
*** aside
NOTE: Use of [Native]
mojom definitions is strictly limited to C++
bindings. If a mojom message depends on such definitions, it cannot be sent or
received by other language bindings. This feature also depends on continued
support for legacy IPC serialization and all uses of it should therefore be
treated as technical debt.
Let's assume we have a mojom file such as this:
module example.mojom;
interface Foo {
SendData(string param1, array<int32> param2);
};
The following GN snippet will generate two concrete targets: example
and
example_blink
:
mojom("example") {
sources = [ "example.mojom" ]
}
The target example
will generate Chromium-style C++ bindings using STL types:
// example.mojom.h
namespace example {
namespace mojom {
class Example {
virtual void SendArray(const std::string& param1, const std::vector<int32_t>& param2) = 0;
}
} // namespace mojom
} // namespace example
The target example_blink
will generate Blink-style C++ bindings using WTF types:
// example.mojom-blink.h
namespace example {
namespace mojom {
namespace blink {
class Example {
virtual void SendArray(const WTF::String& param1, const WTF::Vector<int32_t>& param2) = 0;
}
} // namespace blink
} // namespace mojom
} // namespace example
Thanks to these separate sets of bindings no work is necessary to convert types between Blink-style code and Chromium-style code. It is handled automatically during message serialization and deserialization.
For more information about variants, see this section of the C++ bindings documentation.
Mojo methods that return a value take an instance of base::OnceCallback
.
Use WTF::BindOnce()
and an appropriate wrapper function depending on the type of
object and the callback.
For garbage-collected (Oilpan) classes owning the mojo::Remote
, it is recommended
to use WrapWeakPersistent(this)
for connection error handlers since they
are not guaranteed to get called in a finite time period (wrapping the object
with WrapPersistent
in this case would cause memory leaks).
If the response can be discarded in case the object is not alive by the time
the response is received, use WrapWeakPersistent(this)
for binding the response callback:
// src/third_party/blink/renderer/modules/device_orientation/device_sensor_entry.cc
sensor_.set_connection_error_handler(WTF::BindOnce(
&DeviceSensorEntry::HandleSensorError, WrapWeakPersistent(this)));
sensor_->ConfigureReadingChangeNotifications(/*enabled=*/false);
sensor_->AddConfiguration(
std::move(config), WTF::BindOnce(&DeviceSensorEntry::OnSensorAddConfiguration,
WrapWeakPersistent(this)));
Otherwise (for example, if the response callback is used to resolve a Promise),
use WrapPersistent(this)
to keep the object alive:
// src/third_party/blink/renderer/modules/nfc/nfc.cc
ScriptPromiseResolver* resolver = ScriptPromiseResolver::Create(script_state);
...
nfc_->CancelAllWatches(WTF::BindOnce(&NFC::OnRequestCompleted,
WrapPersistent(this),
WrapPersistent(resolver)));
Non-garbage-collected objects can use WTF::Unretained(this)
for both response
and error handler callbacks when the mojo::Remote
is owned by the object bound
to the callback or the object is guaranteed to outlive the Mojo connection for
another reason. Otherwise a weak pointer should be used. However, it is not a
common pattern since using Oilpan is recommended for all Blink code.
Only a mojo::Receiver
or mojo::ReceiverSet
should be used when implementing a
Mojo interface in an Oilpan-managed object. The object must then have a pre-finalizer
to close any open pipes when the object is about to be swept as lazy sweeping
means that it may be invalid long before the destructor is called. This requires
setup in both the object header and implementation.
// MyObject.h
class MyObject : public GarbageCollected,
public example::mojom::blink::Example {
USING_PRE_FINALIZER(MyObject, Dispose);
public:
MyObject();
void Dispose();
// Implementation of example::mojom::blink::Example.
private:
mojo::Receiver<example::mojom::blink::Example> m_receiver{this};
};
// MyObject.cpp
void MyObject::Dispose() {
m_receiver.Close();
}
For more information about Blink's Garbage Collector, see Blink GC API Reference.
Using typemapping for messages that go between Blink and content browser code can sometimes be tricky due to things like dependency cycles or confusion over the correct place for some definition to live. There are some example CLs provided here, but feel free to also contact [email protected] with specific details if you encounter trouble.
This CL introduces a Mojom
definition and typemap for ui::WindowOpenDisposition
as a precursor to the
IPC conversion below.
The follow-up CL uses that
definition along with several other new typemaps (including native typemaps as
described above) to convert the relatively large ViewHostMsg_CreateWindow
message to Mojo.
If this document was not helpful in some way, please post a message to your friendly [email protected] mailing list.