Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

FFI: Foreign Function Interface #76

Open
bengl opened this issue Jun 25, 2020 · 2 comments
Open

FFI: Foreign Function Interface #76

bengl opened this issue Jun 25, 2020 · 2 comments

Comments

@bengl
Copy link
Member

bengl commented Jun 25, 2020

What does this mean? (Background)

Foreign Function Interfaces are ways to call into native (i.e. C/C++/Rust/etc.) code from a higher level language, in our case JavaScript. There are many reasons to call native code:

  • Some code currently cannot target WASI (for example, code making system calls without counterparts in WASI).
  • Some C libraries are closed-source.
  • Performance concerns.
  • Many more!

We currently have a few approaches available to us:

  1. Node addons, including those using NAPI (C or C++ versions) or NAN.
  2. Userland FFI libraries. (e.g. ffi-napi, it's predecessor node-ffi, and others)

Userland FFI libraries typically depend on some C FFI library (such as libffi or dyncall) to build up the native function call using the appropriate calling conventions dynamically, which isn't otherwise possible in C/C++. Another approach (taken by DragonFFI for example) is to JIT compile function calls.

The advantage of userland FFI libraries is that they don't require the user to write any wrapper native code in order to use a given native function. The user just identifies the function they're trying to call and its signature.

The advantage of node addons is that they are typically quite a bit faster that FFI libraries.

New developments

A new API, v8-fast-api-calls.h is being added to V8, which enables V8 to call native code directly in optimized functions. While it's still incomplete, it's usable today on node@master or nightlies. It's currently used in node@master to call uv_hrtime as fast as currently possible.

Since the fast calls happen only when the functions are optimized, the slow path must also be provided.

Experiments

In an almost quixotical quest for better FFI performance, I've done some experiments on this front. I've made a branch of my FFI library, sbffi which uses v8-fast-api-calls.h and the performance was measurably good. Keep in mind it's still using an FFI technique (in this case using libtcc to JIT compile a function call), wheras a better implementation might pass the native function directly in as the fast path.

A thing about sbffi's implementation details: I originally built this library on the idea that arguments and return values could be stored in a shared buffer to avoid any marshalling overhead. This means building function calls dynamically from values in that buffer and a known signature. This adds a layer of abstraction in between, and requires putting arguments into heap-allocated memory and retrieving it from there, rather than more optimal paths for the data to take.

There are benchmarks provided. I can post results here if desired, but YMMV, and I don't think they're representative of the best we can do.

Why in Node.js core? Why now?

Since the fast API calls are now available, it may time to investigate whether they can be used to build an FFI system in Node.js core. I think it's possible to build it so that it approaches (or maybe even surpasses) the speed of node addons. With that kind of speed available, it makes sense to have first-class support for it, alongside the existing support of addons. This will enable end users to use native libraries without requiring them to write non-JS code do so, or have any extra compile step.

I can imagine this simplifying calling out to things like libsass.

Challenges

  1. The fast API calls API is currently slightly incomplete. Not all types (not even all JS-compatible types) are supported. This limits the utility at the moment, unless arguments and return values are passed through a shared buffer, like in sbffi, but that carries some performance cost.
  2. The slow path must still be provided which means either building up a dynamic call (i.e. with libffi or dyncall, or JIT-compiling a call (i.e. with libtcc or clang or something like that), or building the call manually in handwritten assembly. I don't know if it's possible to work around this, but I'm pretty sure it would need to be handled by V8.
  3. Callbacks are a bit difficult to deal with here. Not impossible though.

Alternatives

Obviously Node.js has been chugging along just fine without FFI in core, so not doing anything like this at all is certainly an option.

On the WG call during the June 2020 collab summit, @mhdawson brought up the idea of enabling node addons to more easily use the fast API calls. This could be done by offering a new C++ API to identify "fast call" versions of exported native functions. This would serve the goal of getting v8-fast-api-calls.h into the native code calling path, but wouldn't give the ease-of-use that we get from more FFI-style libraries, where no native code specific to Node.js needs to be written by the end user. It's probably worth considering this approach whether some kind of FFI ends up being included in core or not.

@bengl
Copy link
Member Author

bengl commented Jan 22, 2021

Just a heads up that:

  1. This is still something I want to pursue, but don't have a whole lot of time for right now.
  2. It should probably be removed from the biweekly agenda, unless someone else wants to tackle it.

Tangentially (although related to the "Alternatives" section above), I've put together a hepler library https://www.npmjs.com/package/fast-native-fn to help folks use v8-fast-api-calls.h when available, and still compile when it isn't.

@sindresorhus
Copy link

Great to see some activity on FFI. I have been hoping for native FFI in Node.js since Node.js 0.10. I have a lot of packages that just spawn a native binary as that was more foolproof than a native addon. I have a lot of ideas for packages, but I put them on hold as they require certain system APIs and I don't want to deal with a native addon.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

4 participants