
# Objective - Contributes to #15460 ## Solution - Added the following features: - `std` (default) - `async_executor` (default) - `edge_executor` - `critical-section` - `portable-atomic` - Gated `tracing` in `bevy_utils` to allow compilation on certain platforms - Switched from `tracing` to `log` for simple message logging within `bevy_ecs`. Note that `tracing` supports capturing from `log` so this should be an uncontroversial change. - Fixed imports and added feature gates as required - Made `bevy_tasks` optional within `bevy_ecs`. Turns out it's only needed for parallel operations which are already gated behind `multi_threaded` anyway. ## Testing - Added to `compile-check-no-std` CI command - `cargo check -p bevy_ecs --no-default-features --features edge_executor,critical-section,portable-atomic --target thumbv6m-none-eabi` - `cargo check -p bevy_ecs --no-default-features --features edge_executor,critical-section` - `cargo check -p bevy_ecs --no-default-features` ## Draft Release Notes Bevy's core ECS now supports `no_std` platforms. In prior versions of Bevy, it was not possible to work with embedded or niche platforms due to our reliance on the standard library, `std`. This has blocked a number of novel use-cases for Bevy, such as an embedded database for IoT devices, or for creating games on retro consoles. With this release, `bevy_ecs` no longer requires `std`. To use Bevy on a `no_std` platform, you must disable default features and enable the new `edge_executor` and `critical-section` features. You may also need to enable `portable-atomic` and `critical-section` if your platform does not natively support all atomic types and operations used by Bevy. ```toml [dependencies] bevy_ecs = { version = "0.16", default-features = false, features = [ # Required for platforms with incomplete atomics (e.g., Raspberry Pi Pico) "portable-atomic", "critical-section", # Optional "bevy_reflect", "serialize", "bevy_debug_stepping", "edge_executor" ] } ``` Currently, this has been tested on bare-metal x86 and the Raspberry Pi Pico. If you have trouble using `bevy_ecs` on a particular platform, please reach out either through a GitHub issue or in the `no_std` working group on the Bevy Discord server. Keep an eye out for future `no_std` updates as we continue to improve the parity between `std` and `no_std`. We look forward to seeing what kinds of applications are now possible with Bevy! ## Notes - Creating PR in draft to ensure CI is passing before requesting reviews. - This implementation has no support for multithreading in `no_std`, especially due to `NonSend` being unsound if allowed in multithreading. The reason is we cannot check the `ThreadId` in `no_std`, so we have no mechanism to at-runtime determine if access is sound. --------- Co-authored-by: Alice Cecile <alice.i.cecile@gmail.com> Co-authored-by: Vic <59878206+Victoronz@users.noreply.github.com>
349 lines
13 KiB
Rust
349 lines
13 KiB
Rust
use proc_macro::{TokenStream, TokenTree};
|
|
use proc_macro2::{Span, TokenStream as TokenStream2};
|
|
use quote::{quote, ToTokens};
|
|
use std::collections::HashSet;
|
|
use syn::{
|
|
parenthesized,
|
|
parse::Parse,
|
|
parse_macro_input, parse_quote,
|
|
punctuated::Punctuated,
|
|
spanned::Spanned,
|
|
token::{Comma, Paren},
|
|
DeriveInput, ExprClosure, ExprPath, Ident, LitStr, Path, Result,
|
|
};
|
|
|
|
pub fn derive_event(input: TokenStream) -> TokenStream {
|
|
let mut ast = parse_macro_input!(input as DeriveInput);
|
|
let bevy_ecs_path: Path = crate::bevy_ecs_path();
|
|
|
|
ast.generics
|
|
.make_where_clause()
|
|
.predicates
|
|
.push(parse_quote! { Self: Send + Sync + 'static });
|
|
|
|
let struct_name = &ast.ident;
|
|
let (impl_generics, type_generics, where_clause) = &ast.generics.split_for_impl();
|
|
|
|
TokenStream::from(quote! {
|
|
impl #impl_generics #bevy_ecs_path::event::Event for #struct_name #type_generics #where_clause {
|
|
type Traversal = ();
|
|
const AUTO_PROPAGATE: bool = false;
|
|
}
|
|
|
|
impl #impl_generics #bevy_ecs_path::component::Component for #struct_name #type_generics #where_clause {
|
|
const STORAGE_TYPE: #bevy_ecs_path::component::StorageType = #bevy_ecs_path::component::StorageType::SparseSet;
|
|
type Mutability = #bevy_ecs_path::component::Mutable;
|
|
}
|
|
})
|
|
}
|
|
|
|
pub fn derive_resource(input: TokenStream) -> TokenStream {
|
|
let mut ast = parse_macro_input!(input as DeriveInput);
|
|
let bevy_ecs_path: Path = crate::bevy_ecs_path();
|
|
|
|
ast.generics
|
|
.make_where_clause()
|
|
.predicates
|
|
.push(parse_quote! { Self: Send + Sync + 'static });
|
|
|
|
let struct_name = &ast.ident;
|
|
let (impl_generics, type_generics, where_clause) = &ast.generics.split_for_impl();
|
|
|
|
TokenStream::from(quote! {
|
|
impl #impl_generics #bevy_ecs_path::system::Resource for #struct_name #type_generics #where_clause {
|
|
}
|
|
})
|
|
}
|
|
|
|
pub fn derive_component(input: TokenStream) -> TokenStream {
|
|
let mut ast = parse_macro_input!(input as DeriveInput);
|
|
let bevy_ecs_path: Path = crate::bevy_ecs_path();
|
|
|
|
let attrs = match parse_component_attr(&ast) {
|
|
Ok(attrs) => attrs,
|
|
Err(e) => return e.into_compile_error().into(),
|
|
};
|
|
|
|
let storage = storage_path(&bevy_ecs_path, attrs.storage);
|
|
|
|
let on_add = hook_register_function_call(quote! {on_add}, attrs.on_add);
|
|
let on_insert = hook_register_function_call(quote! {on_insert}, attrs.on_insert);
|
|
let on_replace = hook_register_function_call(quote! {on_replace}, attrs.on_replace);
|
|
let on_remove = hook_register_function_call(quote! {on_remove}, attrs.on_remove);
|
|
|
|
ast.generics
|
|
.make_where_clause()
|
|
.predicates
|
|
.push(parse_quote! { Self: Send + Sync + 'static });
|
|
|
|
let requires = &attrs.requires;
|
|
let mut register_required = Vec::with_capacity(attrs.requires.iter().len());
|
|
let mut register_recursive_requires = Vec::with_capacity(attrs.requires.iter().len());
|
|
if let Some(requires) = requires {
|
|
for require in requires {
|
|
let ident = &require.path;
|
|
register_recursive_requires.push(quote! {
|
|
<#ident as #bevy_ecs_path::component::Component>::register_required_components(
|
|
requiree,
|
|
components,
|
|
storages,
|
|
required_components,
|
|
inheritance_depth + 1,
|
|
recursion_check_stack
|
|
);
|
|
});
|
|
match &require.func {
|
|
Some(RequireFunc::Path(func)) => {
|
|
register_required.push(quote! {
|
|
components.register_required_components_manual::<Self, #ident>(
|
|
storages,
|
|
required_components,
|
|
|| { let x: #ident = #func().into(); x },
|
|
inheritance_depth,
|
|
recursion_check_stack
|
|
);
|
|
});
|
|
}
|
|
Some(RequireFunc::Closure(func)) => {
|
|
register_required.push(quote! {
|
|
components.register_required_components_manual::<Self, #ident>(
|
|
storages,
|
|
required_components,
|
|
|| { let x: #ident = (#func)().into(); x },
|
|
inheritance_depth,
|
|
recursion_check_stack
|
|
);
|
|
});
|
|
}
|
|
None => {
|
|
register_required.push(quote! {
|
|
components.register_required_components_manual::<Self, #ident>(
|
|
storages,
|
|
required_components,
|
|
<#ident as Default>::default,
|
|
inheritance_depth,
|
|
recursion_check_stack
|
|
);
|
|
});
|
|
}
|
|
}
|
|
}
|
|
}
|
|
let struct_name = &ast.ident;
|
|
let (impl_generics, type_generics, where_clause) = &ast.generics.split_for_impl();
|
|
|
|
let mutable_type = attrs
|
|
.immutable
|
|
.then_some(quote! { #bevy_ecs_path::component::Immutable })
|
|
.unwrap_or(quote! { #bevy_ecs_path::component::Mutable });
|
|
|
|
// This puts `register_required` before `register_recursive_requires` to ensure that the constructors of _all_ top
|
|
// level components are initialized first, giving them precedence over recursively defined constructors for the same component type
|
|
TokenStream::from(quote! {
|
|
impl #impl_generics #bevy_ecs_path::component::Component for #struct_name #type_generics #where_clause {
|
|
const STORAGE_TYPE: #bevy_ecs_path::component::StorageType = #storage;
|
|
type Mutability = #mutable_type;
|
|
fn register_required_components(
|
|
requiree: #bevy_ecs_path::component::ComponentId,
|
|
components: &mut #bevy_ecs_path::component::Components,
|
|
storages: &mut #bevy_ecs_path::storage::Storages,
|
|
required_components: &mut #bevy_ecs_path::component::RequiredComponents,
|
|
inheritance_depth: u16,
|
|
recursion_check_stack: &mut #bevy_ecs_path::__macro_exports::Vec<#bevy_ecs_path::component::ComponentId>
|
|
) {
|
|
#bevy_ecs_path::component::enforce_no_required_components_recursion(components, recursion_check_stack);
|
|
let self_id = components.register_component::<Self>(storages);
|
|
recursion_check_stack.push(self_id);
|
|
#(#register_required)*
|
|
#(#register_recursive_requires)*
|
|
recursion_check_stack.pop();
|
|
}
|
|
|
|
#[allow(unused_variables)]
|
|
fn register_component_hooks(hooks: &mut #bevy_ecs_path::component::ComponentHooks) {
|
|
#on_add
|
|
#on_insert
|
|
#on_replace
|
|
#on_remove
|
|
}
|
|
|
|
fn get_component_clone_handler() -> #bevy_ecs_path::component::ComponentCloneHandler {
|
|
use #bevy_ecs_path::component::{ComponentCloneViaClone, ComponentCloneBase};
|
|
(&&&#bevy_ecs_path::component::ComponentCloneSpecializationWrapper::<Self>::default())
|
|
.get_component_clone_handler()
|
|
}
|
|
}
|
|
})
|
|
}
|
|
|
|
pub fn document_required_components(attr: TokenStream, item: TokenStream) -> TokenStream {
|
|
let paths = parse_macro_input!(attr with Punctuated::<Require, Comma>::parse_terminated)
|
|
.iter()
|
|
.map(|r| format!("[`{}`]", r.path.to_token_stream()))
|
|
.collect::<Vec<_>>()
|
|
.join(", ");
|
|
|
|
let bevy_ecs_path = crate::bevy_ecs_path()
|
|
.to_token_stream()
|
|
.to_string()
|
|
.replace(' ', "");
|
|
let required_components_path = bevy_ecs_path + "::component::Component#required-components";
|
|
|
|
// Insert information about required components after any existing doc comments
|
|
let mut out = TokenStream::new();
|
|
let mut end_of_attributes_reached = false;
|
|
for tt in item {
|
|
if !end_of_attributes_reached & matches!(tt, TokenTree::Ident(_)) {
|
|
end_of_attributes_reached = true;
|
|
let doc: TokenStream = format!("#[doc = \"\n\n# Required Components\n{paths} \n\n A component's [required components]({required_components_path}) are inserted whenever it is inserted. Note that this will also insert the required components _of_ the required components, recursively, in depth-first order.\"]").parse().unwrap();
|
|
out.extend(doc);
|
|
}
|
|
out.extend(Some(tt));
|
|
}
|
|
|
|
out
|
|
}
|
|
|
|
pub const COMPONENT: &str = "component";
|
|
pub const STORAGE: &str = "storage";
|
|
pub const REQUIRE: &str = "require";
|
|
|
|
pub const ON_ADD: &str = "on_add";
|
|
pub const ON_INSERT: &str = "on_insert";
|
|
pub const ON_REPLACE: &str = "on_replace";
|
|
pub const ON_REMOVE: &str = "on_remove";
|
|
|
|
pub const IMMUTABLE: &str = "immutable";
|
|
|
|
struct Attrs {
|
|
storage: StorageTy,
|
|
requires: Option<Punctuated<Require, Comma>>,
|
|
on_add: Option<ExprPath>,
|
|
on_insert: Option<ExprPath>,
|
|
on_replace: Option<ExprPath>,
|
|
on_remove: Option<ExprPath>,
|
|
immutable: bool,
|
|
}
|
|
|
|
#[derive(Clone, Copy)]
|
|
enum StorageTy {
|
|
Table,
|
|
SparseSet,
|
|
}
|
|
|
|
struct Require {
|
|
path: Path,
|
|
func: Option<RequireFunc>,
|
|
}
|
|
|
|
enum RequireFunc {
|
|
Path(Path),
|
|
Closure(ExprClosure),
|
|
}
|
|
|
|
// values for `storage` attribute
|
|
const TABLE: &str = "Table";
|
|
const SPARSE_SET: &str = "SparseSet";
|
|
|
|
fn parse_component_attr(ast: &DeriveInput) -> Result<Attrs> {
|
|
let mut attrs = Attrs {
|
|
storage: StorageTy::Table,
|
|
on_add: None,
|
|
on_insert: None,
|
|
on_replace: None,
|
|
on_remove: None,
|
|
requires: None,
|
|
immutable: false,
|
|
};
|
|
|
|
let mut require_paths = HashSet::new();
|
|
for attr in ast.attrs.iter() {
|
|
if attr.path().is_ident(COMPONENT) {
|
|
attr.parse_nested_meta(|nested| {
|
|
if nested.path.is_ident(STORAGE) {
|
|
attrs.storage = match nested.value()?.parse::<LitStr>()?.value() {
|
|
s if s == TABLE => StorageTy::Table,
|
|
s if s == SPARSE_SET => StorageTy::SparseSet,
|
|
s => {
|
|
return Err(nested.error(format!(
|
|
"Invalid storage type `{s}`, expected '{TABLE}' or '{SPARSE_SET}'.",
|
|
)));
|
|
}
|
|
};
|
|
Ok(())
|
|
} else if nested.path.is_ident(ON_ADD) {
|
|
attrs.on_add = Some(nested.value()?.parse::<ExprPath>()?);
|
|
Ok(())
|
|
} else if nested.path.is_ident(ON_INSERT) {
|
|
attrs.on_insert = Some(nested.value()?.parse::<ExprPath>()?);
|
|
Ok(())
|
|
} else if nested.path.is_ident(ON_REPLACE) {
|
|
attrs.on_replace = Some(nested.value()?.parse::<ExprPath>()?);
|
|
Ok(())
|
|
} else if nested.path.is_ident(ON_REMOVE) {
|
|
attrs.on_remove = Some(nested.value()?.parse::<ExprPath>()?);
|
|
Ok(())
|
|
} else if nested.path.is_ident(IMMUTABLE) {
|
|
attrs.immutable = true;
|
|
Ok(())
|
|
} else {
|
|
Err(nested.error("Unsupported attribute"))
|
|
}
|
|
})?;
|
|
} else if attr.path().is_ident(REQUIRE) {
|
|
let punctuated =
|
|
attr.parse_args_with(Punctuated::<Require, Comma>::parse_terminated)?;
|
|
for require in punctuated.iter() {
|
|
if !require_paths.insert(require.path.to_token_stream().to_string()) {
|
|
return Err(syn::Error::new(
|
|
require.path.span(),
|
|
"Duplicate required components are not allowed.",
|
|
));
|
|
}
|
|
}
|
|
if let Some(current) = &mut attrs.requires {
|
|
current.extend(punctuated);
|
|
} else {
|
|
attrs.requires = Some(punctuated);
|
|
}
|
|
}
|
|
}
|
|
|
|
Ok(attrs)
|
|
}
|
|
|
|
impl Parse for Require {
|
|
fn parse(input: syn::parse::ParseStream) -> Result<Self> {
|
|
let path = input.parse::<Path>()?;
|
|
let func = if input.peek(Paren) {
|
|
let content;
|
|
parenthesized!(content in input);
|
|
if let Ok(func) = content.parse::<ExprClosure>() {
|
|
Some(RequireFunc::Closure(func))
|
|
} else {
|
|
let func = content.parse::<Path>()?;
|
|
Some(RequireFunc::Path(func))
|
|
}
|
|
} else {
|
|
None
|
|
};
|
|
Ok(Require { path, func })
|
|
}
|
|
}
|
|
|
|
fn storage_path(bevy_ecs_path: &Path, ty: StorageTy) -> TokenStream2 {
|
|
let storage_type = match ty {
|
|
StorageTy::Table => Ident::new("Table", Span::call_site()),
|
|
StorageTy::SparseSet => Ident::new("SparseSet", Span::call_site()),
|
|
};
|
|
|
|
quote! { #bevy_ecs_path::component::StorageType::#storage_type }
|
|
}
|
|
|
|
fn hook_register_function_call(
|
|
hook: TokenStream2,
|
|
function: Option<ExprPath>,
|
|
) -> Option<TokenStream2> {
|
|
function.map(|meta| quote! { hooks. #hook (#meta); })
|
|
}
|