summaryrefslogtreecommitdiff
path: root/bzipper_macros/src/impls/serialise_enum.rs
diff options
context:
space:
mode:
Diffstat (limited to 'bzipper_macros/src/impls/serialise_enum.rs')
-rw-r--r--bzipper_macros/src/impls/serialise_enum.rs101
1 files changed, 0 insertions, 101 deletions
diff --git a/bzipper_macros/src/impls/serialise_enum.rs b/bzipper_macros/src/impls/serialise_enum.rs
deleted file mode 100644
index 825886c..0000000
--- a/bzipper_macros/src/impls/serialise_enum.rs
+++ /dev/null
@@ -1,101 +0,0 @@
-// Copyright 2024 Gabriel Bjørnager Jensen.
-//
-// This file is part of bzipper.
-//
-// bzipper is free software: you can redistribute
-// it and/or modify it under the terms of the GNU
-// Lesser General Public License as published by
-// the Free Software Foundation, either version 3
-// of the License, or (at your option) any later
-// version.
-//
-// bzipper is distributed in the hope that it will
-// be useful, but WITHOUT ANY WARRANTY; without
-// even the implied warranty of MERCHANTABILITY or
-// FITNESS FOR A PARTICULAR PURPOSE. See the GNU
-// Lesser General Public License for more details.
-//
-// You should have received a copy of the GNU Less-
-// er General Public License along with bzipper. If
-// not, see <https://www.gnu.org/licenses/>.
-
-use crate::Capture;
-
-use proc_macro2::{Span, TokenStream};
-use quote::quote;
-use syn::{DataEnum, Fields, Ident, Token};
-use syn::punctuated::Punctuated;
-
-#[must_use]
-pub fn serialise_enum(data: &DataEnum) -> TokenStream {
- let mut sizes = Vec::new();
-
- let mut match_arms = Punctuated::<TokenStream, Token![,]>::new();
-
- for (index, variant) in data.variants.iter().enumerate() {
- let mut serialised_size = Punctuated::<TokenStream, Token![+]>::new();
-
- let variant_name = &variant.ident;
-
- let discriminant = u32::try_from(index)
- .expect("enumeration discriminants must be representable as `u32`");
-
- // Discriminant size:
- serialised_size.push(quote! { <u32 as ::bzipper::Serialise>::MAX_SERIALISED_SIZE });
-
- let mut captures = Punctuated::<Capture, Token![,]>::new();
-
- let mut chain_commands = Punctuated::<TokenStream, Token![;]>::new();
- chain_commands.push(quote! { #discriminant.serialise(stream)? });
-
- for (index, field) in variant.fields.iter().enumerate() {
- let field_ty = &field.ty;
-
- let field_name = field.ident
- .as_ref()
- .map_or_else(|| Ident::new(&format!("v{index}"), Span::call_site()), Clone::clone);
-
- serialised_size.push(quote! { <#field_ty as ::bzipper::Serialise>::MAX_SERIALISED_SIZE });
-
- captures.push(Capture {
- ref_token: Token![ref](Span::call_site()),
- ident: field_name.clone(),
- });
-
- chain_commands.push(quote! { #field_name.serialise(stream)? });
- }
-
- chain_commands.push_punct(Token![;](Span::call_site()));
-
- let arm = match variant.fields {
- Fields::Named( ..) => quote! { Self::#variant_name { #captures } => { #chain_commands } },
- Fields::Unnamed(..) => quote! { Self::#variant_name(#captures) => { #chain_commands } },
- Fields::Unit => quote! { Self::#variant_name => { #chain_commands } },
- };
-
- sizes.push(serialised_size);
- match_arms.push(arm);
- }
-
- let mut size_tests = Punctuated::<TokenStream, Token![else]>::new();
-
- for size in &sizes {
- let mut test = Punctuated::<TokenStream, Token![&&]>::new();
-
- for other_size in &sizes { test.push(quote! { #size >= #other_size }) }
-
- size_tests.push(quote! { if #test { #size } });
- }
-
- size_tests.push(quote! { { core::unreachable!(); } });
-
- quote! {
- const MAX_SERIALISED_SIZE: usize = const { #size_tests };
-
- fn serialise(&self, stream: &mut ::bzipper::Sstream) -> ::bzipper::Result<()> {
- match (*self) { #match_arms }
-
- Ok(())
- }
- }
-}