summaryrefslogtreecommitdiff
path: root/bzipper_macros/src/impls/serialise_enum.rs
blob: 825886c8042cb7ba45642a4eae63e2ea05b77405 (plain) (blame)
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
// Copyright 2024 Gabriel Bjørnager Jensen.
//
// This file is part of bzipper.
//
// bzipper is free software: you can redistribute
// it and/or modify it under the terms of the GNU
// Lesser General Public License as published by
// the Free Software Foundation, either version 3
// of the License, or (at your option) any later
// version.
//
// bzipper is distributed in the hope that it will
// be useful, but WITHOUT ANY WARRANTY; without
// even the implied warranty of MERCHANTABILITY or
// FITNESS FOR A PARTICULAR PURPOSE. See the GNU
// Lesser General Public License for more details.
//
// You should have received a copy of the GNU Less-
// er General Public License along with bzipper. If
// not, see <https://www.gnu.org/licenses/>.

use crate::Capture;

use proc_macro2::{Span, TokenStream};
use quote::quote;
use syn::{DataEnum, Fields, Ident, Token};
use syn::punctuated::Punctuated;

#[must_use]
pub fn serialise_enum(data: &DataEnum) -> TokenStream {
	let mut sizes = Vec::new();

	let mut match_arms = Punctuated::<TokenStream, Token![,]>::new();

	for (index, variant) in data.variants.iter().enumerate() {
		let mut serialised_size = Punctuated::<TokenStream, Token![+]>::new();

		let variant_name = &variant.ident;

		let discriminant = u32::try_from(index)
			.expect("enumeration discriminants must be representable as `u32`");

		// Discriminant size:
		serialised_size.push(quote! { <u32 as ::bzipper::Serialise>::MAX_SERIALISED_SIZE });

		let mut captures = Punctuated::<Capture, Token![,]>::new();

		let mut chain_commands = Punctuated::<TokenStream, Token![;]>::new();
		chain_commands.push(quote! { #discriminant.serialise(stream)? });

		for (index, field) in variant.fields.iter().enumerate() {
			let field_ty = &field.ty;

			let field_name = field.ident
				.as_ref()
				.map_or_else(|| Ident::new(&format!("v{index}"), Span::call_site()), Clone::clone);

			serialised_size.push(quote! { <#field_ty as ::bzipper::Serialise>::MAX_SERIALISED_SIZE });

			captures.push(Capture {
				ref_token: Token![ref](Span::call_site()),
				ident:     field_name.clone(),
			});

			chain_commands.push(quote! { #field_name.serialise(stream)? });
		}

		chain_commands.push_punct(Token![;](Span::call_site()));

		let arm = match variant.fields {
			Fields::Named(  ..) => quote! { Self::#variant_name { #captures } => { #chain_commands } },
			Fields::Unnamed(..) => quote! { Self::#variant_name(#captures)    => { #chain_commands } },
			Fields::Unit        => quote! { Self::#variant_name               => { #chain_commands } },
		};

		sizes.push(serialised_size);
		match_arms.push(arm);
	}

	let mut size_tests = Punctuated::<TokenStream, Token![else]>::new();

	for size in &sizes {
		let mut test = Punctuated::<TokenStream, Token![&&]>::new();

		for other_size in &sizes { test.push(quote! { #size >= #other_size }) }

		size_tests.push(quote! { if #test { #size } });
	}

	size_tests.push(quote! { { core::unreachable!(); } });

	quote! {
		const MAX_SERIALISED_SIZE: usize = const { #size_tests };

		fn serialise(&self, stream: &mut ::bzipper::Sstream) -> ::bzipper::Result<()> {
			match (*self) { #match_arms }

			Ok(())
		}
	}
}