aboutsummaryrefslogtreecommitdiff
path: root/core/encoding
diff options
context:
space:
mode:
authorjockus <joakim.hentula@gmail.com>2020-09-15 11:39:34 +0100
committerjockus <joakim.hentula@gmail.com>2020-09-15 11:39:34 +0100
commit195dbd658dd5518ab10c8f143f477d0e29dc751e (patch)
treeae6c6bc21cf972ef25406f4f2ee2704dbe9d17e1 /core/encoding
parent0cd681e6b76896e671d11007966b118b2b263e39 (diff)
Added option to parse number as integer, disabled by default
Diffstat (limited to 'core/encoding')
-rw-r--r--core/encoding/json/parser.odin9
-rw-r--r--core/encoding/json/tokenizer.odin7
-rw-r--r--core/encoding/json/validator.odin2
3 files changed, 10 insertions, 8 deletions
diff --git a/core/encoding/json/parser.odin b/core/encoding/json/parser.odin
index 04e2bab09..448a0e41f 100644
--- a/core/encoding/json/parser.odin
+++ b/core/encoding/json/parser.odin
@@ -11,11 +11,12 @@ Parser :: struct {
spec: Specification,
allocator: mem.Allocator,
unmarshal_data: any,
+ parse_integers: bool,
}
-make_parser :: proc(data: []byte, spec := Specification.JSON, allocator := context.allocator) -> Parser {
+make_parser :: proc(data: []byte, spec := Specification.JSON, parse_integers := false, allocator := context.allocator) -> Parser {
p: Parser;
- p.tok = make_tokenizer(data, spec);
+ p.tok = make_tokenizer(data, spec, parse_integers);
p.spec = spec;
p.allocator = allocator;
assert(p.allocator.procedure != nil);
@@ -23,9 +24,9 @@ make_parser :: proc(data: []byte, spec := Specification.JSON, allocator := conte
return p;
}
-parse :: proc(data: []byte, spec := Specification.JSON, allocator := context.allocator) -> (Value, Error) {
+parse :: proc(data: []byte, spec := Specification.JSON, parse_integers := false, allocator := context.allocator) -> (Value, Error) {
context.allocator = allocator;
- p := make_parser(data, spec, allocator);
+ p := make_parser(data, spec, parse_integers, allocator);
if p.spec == Specification.JSON5 {
return parse_value(&p);
diff --git a/core/encoding/json/tokenizer.odin b/core/encoding/json/tokenizer.odin
index 9527cc07f..f5fab47f2 100644
--- a/core/encoding/json/tokenizer.odin
+++ b/core/encoding/json/tokenizer.odin
@@ -42,12 +42,13 @@ Tokenizer :: struct {
w: int, // current rune width in bytes
curr_line_offset: int,
spec: Specification,
+ parse_integers: bool,
}
-make_tokenizer :: proc(data: []byte, spec := Specification.JSON) -> Tokenizer {
- t := Tokenizer{pos = {line=1}, data = data, spec = spec};
+make_tokenizer :: proc(data: []byte, spec := Specification.JSON, parse_integers := false) -> Tokenizer {
+ t := Tokenizer{pos = {line=1}, data = data, spec = spec, parse_integers = parse_integers};
next_rune(&t);
if t.r == utf8.RUNE_BOM {
next_rune(&t);
@@ -217,7 +218,7 @@ get_token :: proc(t: ^Tokenizer) -> (token: Token, err: Error) {
fallthrough;
case '0'..'9':
- token.kind = .Integer;
+ token.kind = t.parse_integers ? .Integer : .Float;
if t.spec == .JSON5 { // Hexadecimal Numbers
if curr_rune == '0' && (t.r == 'x' || t.r == 'X') {
next_rune(t);
diff --git a/core/encoding/json/validator.odin b/core/encoding/json/validator.odin
index 31cf038b0..ba8c4097f 100644
--- a/core/encoding/json/validator.odin
+++ b/core/encoding/json/validator.odin
@@ -4,7 +4,7 @@ import "core:mem"
// NOTE(bill): is_valid will not check for duplicate keys
is_valid :: proc(data: []byte, spec := Specification.JSON) -> bool {
- p := make_parser(data, spec, mem.nil_allocator());
+ p := make_parser(data, spec, false, mem.nil_allocator());
if p.spec == Specification.JSON5 {
return validate_value(&p);
}