diff options
| author | Ginger Bill <bill@gingerbill.org> | 2016-10-23 14:46:52 +0100 |
|---|---|---|
| committer | Ginger Bill <bill@gingerbill.org> | 2016-10-23 14:46:52 +0100 |
| commit | bd5d095de4a3a3de574f6004751718ee944417cc (patch) | |
| tree | 0eaceb775794fe21073c2b4159287ae08b99c3e3 /src/tokenizer.cpp | |
| parent | f60dc7b0a7f8bf8122df0fa3b4d12603a9775f87 (diff) | |
Separate ssa_struct_gep and ssa_array_gep procedures
Diffstat (limited to 'src/tokenizer.cpp')
| -rw-r--r-- | src/tokenizer.cpp | 14 |
1 files changed, 7 insertions, 7 deletions
diff --git a/src/tokenizer.cpp b/src/tokenizer.cpp index a7f53c0f2..02003d4af 100644 --- a/src/tokenizer.cpp +++ b/src/tokenizer.cpp @@ -348,15 +348,15 @@ void advance_to_next_rune(Tokenizer *t) { TokenizerInitError init_tokenizer(Tokenizer *t, String fullpath) { PROF_PROC(); - char *c_str = gb_alloc_array(gb_heap_allocator(), char, fullpath.len+1); + char *c_str = gb_alloc_array(heap_allocator(), char, fullpath.len+1); memcpy(c_str, fullpath.text, fullpath.len); c_str[fullpath.len] = '\0'; - defer (gb_free(gb_heap_allocator(), c_str)); + defer (gb_free(heap_allocator(), c_str)); // TODO(bill): Memory map rather than copy contents - gbFileContents fc = gb_file_read_contents(gb_heap_allocator(), true, c_str); + gbFileContents fc = gb_file_read_contents(heap_allocator(), true, c_str); gb_zero_item(t); if (fc.data != NULL) { t->start = cast(u8 *)fc.data; @@ -372,7 +372,7 @@ TokenizerInitError init_tokenizer(Tokenizer *t, String fullpath) { advance_to_next_rune(t); // Ignore BOM at file beginning } - array_init(&t->allocated_strings, gb_heap_allocator()); + array_init(&t->allocated_strings, heap_allocator()); return TokenizerInit_None; } @@ -399,10 +399,10 @@ TokenizerInitError init_tokenizer(Tokenizer *t, String fullpath) { gb_inline void destroy_tokenizer(Tokenizer *t) { if (t->start != NULL) { - gb_free(gb_heap_allocator(), t->start); + gb_free(heap_allocator(), t->start); } for_array(i, t->allocated_strings) { - gb_free(gb_heap_allocator(), t->allocated_strings[i].text); + gb_free(heap_allocator(), t->allocated_strings[i].text); } array_free(&t->allocated_strings); } @@ -699,7 +699,7 @@ Token tokenizer_get_token(Tokenizer *t) { } } token.string.len = t->curr - token.string.text; - i32 success = unquote_string(gb_heap_allocator(), &token.string); + i32 success = unquote_string(heap_allocator(), &token.string); if (success > 0) { if (success == 2) { array_add(&t->allocated_strings, token.string); |