@@ -104,7 +104,7 @@ const Tokenizer = struct {
104
104
};
105
105
}
106
106
107
- fn next (self : & Tokenizer ) Token {
107
+ fn next (self : * Tokenizer ) Token {
108
108
var result = Token {
109
109
.id = Token .Id .Eof ,
110
110
.start = self .index ,
@@ -196,7 +196,7 @@ const Tokenizer = struct {
196
196
line_end : usize ,
197
197
};
198
198
199
- fn getTokenLocation (self : & Tokenizer , token : & const Token ) Location {
199
+ fn getTokenLocation (self : * Tokenizer , token : * const Token ) Location {
200
200
var loc = Location {
201
201
.line = 0 ,
202
202
.column = 0 ,
@@ -221,7 +221,7 @@ const Tokenizer = struct {
221
221
}
222
222
};
223
223
224
- fn parseError (tokenizer : & Tokenizer , token : & const Token , comptime fmt : []const u8 , args : ... ) error {
224
+ fn parseError (tokenizer : * Tokenizer , token : * const Token , comptime fmt : []const u8 , args : ... ) error {
225
225
const loc = tokenizer .getTokenLocation (token );
226
226
warn ("{}:{}:{}: error: " ++ fmt ++ "\n " , tokenizer .source_file_name , loc .line + 1 , loc .column + 1 , args );
227
227
if (loc .line_start <= loc .line_end ) {
@@ -244,13 +244,13 @@ fn parseError(tokenizer: &Tokenizer, token: &const Token, comptime fmt: []const
244
244
return error .ParseError ;
245
245
}
246
246
247
- fn assertToken (tokenizer : & Tokenizer , token : & const Token , id : Token.Id ) ! void {
247
+ fn assertToken (tokenizer : * Tokenizer , token : * const Token , id : Token.Id ) ! void {
248
248
if (token .id != id ) {
249
249
return parseError (tokenizer , token , "expected {}, found {}" , @tagName (id ), @tagName (token .id ));
250
250
}
251
251
}
252
252
253
- fn eatToken (tokenizer : & Tokenizer , id : Token.Id ) ! Token {
253
+ fn eatToken (tokenizer : * Tokenizer , id : Token.Id ) ! Token {
254
254
const token = tokenizer .next ();
255
255
try assertToken (tokenizer , token , id );
256
256
return token ;
@@ -317,7 +317,7 @@ const Action = enum {
317
317
Close ,
318
318
};
319
319
320
- fn genToc (allocator : & mem.Allocator , tokenizer : & Tokenizer ) ! Toc {
320
+ fn genToc (allocator : * mem.Allocator , tokenizer : * Tokenizer ) ! Toc {
321
321
var urls = std .HashMap ([]const u8 , Token , mem .hash_slice_u8 , mem .eql_slice_u8 ).init (allocator );
322
322
errdefer urls .deinit ();
323
323
@@ -546,7 +546,7 @@ fn genToc(allocator: &mem.Allocator, tokenizer: &Tokenizer) !Toc {
546
546
};
547
547
}
548
548
549
- fn urlize (allocator : & mem.Allocator , input : []const u8 ) ! []u8 {
549
+ fn urlize (allocator : * mem.Allocator , input : []const u8 ) ! []u8 {
550
550
var buf = try std .Buffer .initSize (allocator , 0 );
551
551
defer buf .deinit ();
552
552
@@ -566,7 +566,7 @@ fn urlize(allocator: &mem.Allocator, input: []const u8) ![]u8 {
566
566
return buf .toOwnedSlice ();
567
567
}
568
568
569
- fn escapeHtml (allocator : & mem.Allocator , input : []const u8 ) ! []u8 {
569
+ fn escapeHtml (allocator : * mem.Allocator , input : []const u8 ) ! []u8 {
570
570
var buf = try std .Buffer .initSize (allocator , 0 );
571
571
defer buf .deinit ();
572
572
@@ -608,7 +608,7 @@ test "term color" {
608
608
assert (mem .eql (u8 , result , "A<span class=\" t32\" >green</span>B" ));
609
609
}
610
610
611
- fn termColor (allocator : & mem.Allocator , input : []const u8 ) ! []u8 {
611
+ fn termColor (allocator : * mem.Allocator , input : []const u8 ) ! []u8 {
612
612
var buf = try std .Buffer .initSize (allocator , 0 );
613
613
defer buf .deinit ();
614
614
@@ -688,7 +688,7 @@ fn termColor(allocator: &mem.Allocator, input: []const u8) ![]u8 {
688
688
return buf .toOwnedSlice ();
689
689
}
690
690
691
- fn genHtml (allocator : & mem.Allocator , tokenizer : & Tokenizer , toc : & Toc , out : var , zig_exe : []const u8 ) ! void {
691
+ fn genHtml (allocator : * mem.Allocator , tokenizer : * Tokenizer , toc : * Toc , out : var , zig_exe : []const u8 ) ! void {
692
692
var code_progress_index : usize = 0 ;
693
693
for (toc .nodes ) | node | {
694
694
switch (node ) {
@@ -1036,7 +1036,7 @@ fn genHtml(allocator: &mem.Allocator, tokenizer: &Tokenizer, toc: &Toc, out: var
1036
1036
}
1037
1037
}
1038
1038
1039
- fn exec (allocator : & mem.Allocator , args : []const []const u8 ) ! os.ChildProcess.ExecResult {
1039
+ fn exec (allocator : * mem.Allocator , args : []const []const u8 ) ! os.ChildProcess.ExecResult {
1040
1040
const result = try os .ChildProcess .exec (allocator , args , null , null , max_doc_file_size );
1041
1041
switch (result .term ) {
1042
1042
os .ChildProcess .Term .Exited = > | exit_code | {
0 commit comments