mirror of
https://github.com/ziglang/zig.git
synced 2025-12-16 19:23:08 +00:00
Merge pull request #14202 from ziglang/std.http
std.http.Client: support HTTP redirects
This commit is contained in:
commit
b3e495a38a
@ -1,5 +1,10 @@
|
|||||||
pub const Client = @import("http/Client.zig");
|
pub const Client = @import("http/Client.zig");
|
||||||
|
|
||||||
|
pub const Version = enum {
|
||||||
|
@"HTTP/1.0",
|
||||||
|
@"HTTP/1.1",
|
||||||
|
};
|
||||||
|
|
||||||
/// https://developer.mozilla.org/en-US/docs/Web/HTTP/Methods
|
/// https://developer.mozilla.org/en-US/docs/Web/HTTP/Methods
|
||||||
/// https://datatracker.ietf.org/doc/html/rfc7231#section-4 Initial definiton
|
/// https://datatracker.ietf.org/doc/html/rfc7231#section-4 Initial definiton
|
||||||
/// https://datatracker.ietf.org/doc/html/rfc5789#section-2 PATCH
|
/// https://datatracker.ietf.org/doc/html/rfc5789#section-2 PATCH
|
||||||
@ -220,14 +225,13 @@ pub const Status = enum(u10) {
|
|||||||
server_error,
|
server_error,
|
||||||
};
|
};
|
||||||
|
|
||||||
pub fn class(self: Status) ?Class {
|
pub fn class(self: Status) Class {
|
||||||
return switch (@enumToInt(self)) {
|
return switch (@enumToInt(self)) {
|
||||||
100...199 => .informational,
|
100...199 => .informational,
|
||||||
200...299 => .success,
|
200...299 => .success,
|
||||||
300...399 => .redirect,
|
300...399 => .redirect,
|
||||||
400...499 => .client_error,
|
400...499 => .client_error,
|
||||||
500...599 => .server_error,
|
else => .server_error,
|
||||||
else => null,
|
|
||||||
};
|
};
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -242,60 +246,10 @@ pub const Status = enum(u10) {
|
|||||||
}
|
}
|
||||||
};
|
};
|
||||||
|
|
||||||
pub const Headers = struct {
|
|
||||||
state: State = .start,
|
|
||||||
invalid_index: u32 = undefined,
|
|
||||||
|
|
||||||
pub const State = enum { invalid, start, line, nl_r, nl_n, nl2_r, finished };
|
|
||||||
|
|
||||||
/// Returns how many bytes are processed into headers. Always less than or
|
|
||||||
/// equal to bytes.len. If the amount returned is less than bytes.len, it
|
|
||||||
/// means the headers ended and the first byte after the double \r\n\r\n is
|
|
||||||
/// located at `bytes[result]`.
|
|
||||||
pub fn feed(h: *Headers, bytes: []const u8) usize {
|
|
||||||
for (bytes) |b, i| {
|
|
||||||
switch (h.state) {
|
|
||||||
.start => switch (b) {
|
|
||||||
'\r' => h.state = .nl_r,
|
|
||||||
'\n' => return invalid(h, i),
|
|
||||||
else => {},
|
|
||||||
},
|
|
||||||
.nl_r => switch (b) {
|
|
||||||
'\n' => h.state = .nl_n,
|
|
||||||
else => return invalid(h, i),
|
|
||||||
},
|
|
||||||
.nl_n => switch (b) {
|
|
||||||
'\r' => h.state = .nl2_r,
|
|
||||||
else => h.state = .line,
|
|
||||||
},
|
|
||||||
.nl2_r => switch (b) {
|
|
||||||
'\n' => h.state = .finished,
|
|
||||||
else => return invalid(h, i),
|
|
||||||
},
|
|
||||||
.line => switch (b) {
|
|
||||||
'\r' => h.state = .nl_r,
|
|
||||||
'\n' => return invalid(h, i),
|
|
||||||
else => {},
|
|
||||||
},
|
|
||||||
.invalid => return i,
|
|
||||||
.finished => return i,
|
|
||||||
}
|
|
||||||
}
|
|
||||||
return bytes.len;
|
|
||||||
}
|
|
||||||
|
|
||||||
fn invalid(h: *Headers, i: usize) usize {
|
|
||||||
h.invalid_index = @intCast(u32, i);
|
|
||||||
h.state = .invalid;
|
|
||||||
return i;
|
|
||||||
}
|
|
||||||
};
|
|
||||||
|
|
||||||
const std = @import("std.zig");
|
const std = @import("std.zig");
|
||||||
|
|
||||||
test {
|
test {
|
||||||
_ = Client;
|
_ = Client;
|
||||||
_ = Method;
|
_ = Method;
|
||||||
_ = Status;
|
_ = Status;
|
||||||
_ = Headers;
|
|
||||||
}
|
}
|
||||||
|
|||||||
@ -1,64 +1,449 @@
|
|||||||
//! This API is a barely-touched, barely-functional http client, just the
|
//! This API is a barely-touched, barely-functional http client, just the
|
||||||
//! absolute minimum thing I needed in order to test `std.crypto.tls`. Bear
|
//! absolute minimum thing I needed in order to test `std.crypto.tls`. Bear
|
||||||
//! with me and I promise the API will become useful and streamlined.
|
//! with me and I promise the API will become useful and streamlined.
|
||||||
|
//!
|
||||||
|
//! TODO: send connection: keep-alive and LRU cache a configurable number of
|
||||||
|
//! open connections to skip DNS and TLS handshake for subsequent requests.
|
||||||
|
|
||||||
const std = @import("../std.zig");
|
const std = @import("../std.zig");
|
||||||
|
const mem = std.mem;
|
||||||
const assert = std.debug.assert;
|
const assert = std.debug.assert;
|
||||||
const http = std.http;
|
const http = std.http;
|
||||||
const net = std.net;
|
const net = std.net;
|
||||||
const Client = @This();
|
const Client = @This();
|
||||||
const Url = std.Url;
|
const Url = std.Url;
|
||||||
|
const Allocator = std.mem.Allocator;
|
||||||
|
const testing = std.testing;
|
||||||
|
|
||||||
allocator: std.mem.Allocator,
|
/// Used for tcpConnectToHost and storing HTTP headers when an externally
|
||||||
headers: std.ArrayListUnmanaged(u8) = .{},
|
/// managed buffer is not provided.
|
||||||
active_requests: usize = 0,
|
allocator: Allocator,
|
||||||
ca_bundle: std.crypto.Certificate.Bundle = .{},
|
ca_bundle: std.crypto.Certificate.Bundle = .{},
|
||||||
|
|
||||||
|
pub const Connection = struct {
|
||||||
|
stream: net.Stream,
|
||||||
|
/// undefined unless protocol is tls.
|
||||||
|
tls_client: std.crypto.tls.Client,
|
||||||
|
protocol: Protocol,
|
||||||
|
|
||||||
|
pub const Protocol = enum { plain, tls };
|
||||||
|
|
||||||
|
pub fn read(conn: *Connection, buffer: []u8) !usize {
|
||||||
|
switch (conn.protocol) {
|
||||||
|
.plain => return conn.stream.read(buffer),
|
||||||
|
.tls => return conn.tls_client.read(conn.stream, buffer),
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
pub fn readAtLeast(conn: *Connection, buffer: []u8, len: usize) !usize {
|
||||||
|
switch (conn.protocol) {
|
||||||
|
.plain => return conn.stream.readAtLeast(buffer, len),
|
||||||
|
.tls => return conn.tls_client.readAtLeast(conn.stream, buffer, len),
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
pub fn writeAll(conn: *Connection, buffer: []const u8) !void {
|
||||||
|
switch (conn.protocol) {
|
||||||
|
.plain => return conn.stream.writeAll(buffer),
|
||||||
|
.tls => return conn.tls_client.writeAll(conn.stream, buffer),
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
pub fn write(conn: *Connection, buffer: []const u8) !usize {
|
||||||
|
switch (conn.protocol) {
|
||||||
|
.plain => return conn.stream.write(buffer),
|
||||||
|
.tls => return conn.tls_client.write(conn.stream, buffer),
|
||||||
|
}
|
||||||
|
}
|
||||||
|
};
|
||||||
|
|
||||||
/// TODO: emit error.UnexpectedEndOfStream or something like that when the read
|
/// TODO: emit error.UnexpectedEndOfStream or something like that when the read
|
||||||
/// data does not match the content length. This is necessary since HTTPS disables
|
/// data does not match the content length. This is necessary since HTTPS disables
|
||||||
/// close_notify protection on underlying TLS streams.
|
/// close_notify protection on underlying TLS streams.
|
||||||
pub const Request = struct {
|
pub const Request = struct {
|
||||||
client: *Client,
|
client: *Client,
|
||||||
stream: net.Stream,
|
connection: Connection,
|
||||||
headers: std.ArrayListUnmanaged(u8) = .{},
|
redirects_left: u32,
|
||||||
tls_client: std.crypto.tls.Client,
|
response: Response,
|
||||||
protocol: Protocol,
|
/// These are stored in Request so that they are available when following
|
||||||
response_headers: http.Headers = .{},
|
/// redirects.
|
||||||
|
headers: Headers,
|
||||||
|
|
||||||
pub const Protocol = enum { http, https };
|
pub const Response = struct {
|
||||||
|
headers: Response.Headers,
|
||||||
|
state: State,
|
||||||
|
header_bytes_owned: bool,
|
||||||
|
/// This could either be a fixed buffer provided by the API user or it
|
||||||
|
/// could be our own array list.
|
||||||
|
header_bytes: std.ArrayListUnmanaged(u8),
|
||||||
|
max_header_bytes: usize,
|
||||||
|
|
||||||
pub const Options = struct {
|
pub const Headers = struct {
|
||||||
|
location: ?[]const u8 = null,
|
||||||
|
status: http.Status,
|
||||||
|
version: http.Version,
|
||||||
|
content_length: ?u64 = null,
|
||||||
|
|
||||||
|
pub fn parse(bytes: []const u8) !Response.Headers {
|
||||||
|
var it = mem.split(u8, bytes[0 .. bytes.len - 4], "\r\n");
|
||||||
|
|
||||||
|
const first_line = it.first();
|
||||||
|
if (first_line.len < 12)
|
||||||
|
return error.ShortHttpStatusLine;
|
||||||
|
|
||||||
|
const version: http.Version = switch (int64(first_line[0..8])) {
|
||||||
|
int64("HTTP/1.0") => .@"HTTP/1.0",
|
||||||
|
int64("HTTP/1.1") => .@"HTTP/1.1",
|
||||||
|
else => return error.BadHttpVersion,
|
||||||
|
};
|
||||||
|
if (first_line[8] != ' ') return error.HttpHeadersInvalid;
|
||||||
|
const status = @intToEnum(http.Status, parseInt3(first_line[9..12].*));
|
||||||
|
|
||||||
|
var headers: Response.Headers = .{
|
||||||
|
.version = version,
|
||||||
|
.status = status,
|
||||||
|
};
|
||||||
|
|
||||||
|
while (it.next()) |line| {
|
||||||
|
if (line.len == 0) return error.HttpHeadersInvalid;
|
||||||
|
switch (line[0]) {
|
||||||
|
' ', '\t' => return error.HttpHeaderContinuationsUnsupported,
|
||||||
|
else => {},
|
||||||
|
}
|
||||||
|
var line_it = mem.split(u8, line, ": ");
|
||||||
|
const header_name = line_it.first();
|
||||||
|
const header_value = line_it.rest();
|
||||||
|
if (std.ascii.eqlIgnoreCase(header_name, "location")) {
|
||||||
|
if (headers.location != null) return error.HttpHeadersInvalid;
|
||||||
|
headers.location = header_value;
|
||||||
|
} else if (std.ascii.eqlIgnoreCase(header_name, "content-length")) {
|
||||||
|
if (headers.content_length != null) return error.HttpHeadersInvalid;
|
||||||
|
headers.content_length = try std.fmt.parseInt(u64, header_value, 10);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
return headers;
|
||||||
|
}
|
||||||
|
|
||||||
|
test "parse headers" {
|
||||||
|
const example =
|
||||||
|
"HTTP/1.1 301 Moved Permanently\r\n" ++
|
||||||
|
"Location: https://www.example.com/\r\n" ++
|
||||||
|
"Content-Type: text/html; charset=UTF-8\r\n" ++
|
||||||
|
"Content-Length: 220\r\n\r\n";
|
||||||
|
const parsed = try Response.Headers.parse(example);
|
||||||
|
try testing.expectEqual(http.Version.@"HTTP/1.1", parsed.version);
|
||||||
|
try testing.expectEqual(http.Status.moved_permanently, parsed.status);
|
||||||
|
try testing.expectEqualStrings("https://www.example.com/", parsed.location orelse
|
||||||
|
return error.TestFailed);
|
||||||
|
try testing.expectEqual(@as(?u64, 220), parsed.content_length);
|
||||||
|
}
|
||||||
|
|
||||||
|
test "header continuation" {
|
||||||
|
const example =
|
||||||
|
"HTTP/1.0 200 OK\r\n" ++
|
||||||
|
"Content-Type: text/html;\r\n charset=UTF-8\r\n" ++
|
||||||
|
"Content-Length: 220\r\n\r\n";
|
||||||
|
try testing.expectError(
|
||||||
|
error.HttpHeaderContinuationsUnsupported,
|
||||||
|
Response.Headers.parse(example),
|
||||||
|
);
|
||||||
|
}
|
||||||
|
|
||||||
|
test "extra content length" {
|
||||||
|
const example =
|
||||||
|
"HTTP/1.0 200 OK\r\n" ++
|
||||||
|
"Content-Length: 220\r\n" ++
|
||||||
|
"Content-Type: text/html; charset=UTF-8\r\n" ++
|
||||||
|
"content-length: 220\r\n\r\n";
|
||||||
|
try testing.expectError(
|
||||||
|
error.HttpHeadersInvalid,
|
||||||
|
Response.Headers.parse(example),
|
||||||
|
);
|
||||||
|
}
|
||||||
|
};
|
||||||
|
|
||||||
|
pub const State = enum {
|
||||||
|
invalid,
|
||||||
|
finished,
|
||||||
|
start,
|
||||||
|
seen_r,
|
||||||
|
seen_rn,
|
||||||
|
seen_rnr,
|
||||||
|
};
|
||||||
|
|
||||||
|
pub fn initDynamic(max: usize) Response {
|
||||||
|
return .{
|
||||||
|
.state = .start,
|
||||||
|
.headers = undefined,
|
||||||
|
.header_bytes = .{},
|
||||||
|
.max_header_bytes = max,
|
||||||
|
.header_bytes_owned = true,
|
||||||
|
};
|
||||||
|
}
|
||||||
|
|
||||||
|
pub fn initStatic(buf: []u8) Response {
|
||||||
|
return .{
|
||||||
|
.state = .start,
|
||||||
|
.headers = undefined,
|
||||||
|
.header_bytes = .{ .items = buf[0..0], .capacity = buf.len },
|
||||||
|
.max_header_bytes = buf.len,
|
||||||
|
.header_bytes_owned = false,
|
||||||
|
};
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Returns how many bytes are part of HTTP headers. Always less than or
|
||||||
|
/// equal to bytes.len. If the amount returned is less than bytes.len, it
|
||||||
|
/// means the headers ended and the first byte after the double \r\n\r\n is
|
||||||
|
/// located at `bytes[result]`.
|
||||||
|
pub fn findHeadersEnd(r: *Response, bytes: []const u8) usize {
|
||||||
|
var index: usize = 0;
|
||||||
|
|
||||||
|
// TODO: https://github.com/ziglang/zig/issues/8220
|
||||||
|
state: while (true) {
|
||||||
|
switch (r.state) {
|
||||||
|
.invalid => unreachable,
|
||||||
|
.finished => unreachable,
|
||||||
|
.start => while (true) {
|
||||||
|
switch (bytes.len - index) {
|
||||||
|
0 => return index,
|
||||||
|
1 => {
|
||||||
|
if (bytes[index] == '\r')
|
||||||
|
r.state = .seen_r;
|
||||||
|
return index + 1;
|
||||||
|
},
|
||||||
|
2 => {
|
||||||
|
if (int16(bytes[index..][0..2]) == int16("\r\n")) {
|
||||||
|
r.state = .seen_rn;
|
||||||
|
} else if (bytes[index + 1] == '\r') {
|
||||||
|
r.state = .seen_r;
|
||||||
|
}
|
||||||
|
return index + 2;
|
||||||
|
},
|
||||||
|
3 => {
|
||||||
|
if (int16(bytes[index..][0..2]) == int16("\r\n") and
|
||||||
|
bytes[index + 2] == '\r')
|
||||||
|
{
|
||||||
|
r.state = .seen_rnr;
|
||||||
|
} else if (int16(bytes[index + 1 ..][0..2]) == int16("\r\n")) {
|
||||||
|
r.state = .seen_rn;
|
||||||
|
} else if (bytes[index + 2] == '\r') {
|
||||||
|
r.state = .seen_r;
|
||||||
|
}
|
||||||
|
return index + 3;
|
||||||
|
},
|
||||||
|
4...15 => {
|
||||||
|
if (int32(bytes[index..][0..4]) == int32("\r\n\r\n")) {
|
||||||
|
r.state = .finished;
|
||||||
|
return index + 4;
|
||||||
|
} else if (int16(bytes[index + 1 ..][0..2]) == int16("\r\n") and
|
||||||
|
bytes[index + 3] == '\r')
|
||||||
|
{
|
||||||
|
r.state = .seen_rnr;
|
||||||
|
index += 4;
|
||||||
|
continue :state;
|
||||||
|
} else if (int16(bytes[index + 2 ..][0..2]) == int16("\r\n")) {
|
||||||
|
r.state = .seen_rn;
|
||||||
|
index += 4;
|
||||||
|
continue :state;
|
||||||
|
} else if (bytes[index + 3] == '\r') {
|
||||||
|
r.state = .seen_r;
|
||||||
|
index += 4;
|
||||||
|
continue :state;
|
||||||
|
}
|
||||||
|
index += 4;
|
||||||
|
continue;
|
||||||
|
},
|
||||||
|
else => {
|
||||||
|
const chunk = bytes[index..][0..16];
|
||||||
|
const v: @Vector(16, u8) = chunk.*;
|
||||||
|
const matches_r = v == @splat(16, @as(u8, '\r'));
|
||||||
|
const iota = std.simd.iota(u8, 16);
|
||||||
|
const default = @splat(16, @as(u8, 16));
|
||||||
|
const sub_index = @reduce(.Min, @select(u8, matches_r, iota, default));
|
||||||
|
switch (sub_index) {
|
||||||
|
0...12 => {
|
||||||
|
index += sub_index + 4;
|
||||||
|
if (int32(chunk[sub_index..][0..4]) == int32("\r\n\r\n")) {
|
||||||
|
r.state = .finished;
|
||||||
|
return index;
|
||||||
|
}
|
||||||
|
continue;
|
||||||
|
},
|
||||||
|
13 => {
|
||||||
|
index += 16;
|
||||||
|
if (int16(chunk[14..][0..2]) == int16("\n\r")) {
|
||||||
|
r.state = .seen_rnr;
|
||||||
|
continue :state;
|
||||||
|
}
|
||||||
|
continue;
|
||||||
|
},
|
||||||
|
14 => {
|
||||||
|
index += 16;
|
||||||
|
if (chunk[15] == '\n') {
|
||||||
|
r.state = .seen_rn;
|
||||||
|
continue :state;
|
||||||
|
}
|
||||||
|
continue;
|
||||||
|
},
|
||||||
|
15 => {
|
||||||
|
r.state = .seen_r;
|
||||||
|
index += 16;
|
||||||
|
continue :state;
|
||||||
|
},
|
||||||
|
16 => {
|
||||||
|
index += 16;
|
||||||
|
continue;
|
||||||
|
},
|
||||||
|
else => unreachable,
|
||||||
|
}
|
||||||
|
},
|
||||||
|
}
|
||||||
|
},
|
||||||
|
|
||||||
|
.seen_r => switch (bytes.len - index) {
|
||||||
|
0 => return index,
|
||||||
|
1 => {
|
||||||
|
switch (bytes[index]) {
|
||||||
|
'\n' => r.state = .seen_rn,
|
||||||
|
'\r' => r.state = .seen_r,
|
||||||
|
else => r.state = .start,
|
||||||
|
}
|
||||||
|
return index + 1;
|
||||||
|
},
|
||||||
|
2 => {
|
||||||
|
if (int16(bytes[index..][0..2]) == int16("\n\r")) {
|
||||||
|
r.state = .seen_rnr;
|
||||||
|
return index + 2;
|
||||||
|
}
|
||||||
|
r.state = .start;
|
||||||
|
return index + 2;
|
||||||
|
},
|
||||||
|
else => {
|
||||||
|
if (int16(bytes[index..][0..2]) == int16("\n\r") and
|
||||||
|
bytes[index + 2] == '\n')
|
||||||
|
{
|
||||||
|
r.state = .finished;
|
||||||
|
return index + 3;
|
||||||
|
}
|
||||||
|
index += 3;
|
||||||
|
r.state = .start;
|
||||||
|
continue :state;
|
||||||
|
},
|
||||||
|
},
|
||||||
|
.seen_rn => switch (bytes.len - index) {
|
||||||
|
0 => return index,
|
||||||
|
1 => {
|
||||||
|
switch (bytes[index]) {
|
||||||
|
'\r' => r.state = .seen_rnr,
|
||||||
|
else => r.state = .start,
|
||||||
|
}
|
||||||
|
return index + 1;
|
||||||
|
},
|
||||||
|
else => {
|
||||||
|
if (int16(bytes[index..][0..2]) == int16("\r\n")) {
|
||||||
|
r.state = .finished;
|
||||||
|
return index + 2;
|
||||||
|
}
|
||||||
|
index += 2;
|
||||||
|
r.state = .start;
|
||||||
|
continue :state;
|
||||||
|
},
|
||||||
|
},
|
||||||
|
.seen_rnr => switch (bytes.len - index) {
|
||||||
|
0 => return index,
|
||||||
|
else => {
|
||||||
|
if (bytes[index] == '\n') {
|
||||||
|
r.state = .finished;
|
||||||
|
return index + 1;
|
||||||
|
}
|
||||||
|
index += 1;
|
||||||
|
r.state = .start;
|
||||||
|
continue :state;
|
||||||
|
},
|
||||||
|
},
|
||||||
|
}
|
||||||
|
|
||||||
|
return index;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
fn parseInt3(nnn: @Vector(3, u8)) u10 {
|
||||||
|
const zero: @Vector(3, u8) = .{ '0', '0', '0' };
|
||||||
|
const mmm: @Vector(3, u10) = .{ 100, 10, 1 };
|
||||||
|
return @reduce(.Add, @as(@Vector(3, u10), nnn -% zero) *% mmm);
|
||||||
|
}
|
||||||
|
|
||||||
|
test parseInt3 {
|
||||||
|
const expectEqual = std.testing.expectEqual;
|
||||||
|
try expectEqual(@as(u10, 0), parseInt3("000".*));
|
||||||
|
try expectEqual(@as(u10, 418), parseInt3("418".*));
|
||||||
|
try expectEqual(@as(u10, 999), parseInt3("999".*));
|
||||||
|
}
|
||||||
|
|
||||||
|
inline fn int16(array: *const [2]u8) u16 {
|
||||||
|
return @bitCast(u16, array.*);
|
||||||
|
}
|
||||||
|
|
||||||
|
inline fn int32(array: *const [4]u8) u32 {
|
||||||
|
return @bitCast(u32, array.*);
|
||||||
|
}
|
||||||
|
|
||||||
|
inline fn int64(array: *const [8]u8) u64 {
|
||||||
|
return @bitCast(u64, array.*);
|
||||||
|
}
|
||||||
|
|
||||||
|
test "find headers end basic" {
|
||||||
|
var buffer: [1]u8 = undefined;
|
||||||
|
var r = Response.initStatic(&buffer);
|
||||||
|
try testing.expectEqual(@as(usize, 10), r.findHeadersEnd("HTTP/1.1 4"));
|
||||||
|
try testing.expectEqual(@as(usize, 2), r.findHeadersEnd("18"));
|
||||||
|
try testing.expectEqual(@as(usize, 8), r.findHeadersEnd(" lol\r\n\r\nblah blah"));
|
||||||
|
}
|
||||||
|
|
||||||
|
test "find headers end vectorized" {
|
||||||
|
var buffer: [1]u8 = undefined;
|
||||||
|
var r = Response.initStatic(&buffer);
|
||||||
|
const example =
|
||||||
|
"HTTP/1.1 301 Moved Permanently\r\n" ++
|
||||||
|
"Location: https://www.example.com/\r\n" ++
|
||||||
|
"Content-Type: text/html; charset=UTF-8\r\n" ++
|
||||||
|
"Content-Length: 220\r\n" ++
|
||||||
|
"\r\ncontent";
|
||||||
|
try testing.expectEqual(@as(usize, 131), r.findHeadersEnd(example));
|
||||||
|
}
|
||||||
|
};
|
||||||
|
|
||||||
|
pub const Headers = struct {
|
||||||
method: http.Method = .GET,
|
method: http.Method = .GET,
|
||||||
};
|
};
|
||||||
|
|
||||||
|
pub const Options = struct {
|
||||||
|
max_redirects: u32 = 3,
|
||||||
|
header_strategy: HeaderStrategy = .{ .dynamic = 16 * 1024 },
|
||||||
|
|
||||||
|
pub const HeaderStrategy = union(enum) {
|
||||||
|
/// In this case, the client's Allocator will be used to store the
|
||||||
|
/// entire HTTP header. This value is the maximum total size of
|
||||||
|
/// HTTP headers allowed, otherwise
|
||||||
|
/// error.HttpHeadersExceededSizeLimit is returned from read().
|
||||||
|
dynamic: usize,
|
||||||
|
/// This is used to store the entire HTTP header. If the HTTP
|
||||||
|
/// header is too big to fit, `error.HttpHeadersExceededSizeLimit`
|
||||||
|
/// is returned from read(). When this is used, `error.OutOfMemory`
|
||||||
|
/// cannot be returned from `read()`.
|
||||||
|
static: []u8,
|
||||||
|
};
|
||||||
|
};
|
||||||
|
|
||||||
|
/// May be skipped if header strategy is buffer.
|
||||||
pub fn deinit(req: *Request) void {
|
pub fn deinit(req: *Request) void {
|
||||||
req.client.active_requests -= 1;
|
if (req.response.header_bytes_owned) {
|
||||||
req.headers.deinit(req.client.allocator);
|
req.response.header_bytes.deinit(req.client.allocator);
|
||||||
|
}
|
||||||
req.* = undefined;
|
req.* = undefined;
|
||||||
}
|
}
|
||||||
|
|
||||||
pub fn addHeader(req: *Request, name: []const u8, value: []const u8) !void {
|
|
||||||
const gpa = req.client.allocator;
|
|
||||||
// Ensure an extra +2 for the \r\n in end()
|
|
||||||
try req.headers.ensureUnusedCapacity(gpa, name.len + value.len + 6);
|
|
||||||
req.headers.appendSliceAssumeCapacity(name);
|
|
||||||
req.headers.appendSliceAssumeCapacity(": ");
|
|
||||||
req.headers.appendSliceAssumeCapacity(value);
|
|
||||||
req.headers.appendSliceAssumeCapacity("\r\n");
|
|
||||||
}
|
|
||||||
|
|
||||||
pub fn end(req: *Request) !void {
|
|
||||||
req.headers.appendSliceAssumeCapacity("\r\n");
|
|
||||||
switch (req.protocol) {
|
|
||||||
.http => {
|
|
||||||
try req.stream.writeAll(req.headers.items);
|
|
||||||
},
|
|
||||||
.https => {
|
|
||||||
try req.tls_client.writeAll(req.stream, req.headers.items);
|
|
||||||
},
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
pub fn readAll(req: *Request, buffer: []u8) !usize {
|
pub fn readAll(req: *Request, buffer: []u8) !usize {
|
||||||
return readAtLeast(req, buffer, buffer.len);
|
return readAtLeast(req, buffer, buffer.len);
|
||||||
}
|
}
|
||||||
@ -71,7 +456,7 @@ pub const Request = struct {
|
|||||||
assert(len <= buffer.len);
|
assert(len <= buffer.len);
|
||||||
var index: usize = 0;
|
var index: usize = 0;
|
||||||
while (index < len) {
|
while (index < len) {
|
||||||
const headers_finished = req.response_headers.state == .finished;
|
const headers_finished = req.response.state == .finished;
|
||||||
const amt = try readAdvanced(req, buffer[index..]);
|
const amt = try readAdvanced(req, buffer[index..]);
|
||||||
if (amt == 0 and headers_finished) break;
|
if (amt == 0 and headers_finished) break;
|
||||||
index += amt;
|
index += amt;
|
||||||
@ -82,100 +467,126 @@ pub const Request = struct {
|
|||||||
/// This one can return 0 without meaning EOF.
|
/// This one can return 0 without meaning EOF.
|
||||||
/// TODO change to readvAdvanced
|
/// TODO change to readvAdvanced
|
||||||
pub fn readAdvanced(req: *Request, buffer: []u8) !usize {
|
pub fn readAdvanced(req: *Request, buffer: []u8) !usize {
|
||||||
if (req.response_headers.state == .finished) return readRaw(req, buffer);
|
if (req.response.state == .finished) return req.connection.read(buffer);
|
||||||
|
|
||||||
const amt = try readRaw(req, buffer);
|
const amt = try req.connection.read(buffer);
|
||||||
const data = buffer[0..amt];
|
const data = buffer[0..amt];
|
||||||
const i = req.response_headers.feed(data);
|
const i = req.response.findHeadersEnd(data);
|
||||||
if (req.response_headers.state == .invalid) return error.InvalidHttpHeaders;
|
if (req.response.state == .invalid) return error.HttpHeadersInvalid;
|
||||||
if (i < data.len) {
|
|
||||||
const rest = data[i..];
|
const headers_data = data[0..i];
|
||||||
std.mem.copy(u8, buffer, rest);
|
if (req.response.header_bytes.items.len + headers_data.len > req.response.max_header_bytes) {
|
||||||
return rest.len;
|
return error.HttpHeadersExceededSizeLimit;
|
||||||
|
}
|
||||||
|
try req.response.header_bytes.appendSlice(req.client.allocator, headers_data);
|
||||||
|
|
||||||
|
if (req.response.state == .finished) {
|
||||||
|
req.response.headers = try Response.Headers.parse(req.response.header_bytes.items);
|
||||||
|
}
|
||||||
|
|
||||||
|
if (req.response.headers.status.class() == .redirect) {
|
||||||
|
if (req.redirects_left == 0) return error.TooManyHttpRedirects;
|
||||||
|
const location = req.response.headers.location orelse
|
||||||
|
return error.HttpRedirectMissingLocation;
|
||||||
|
const new_url = try std.Url.parse(location);
|
||||||
|
const new_req = try req.client.request(new_url, req.headers, .{
|
||||||
|
.max_redirects = req.redirects_left - 1,
|
||||||
|
.header_strategy = if (req.response.header_bytes_owned) .{
|
||||||
|
.dynamic = req.response.max_header_bytes,
|
||||||
|
} else .{
|
||||||
|
.static = req.response.header_bytes.unusedCapacitySlice(),
|
||||||
|
},
|
||||||
|
});
|
||||||
|
req.deinit();
|
||||||
|
req.* = new_req;
|
||||||
|
return readAdvanced(req, buffer);
|
||||||
|
}
|
||||||
|
|
||||||
|
const body_data = data[i..];
|
||||||
|
if (body_data.len > 0) {
|
||||||
|
mem.copy(u8, buffer, body_data);
|
||||||
|
return body_data.len;
|
||||||
}
|
}
|
||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Only abstracts over http/https.
|
test {
|
||||||
fn readRaw(req: *Request, buffer: []u8) !usize {
|
_ = Response;
|
||||||
switch (req.protocol) {
|
|
||||||
.http => return req.stream.read(buffer),
|
|
||||||
.https => return req.tls_client.read(req.stream, buffer),
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
/// Only abstracts over http/https.
|
|
||||||
fn readAtLeastRaw(req: *Request, buffer: []u8, len: usize) !usize {
|
|
||||||
switch (req.protocol) {
|
|
||||||
.http => return req.stream.readAtLeast(buffer, len),
|
|
||||||
.https => return req.tls_client.readAtLeast(req.stream, buffer, len),
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
};
|
};
|
||||||
|
|
||||||
pub fn deinit(client: *Client) void {
|
pub fn deinit(client: *Client, gpa: Allocator) void {
|
||||||
assert(client.active_requests == 0);
|
client.ca_bundle.deinit(gpa);
|
||||||
client.headers.deinit(client.allocator);
|
|
||||||
client.* = undefined;
|
client.* = undefined;
|
||||||
}
|
}
|
||||||
|
|
||||||
pub fn request(client: *Client, url: Url, options: Request.Options) !Request {
|
pub fn connect(client: *Client, host: []const u8, port: u16, protocol: Connection.Protocol) !Connection {
|
||||||
const protocol = std.meta.stringToEnum(Request.Protocol, url.scheme) orelse
|
var conn: Connection = .{
|
||||||
|
.stream = try net.tcpConnectToHost(client.allocator, host, port),
|
||||||
|
.tls_client = undefined,
|
||||||
|
.protocol = protocol,
|
||||||
|
};
|
||||||
|
|
||||||
|
switch (protocol) {
|
||||||
|
.plain => {},
|
||||||
|
.tls => {
|
||||||
|
conn.tls_client = try std.crypto.tls.Client.init(conn.stream, client.ca_bundle, host);
|
||||||
|
// This is appropriate for HTTPS because the HTTP headers contain
|
||||||
|
// the content length which is used to detect truncation attacks.
|
||||||
|
conn.tls_client.allow_truncation_attacks = true;
|
||||||
|
},
|
||||||
|
}
|
||||||
|
|
||||||
|
return conn;
|
||||||
|
}
|
||||||
|
|
||||||
|
pub fn request(client: *Client, url: Url, headers: Request.Headers, options: Request.Options) !Request {
|
||||||
|
const protocol: Connection.Protocol = if (mem.eql(u8, url.scheme, "http"))
|
||||||
|
.plain
|
||||||
|
else if (mem.eql(u8, url.scheme, "https"))
|
||||||
|
.tls
|
||||||
|
else
|
||||||
return error.UnsupportedUrlScheme;
|
return error.UnsupportedUrlScheme;
|
||||||
|
|
||||||
const port: u16 = url.port orelse switch (protocol) {
|
const port: u16 = url.port orelse switch (protocol) {
|
||||||
.http => 80,
|
.plain => 80,
|
||||||
.https => 443,
|
.tls => 443,
|
||||||
};
|
};
|
||||||
|
|
||||||
var req: Request = .{
|
var req: Request = .{
|
||||||
.client = client,
|
.client = client,
|
||||||
.stream = try net.tcpConnectToHost(client.allocator, url.host, port),
|
.headers = headers,
|
||||||
.protocol = protocol,
|
.connection = try client.connect(url.host, port, protocol),
|
||||||
.tls_client = undefined,
|
.redirects_left = options.max_redirects,
|
||||||
};
|
.response = switch (options.header_strategy) {
|
||||||
client.active_requests += 1;
|
.dynamic => |max| Request.Response.initDynamic(max),
|
||||||
errdefer req.deinit();
|
.static => |buf| Request.Response.initStatic(buf),
|
||||||
|
|
||||||
switch (protocol) {
|
|
||||||
.http => {},
|
|
||||||
.https => {
|
|
||||||
req.tls_client = try std.crypto.tls.Client.init(req.stream, client.ca_bundle, url.host);
|
|
||||||
// This is appropriate for HTTPS because the HTTP headers contain
|
|
||||||
// the content length which is used to detect truncation attacks.
|
|
||||||
req.tls_client.allow_truncation_attacks = true;
|
|
||||||
},
|
},
|
||||||
}
|
};
|
||||||
|
|
||||||
try req.headers.ensureUnusedCapacity(
|
{
|
||||||
client.allocator,
|
var h = try std.BoundedArray(u8, 1000).init(0);
|
||||||
@tagName(options.method).len +
|
try h.appendSlice(@tagName(headers.method));
|
||||||
1 +
|
try h.appendSlice(" ");
|
||||||
url.path.len +
|
try h.appendSlice(url.path);
|
||||||
" HTTP/1.1\r\nHost: ".len +
|
try h.appendSlice(" HTTP/1.1\r\nHost: ");
|
||||||
url.host.len +
|
try h.appendSlice(url.host);
|
||||||
"\r\nUpgrade-Insecure-Requests: 1\r\n".len +
|
try h.appendSlice("\r\nConnection: close\r\n\r\n");
|
||||||
client.headers.items.len +
|
|
||||||
2, // for the \r\n at the end of headers
|
const header_bytes = h.slice();
|
||||||
);
|
try req.connection.writeAll(header_bytes);
|
||||||
req.headers.appendSliceAssumeCapacity(@tagName(options.method));
|
|
||||||
req.headers.appendSliceAssumeCapacity(" ");
|
|
||||||
req.headers.appendSliceAssumeCapacity(url.path);
|
|
||||||
req.headers.appendSliceAssumeCapacity(" HTTP/1.1\r\nHost: ");
|
|
||||||
req.headers.appendSliceAssumeCapacity(url.host);
|
|
||||||
switch (protocol) {
|
|
||||||
.https => req.headers.appendSliceAssumeCapacity("\r\nUpgrade-Insecure-Requests: 1\r\n"),
|
|
||||||
.http => req.headers.appendSliceAssumeCapacity("\r\n"),
|
|
||||||
}
|
}
|
||||||
req.headers.appendSliceAssumeCapacity(client.headers.items);
|
|
||||||
|
|
||||||
return req;
|
return req;
|
||||||
}
|
}
|
||||||
|
|
||||||
pub fn addHeader(client: *Client, name: []const u8, value: []const u8) !void {
|
test {
|
||||||
const gpa = client.allocator;
|
const builtin = @import("builtin");
|
||||||
try client.headers.ensureUnusedCapacity(gpa, name.len + value.len + 4);
|
const native_endian = comptime builtin.cpu.arch.endian();
|
||||||
client.headers.appendSliceAssumeCapacity(name);
|
if (builtin.zig_backend == .stage2_llvm and native_endian == .Big) {
|
||||||
client.headers.appendSliceAssumeCapacity(": ");
|
// https://github.com/ziglang/zig/issues/13782
|
||||||
client.headers.appendSliceAssumeCapacity(value);
|
return error.SkipZigTest;
|
||||||
client.headers.appendSliceAssumeCapacity("\r\n");
|
}
|
||||||
|
|
||||||
|
_ = Request;
|
||||||
}
|
}
|
||||||
|
|||||||
@ -86,7 +86,8 @@ pub fn VectorCount(comptime VectorType: type) type {
|
|||||||
|
|
||||||
/// Returns a vector containing the first `len` integers in order from 0 to `len`-1.
|
/// Returns a vector containing the first `len` integers in order from 0 to `len`-1.
|
||||||
/// For example, `iota(i32, 8)` will return a vector containing `.{0, 1, 2, 3, 4, 5, 6, 7}`.
|
/// For example, `iota(i32, 8)` will return a vector containing `.{0, 1, 2, 3, 4, 5, 6, 7}`.
|
||||||
pub fn iota(comptime T: type, comptime len: usize) @Vector(len, T) {
|
pub inline fn iota(comptime T: type, comptime len: usize) @Vector(len, T) {
|
||||||
|
comptime {
|
||||||
var out: [len]T = undefined;
|
var out: [len]T = undefined;
|
||||||
for (out) |*element, i| {
|
for (out) |*element, i| {
|
||||||
element.* = switch (@typeInfo(T)) {
|
element.* = switch (@typeInfo(T)) {
|
||||||
@ -97,6 +98,7 @@ pub fn iota(comptime T: type, comptime len: usize) @Vector(len, T) {
|
|||||||
}
|
}
|
||||||
return @as(@Vector(len, T), out);
|
return @as(@Vector(len, T), out);
|
||||||
}
|
}
|
||||||
|
}
|
||||||
|
|
||||||
/// Returns a vector containing the same elements as the input, but repeated until the desired length is reached.
|
/// Returns a vector containing the same elements as the input, but repeated until the desired length is reached.
|
||||||
/// For example, `repeat(8, [_]u32{1, 2, 3})` will return a vector containing `.{1, 2, 3, 1, 2, 3, 1, 2}`.
|
/// For example, `repeat(8, [_]u32{1, 2, 3})` will return a vector containing `.{1, 2, 3, 1, 2, 3, 1, 2}`.
|
||||||
|
|||||||
@ -109,7 +109,7 @@ fn testBitCastuXToBytes(comptime N: usize) !void {
|
|||||||
const bytes = std.mem.asBytes(&x);
|
const bytes = std.mem.asBytes(&x);
|
||||||
|
|
||||||
const byte_count = (N + 7) / 8;
|
const byte_count = (N + 7) / 8;
|
||||||
switch (builtin.cpu.arch.endian()) {
|
switch (native_endian) {
|
||||||
.Little => {
|
.Little => {
|
||||||
var byte_i = 0;
|
var byte_i = 0;
|
||||||
while (byte_i < (byte_count - 1)) : (byte_i += 1) {
|
while (byte_i < (byte_count - 1)) : (byte_i += 1) {
|
||||||
@ -333,7 +333,7 @@ test "comptime @bitCast packed struct to int and back" {
|
|||||||
if (builtin.zig_backend == .stage2_arm) return error.SkipZigTest;
|
if (builtin.zig_backend == .stage2_arm) return error.SkipZigTest;
|
||||||
if (builtin.zig_backend == .stage2_sparc64) return error.SkipZigTest; // TODO
|
if (builtin.zig_backend == .stage2_sparc64) return error.SkipZigTest; // TODO
|
||||||
|
|
||||||
if (comptime builtin.zig_backend == .stage2_llvm and builtin.cpu.arch.endian() == .Big) {
|
if (builtin.zig_backend == .stage2_llvm and native_endian == .Big) {
|
||||||
// https://github.com/ziglang/zig/issues/13782
|
// https://github.com/ziglang/zig/issues/13782
|
||||||
return error.SkipZigTest;
|
return error.SkipZigTest;
|
||||||
}
|
}
|
||||||
|
|||||||
Loading…
x
Reference in New Issue
Block a user