Prettier backfill

This commit is contained in:
Matt Rubens
2025-01-17 14:11:28 -05:00
parent 3bcb4ff8c5
commit 60a0a824b9
174 changed files with 15715 additions and 15428 deletions

View File

@@ -1,114 +1,114 @@
import { ApiStreamChunk } from '../stream';
import { ApiStreamChunk } from "../stream"
describe('API Stream Types', () => {
describe('ApiStreamChunk', () => {
it('should correctly handle text chunks', () => {
const textChunk: ApiStreamChunk = {
type: 'text',
text: 'Hello world'
};
describe("API Stream Types", () => {
describe("ApiStreamChunk", () => {
it("should correctly handle text chunks", () => {
const textChunk: ApiStreamChunk = {
type: "text",
text: "Hello world",
}
expect(textChunk.type).toBe('text');
expect(textChunk.text).toBe('Hello world');
});
expect(textChunk.type).toBe("text")
expect(textChunk.text).toBe("Hello world")
})
it('should correctly handle usage chunks with cache information', () => {
const usageChunk: ApiStreamChunk = {
type: 'usage',
inputTokens: 100,
outputTokens: 50,
cacheWriteTokens: 20,
cacheReadTokens: 10
};
it("should correctly handle usage chunks with cache information", () => {
const usageChunk: ApiStreamChunk = {
type: "usage",
inputTokens: 100,
outputTokens: 50,
cacheWriteTokens: 20,
cacheReadTokens: 10,
}
expect(usageChunk.type).toBe('usage');
expect(usageChunk.inputTokens).toBe(100);
expect(usageChunk.outputTokens).toBe(50);
expect(usageChunk.cacheWriteTokens).toBe(20);
expect(usageChunk.cacheReadTokens).toBe(10);
});
expect(usageChunk.type).toBe("usage")
expect(usageChunk.inputTokens).toBe(100)
expect(usageChunk.outputTokens).toBe(50)
expect(usageChunk.cacheWriteTokens).toBe(20)
expect(usageChunk.cacheReadTokens).toBe(10)
})
it('should handle usage chunks without cache tokens', () => {
const usageChunk: ApiStreamChunk = {
type: 'usage',
inputTokens: 100,
outputTokens: 50
};
it("should handle usage chunks without cache tokens", () => {
const usageChunk: ApiStreamChunk = {
type: "usage",
inputTokens: 100,
outputTokens: 50,
}
expect(usageChunk.type).toBe('usage');
expect(usageChunk.inputTokens).toBe(100);
expect(usageChunk.outputTokens).toBe(50);
expect(usageChunk.cacheWriteTokens).toBeUndefined();
expect(usageChunk.cacheReadTokens).toBeUndefined();
});
expect(usageChunk.type).toBe("usage")
expect(usageChunk.inputTokens).toBe(100)
expect(usageChunk.outputTokens).toBe(50)
expect(usageChunk.cacheWriteTokens).toBeUndefined()
expect(usageChunk.cacheReadTokens).toBeUndefined()
})
it('should handle text chunks with empty strings', () => {
const emptyTextChunk: ApiStreamChunk = {
type: 'text',
text: ''
};
it("should handle text chunks with empty strings", () => {
const emptyTextChunk: ApiStreamChunk = {
type: "text",
text: "",
}
expect(emptyTextChunk.type).toBe('text');
expect(emptyTextChunk.text).toBe('');
});
expect(emptyTextChunk.type).toBe("text")
expect(emptyTextChunk.text).toBe("")
})
it('should handle usage chunks with zero tokens', () => {
const zeroUsageChunk: ApiStreamChunk = {
type: 'usage',
inputTokens: 0,
outputTokens: 0
};
it("should handle usage chunks with zero tokens", () => {
const zeroUsageChunk: ApiStreamChunk = {
type: "usage",
inputTokens: 0,
outputTokens: 0,
}
expect(zeroUsageChunk.type).toBe('usage');
expect(zeroUsageChunk.inputTokens).toBe(0);
expect(zeroUsageChunk.outputTokens).toBe(0);
});
expect(zeroUsageChunk.type).toBe("usage")
expect(zeroUsageChunk.inputTokens).toBe(0)
expect(zeroUsageChunk.outputTokens).toBe(0)
})
it('should handle usage chunks with large token counts', () => {
const largeUsageChunk: ApiStreamChunk = {
type: 'usage',
inputTokens: 1000000,
outputTokens: 500000,
cacheWriteTokens: 200000,
cacheReadTokens: 100000
};
it("should handle usage chunks with large token counts", () => {
const largeUsageChunk: ApiStreamChunk = {
type: "usage",
inputTokens: 1000000,
outputTokens: 500000,
cacheWriteTokens: 200000,
cacheReadTokens: 100000,
}
expect(largeUsageChunk.type).toBe('usage');
expect(largeUsageChunk.inputTokens).toBe(1000000);
expect(largeUsageChunk.outputTokens).toBe(500000);
expect(largeUsageChunk.cacheWriteTokens).toBe(200000);
expect(largeUsageChunk.cacheReadTokens).toBe(100000);
});
expect(largeUsageChunk.type).toBe("usage")
expect(largeUsageChunk.inputTokens).toBe(1000000)
expect(largeUsageChunk.outputTokens).toBe(500000)
expect(largeUsageChunk.cacheWriteTokens).toBe(200000)
expect(largeUsageChunk.cacheReadTokens).toBe(100000)
})
it('should handle text chunks with special characters', () => {
const specialCharsChunk: ApiStreamChunk = {
type: 'text',
text: '!@#$%^&*()_+-=[]{}|;:,.<>?`~'
};
it("should handle text chunks with special characters", () => {
const specialCharsChunk: ApiStreamChunk = {
type: "text",
text: "!@#$%^&*()_+-=[]{}|;:,.<>?`~",
}
expect(specialCharsChunk.type).toBe('text');
expect(specialCharsChunk.text).toBe('!@#$%^&*()_+-=[]{}|;:,.<>?`~');
});
expect(specialCharsChunk.type).toBe("text")
expect(specialCharsChunk.text).toBe("!@#$%^&*()_+-=[]{}|;:,.<>?`~")
})
it('should handle text chunks with unicode characters', () => {
const unicodeChunk: ApiStreamChunk = {
type: 'text',
text: '你好世界👋🌍'
};
it("should handle text chunks with unicode characters", () => {
const unicodeChunk: ApiStreamChunk = {
type: "text",
text: "你好世界👋🌍",
}
expect(unicodeChunk.type).toBe('text');
expect(unicodeChunk.text).toBe('你好世界👋🌍');
});
expect(unicodeChunk.type).toBe("text")
expect(unicodeChunk.text).toBe("你好世界👋🌍")
})
it('should handle text chunks with multiline content', () => {
const multilineChunk: ApiStreamChunk = {
type: 'text',
text: 'Line 1\nLine 2\nLine 3'
};
it("should handle text chunks with multiline content", () => {
const multilineChunk: ApiStreamChunk = {
type: "text",
text: "Line 1\nLine 2\nLine 3",
}
expect(multilineChunk.type).toBe('text');
expect(multilineChunk.text).toBe('Line 1\nLine 2\nLine 3');
expect(multilineChunk.text.split('\n')).toHaveLength(3);
});
});
});
expect(multilineChunk.type).toBe("text")
expect(multilineChunk.text).toBe("Line 1\nLine 2\nLine 3")
expect(multilineChunk.text.split("\n")).toHaveLength(3)
})
})
})