Compare commits

..

No commits in common. "c6365b5c03de2daf704facaaf9c2e9383100d437" and "ec13a4a5b376dbfd4d7774d562070a60a99341e2" have entirely different histories.

2 changed files with 112 additions and 119 deletions

View File

@ -32,21 +32,33 @@
= Skills = Skills
#indent[ #indent[
*Languages:* Python, TypeScript, JavaScript, SQL, Rust, C/C++, Java, HTML/CSS \ *Languages:* Python, TypeScript, JavaScript, SQL, Rust, C++, Java, HTML, CSS \
// Dart, Ruby // Dart, Ruby
*Technologies:* Node.js, React, Vue, Next.js, Spring Boot, Express, Prisma, Axum *Frameworks:* React, Node.js, Vue, Flutter, Spring Boot, Next.js, Express,
\ Prisma, Axum \
// Qt, Jest / Pytest, Apollo, Prisma, Rails // Qt, Jest / Pytest, Apollo, Prisma, Rails
*Tools:* Linux, Git, Docker, PostgreSQL, Pytest, Jest, AWS/GCP, Kubernetes, *Tools:* Linux, Git, Docker, PostgreSQL, AWS/GCP, Redis, Kubernetes, Terraform,
Terraform // Jenkins, systemd, Jira GraphQL, nginx // Jenkins, systemd, Jira
] ]
= Experience = Experience
#indent[ #indent[
// #work_entry(
// "Technical Volunteer",
// "Organization for Transformative Works",
// start_date: "Feb 2024",
// end_date: "present",
// location: "Remote",
// tools: "Python, Ruby on Rails, React, MySQL",
// tasks: (
// [Upgrading archive import pipelines to migrate 4000+ user works to be accessible
// to *7 million registered users*],
// ),
// )
#work_entry( #work_entry(
"Software Engineering Intern", "Software Developer Intern",
"Immigr8", "Immigr8",
start_date: "Jan 2024", start_date: "Jan 2024",
end_date: "Apr 2024", end_date: "Apr 2024",
@ -54,17 +66,15 @@
tools: "Next.js, Spring Boot, PostgreSQL, AWS, Terraform", tools: "Next.js, Spring Boot, PostgreSQL, AWS, Terraform",
tasks: ( tasks: (
// [Improved client responsiveness by *80* by employing stale-while-revalidate hooks for caching and ?? SWR performance, caching, error handling], // [Improved client responsiveness by *80* by employing stale-while-revalidate hooks for caching and ?? SWR performance, caching, error handling],
[Optimised dashboard performance by *80%* by employing stale-while-revalidate [Reduced manual user inputs by *40%* by extracting data from users' uploaded
hooks in React and lazily loading backend models with Hibernate, improving user documents in S3 with Textract],
retention by *30%*], [Planned and executed a roadmap to SOC 2 security compliance by *implementing
[Achieved SOC 2 security compliance by implementing *role-based permissions with role-based permissions with JWTs*, sending backend logs and alerts to
JWTs* in Spring Boot, logging alerts to CloudWatch, and introducing new privacy CloudWatch, and introducing new privacy controls for user data],
controls for user data], [*Introduced CI/CD for AWS Lambdas* by codifying them in Terraform for
[Owned the creation of an end-to-end document vault feature for users to upload reproducible and automatic deployments],
and verify supplementary identity documents with Textract, reducing support [Set up a staging environment via *EC2 and Jenkins* with parallel builds to
tickets by *25%*], *reduce production errors by 90%*],
[Created CI/CD pipelines to automatically deploy AWS Lambdas with Terraform and
Jenkins],
), ),
) )
@ -76,16 +86,16 @@
location: "Remote", location: "Remote",
tools: "React, Python, PostgreSQL, AWS", tools: "React, Python, PostgreSQL, AWS",
tasks: ( tasks: (
[Owned the creation of a new feature to sync *46000+* enterprise users' [Owned and launched a HubSpot/Airtable to Google Sheets sync feature to *46000+
HubSpot/Airtable data to Google Sheets and emailed CSVs to databases with AWS enterprise users* by writing AWS Lambda functions in Python, abstracting
Lambda and Python], interfaces to easily support future sources],
[Engineered a new React data query builder and FastAPI backend for our business [Wrote and deployed a new *React* data query builder for our business
intelligence visualisation service to *33000+* users, eliminating the need for intelligence visualisation service to *33000+ users*, eliminating the need for
clients to manually write API calls], clients to manually write API calls or Python code],
[Designed a conflict resolution algorithm to merge databases, reducing [Designed a conflict resolution algorithm for merging databases, *reducing
unrecoverable errors by *75%*], unrecoverable errors by 75%*],
// [?? new tool dDeveloped a *serverless microservice* that asynchronously ingests [Developed a *serverless microservice* that asynchronously ingests email
// email attachments directly into ETL pipelines], attachments directly into ETL pipelines],
// [??? emphasise test driven development or remove line *Reduced request errors by 30%* by creating backend testing infrastructure with // [??? emphasise test driven development or remove line *Reduced request errors by 30%* by creating backend testing infrastructure with
// *Pytest* to catch regressions and test new features across database engines with // *Pytest* to catch regressions and test new features across database engines with
// *90% code coverage*], // *90% code coverage*],
@ -93,39 +103,23 @@
) )
#work_entry( #work_entry(
"Backend Developer", "Backend Lead",
"Hack the North", "Hack the North",
start_date: "Mar 2023", start_date: "Mar 2023",
end_date: "present", end_date: "present",
location: "Waterloo, ON", location: "Waterloo, ON",
tools: "Express, Redis, GraphQL, Docker, Kubernetes", tools: "Express, Redis, GraphQL, Docker, Kubernetes",
tasks: ( tasks: (
[Built a new travel tool application leveraging GraphQL and Apollo Server to [Shipped a new *bus registration tool* for attendees by creating new database
track bus capacity for *500+* attendees through real-time registration processes], tables and GraphQL resolvers],
[Streamlined event check-in times by *80%* for *1900+* guests by generating [Created a real-time event statistics dashboard by implementing GraphQL
Apple/Google Wallet passes over Express], *subscriptions with Redis PubSub*],
[Diagnosed and recovered from crashes due to load spikes up to 800k updates per [Diagnosed and *recovered from load spike crashes* by debouncing queries, using
hour at the world's largest student-run hackathon by autoscaling Redis and rate limiting as a stopgap],
Postgres on Kubernetes], [Reduced event check-in times by *80%* for *1900+ attendees* by generating
[Slashed dev server reload time from *130 → 9 s* by introducing threading, digital wallet passes over Express],
profiling, and pre-transpiling generated code], [*Led and mentored a team of 6* to ideate, scope, and deliver features for
// [Upgraded the event statistics dashboard to update in real time with GraphQL *Canada's largest hackathon*],
// *subscriptions with Redis PubSub*],
),
)
#work_entry(
"Technical Volunteer",
"Organization for Transformative Works",
start_date: "Feb 2024",
end_date: "present",
location: "Remote",
tools: "Python, Ruby, React, MySQL",
tasks: (
[Upgraded archive import pipelines in Python to make *4000+* user works
accessible to *7000000+* registered users],
[Developed a new React frontend for users to browse and search works with *Ruby
on Rails*],
), ),
) )
] ]
@ -138,27 +132,23 @@
"Mandown Comic Downloader", "Mandown Comic Downloader",
"Python, Qt", "Python, Qt",
repo_link: "https://github.com/potatoeggy/mandown", repo_link: "https://github.com/potatoeggy/mandown",
tasks: (
[Created a CLI and GUI to scrape, download, and convert webcomics into
e-ink-optimised PDF/EPUBs],
[Designed a scalable ebook converter with a *plugin-based architecture* to
improve compatibility],
[Implemented *image caching and multithreading* to increase download and update
speeds by *300%*],
),
) )
- Created a CLI and GUI to scrape, download, and convert webcomics into
e-ink-optimised ebooks (EPUB, PDF)
- Designed a *scalable data conversion system* via a canonical format, allowing
any format to convert to any other
- Implemented *image caching and multithreading* to increase download and update
speeds by *300%*
#project( #project(
"Kobo Sync Server", "Kobo Sync Server",
"Rust, Axum", "Rust, Axum",
repo_link: "https://github.com/potatoeggy/kobink", repo_link: "https://github.com/potatoeggy/kobink",
tasks: (
[Developed a *Rust* backend to sync custom files to Kobos by reimplementing the
Kobo Sync API],
[Deployed the service behind an nginx reverse proxy and Cloudflare DNS to sync
over the internet],
),
) )
- *Reimplemented the Kobo Store Sync protocol* in an Axum backend to instead sync
personal ebook files
- Deployed the service behind an nginx reverse proxy and Cloudflare DNS to allow
syncing over the internet
// #project( // #project(
// "Genshin Rewards Simulator", // "Genshin Rewards Simulator",
@ -176,13 +166,11 @@
"Vue, TypeScript, WebSockets", "Vue, TypeScript, WebSockets",
repo_link: "https://github.com/potatoeggy/jeopardy", repo_link: "https://github.com/potatoeggy/jeopardy",
demo_link: "https://jeopardy.eggworld.me/host", demo_link: "https://jeopardy.eggworld.me/host",
tasks: (
[Developed a beautiful and polished game where participants buzz in on their
devices over *WebSockets*],
[Synchronised game state between clients and the host with shared types and the
*observer pattern*],
),
) )
- Created a beautiful and polished remote game where participants buzz in on their
devices over WebSockets
- Synchronised game state between clients and the host with shared TypeScript
types and the observer pattern
// #project( // #project(
// "RecipeReady", // "RecipeReady",
@ -217,8 +205,8 @@
#indent[ #indent[
#work_entry( #work_entry(
"B.A.Sc. in Computer Engineering (Honours, Co-op)",
"University of Waterloo", "University of Waterloo",
"B.A.Sc. in Computer Engineering",
start_date: "Sep 2022", start_date: "Sep 2022",
end_date: "Apr 2027", end_date: "Apr 2027",
location: "Waterloo, ON", location: "Waterloo, ON",

77
lib.typ
View File

@ -5,22 +5,17 @@
#let TITLE_FONT = "Bitter" #let TITLE_FONT = "Bitter"
#let HEADING_FONT = TITLE_FONT #let HEADING_FONT = TITLE_FONT
#let BODY_FONT = "Calibri" #let BODY_FONT = "Open Sans"
#let BODY_FONT_SIZE = 11pt #let BODY_FONT_SIZE = 10pt
#let HEADING_FONT_SIZE = 0.8em #let HEADING_FONT_SIZE = 0.95em
#let TITLE_FONT_SIZE = 1.8em #let TITLE_FONT_SIZE = 2em
#let HEADING_LINE_GAP = 3mm #let HEADING_LINE_GAP = 3mm
#let BODY_LINE_HEIGHT = 0.65em #let BODY_LINE_HEIGHT = 0.8em
#let HEADING_DETAILS_LIST_SPACING = 0.8em
#let DETAILS_TITLE_FONT_SIZE = 1em
#let DETAILS_LIST_SPACING = 0.75em
// rest = not top // rest = not top
#let PAGE_MARGINS = (right: 0.4in, left: 0.3in, rest: 0.4in) #let PAGE_MARGINS = (rest: 0.5in, top: 0.4in, left: 0.4in, bottom: 0.4in)
// Format locations. // Format locations.
#let format_location(location) = { #let format_location(location) = {
@ -28,15 +23,15 @@
} }
#let indent(content) = { #let indent(content) = {
block(inset: (left: 0em, right: 0em), content) block(inset: (left: 0.75em, right: 0.5em), content)
} }
// General entry that is split into a left and right half (for experience and education). // General entry that is split into a left and right half (for experience and education).
#let cv_entry(left_content: none, right_content: none, details: none) = { #let cv_entry(left_content: none, right_content: none, details: none) = {
stack( stack(
dir: ttb, dir: ttb,
spacing: HEADING_DETAILS_LIST_SPACING, spacing: 0.9em,
grid(columns: (13fr, 3fr), column-gutter: 0.5cm, { grid(columns: (11fr, 3fr), column-gutter: 1cm, {
set strong(delta: DEFAULT_STRONG) set strong(delta: DEFAULT_STRONG)
set align(left) set align(left)
left_content left_content
@ -46,8 +41,9 @@
right_content right_content
}), }),
{ {
set strong(delta: DEFAULT_STRONG) set align(left)
list(tight: false, spacing: DETAILS_LIST_SPACING, marker: [•], ..details) set strong(delta: SMALL_STRONG)
details
}, },
) )
} }
@ -62,43 +58,52 @@
end_date: none, end_date: none,
location: none, location: none,
) = { ) = {
let task_list = if tasks != none {
let list = []
for task in tasks {
list += [- #task]
}
list
} else {
none
}
cv_entry(left_content: { cv_entry(left_content: {
set par(justify: true) set par(leading: 0.75em)
text(DETAILS_TITLE_FONT_SIZE)[*#company* | _#role _] text(1.1em)[*#role*]
if tools != none { if tools != none {
text(DETAILS_TITLE_FONT_SIZE)[ | *#tools* ] text[ | _ #tools _ ]
} }
"\n"
set strong(delta: SMALL_STRONG)
text[*#company*]
}, right_content: { }, right_content: {
[_#start_date -- #end_date _] [*#start_date -- #end_date*]
// if location != none {
// format_location(location) if location != none {
// } "\n" + format_location(location)
}, details: tasks) }
}, details: task_list)
} }
#let project(title, tools, repo_link: none, demo_link: none, tasks: none) = { #let project(title, tools, repo_link: none, demo_link: none) = {
set strong(delta: DEFAULT_STRONG) set strong(delta: DEFAULT_STRONG)
let text_link = if demo_link != none { demo_link } else { repo_link } let text_link = if demo_link != none { demo_link } else { repo_link }
let content_title = {
if text_link != none { if text_link != none {
link(text_link)[#text[*#title*]] link(text_link)[#text(1.1em)[*#title*]]
} else { } else {
text[*#title*] text(1.1em)[*#title*]
} }
if demo_link != none { if demo_link != none {
link(demo_link)[ #fa(link-icon) ] link(demo_link)[#text(1.1em)[ #fa(link-icon) ]]
} }
if repo_link != none { if repo_link != none {
link(repo_link)[ #fa(github) ] link(repo_link)[#text(1.1em)[ #fa(github) ]]
} }
[ | _ #tools _ ] [ | _ #tools _ ]
}
cv_entry(left_content: content_title, right_content: none, details: tasks)
} }
#let render_contact_data(data) = { #let render_contact_data(data) = {
@ -113,9 +118,9 @@
// Set name and contact data and format headings // Set name and contact data and format headings
#let template(name, color, doc) = { #let template(name, color, doc) = {
set page(margin: PAGE_MARGINS, paper: "us-letter") set page(margin: PAGE_MARGINS, paper: "us-letter")
set list(indent: 0.75em, marker: [•]) set list(indent: 1em, marker: [•])
set text(font: (BODY_FONT), BODY_FONT_SIZE) set text(font: (BODY_FONT), BODY_FONT_SIZE)
set par(leading: BODY_LINE_HEIGHT) set par(justify: true, leading: BODY_LINE_HEIGHT)
align(center)[ align(center)[
#text(size: TITLE_FONT_SIZE, font: TITLE_FONT, fill: color)[*#name*] #text(size: TITLE_FONT_SIZE, font: TITLE_FONT, fill: color)[*#name*]
#block(above: 0em, below: 1em) #block(above: 0em, below: 1em)