Compare commits
11 Commits
ec13a4a5b3
...
c6365b5c03
Author | SHA1 | Date | |
---|---|---|---|
|
c6365b5c03 | ||
|
f056b8f434 | ||
|
33f4f2520c | ||
|
1984d47151 | ||
|
bfca633917 | ||
|
af0f13008a | ||
|
d69e958304 | ||
|
509bd8963d | ||
|
3e5be4cf8f | ||
|
dd4006d03d | ||
|
b92ddd66b2 |
138
Daniel Chen.typ
138
Daniel Chen.typ
@ -32,33 +32,21 @@
|
|||||||
= Skills
|
= Skills
|
||||||
|
|
||||||
#indent[
|
#indent[
|
||||||
*Languages:* Python, TypeScript, JavaScript, SQL, Rust, C++, Java, HTML, CSS \
|
*Languages:* Python, TypeScript, JavaScript, SQL, Rust, C/C++, Java, HTML/CSS \
|
||||||
// Dart, Ruby
|
// Dart, Ruby
|
||||||
*Frameworks:* React, Node.js, Vue, Flutter, Spring Boot, Next.js, Express,
|
*Technologies:* Node.js, React, Vue, Next.js, Spring Boot, Express, Prisma, Axum
|
||||||
Prisma, Axum \
|
\
|
||||||
// Qt, Jest / Pytest, Apollo, Prisma, Rails
|
// Qt, Jest / Pytest, Apollo, Prisma, Rails
|
||||||
*Tools:* Linux, Git, Docker, PostgreSQL, AWS/GCP, Redis, Kubernetes, Terraform,
|
*Tools:* Linux, Git, Docker, PostgreSQL, Pytest, Jest, AWS/GCP, Kubernetes,
|
||||||
GraphQL, nginx // Jenkins, systemd, Jira
|
Terraform // Jenkins, systemd, Jira
|
||||||
]
|
]
|
||||||
|
|
||||||
= Experience
|
= Experience
|
||||||
|
|
||||||
#indent[
|
#indent[
|
||||||
// #work_entry(
|
|
||||||
// "Technical Volunteer",
|
|
||||||
// "Organization for Transformative Works",
|
|
||||||
// start_date: "Feb 2024",
|
|
||||||
// end_date: "present",
|
|
||||||
// location: "Remote",
|
|
||||||
// tools: "Python, Ruby on Rails, React, MySQL",
|
|
||||||
// tasks: (
|
|
||||||
// [Upgrading archive import pipelines to migrate 4000+ user works to be accessible
|
|
||||||
// to *7 million registered users*],
|
|
||||||
// ),
|
|
||||||
// )
|
|
||||||
|
|
||||||
#work_entry(
|
#work_entry(
|
||||||
"Software Developer Intern",
|
"Software Engineering Intern",
|
||||||
"Immigr8",
|
"Immigr8",
|
||||||
start_date: "Jan 2024",
|
start_date: "Jan 2024",
|
||||||
end_date: "Apr 2024",
|
end_date: "Apr 2024",
|
||||||
@ -66,15 +54,17 @@
|
|||||||
tools: "Next.js, Spring Boot, PostgreSQL, AWS, Terraform",
|
tools: "Next.js, Spring Boot, PostgreSQL, AWS, Terraform",
|
||||||
tasks: (
|
tasks: (
|
||||||
// [Improved client responsiveness by *80* by employing stale-while-revalidate hooks for caching and ?? SWR performance, caching, error handling],
|
// [Improved client responsiveness by *80* by employing stale-while-revalidate hooks for caching and ?? SWR performance, caching, error handling],
|
||||||
[Reduced manual user inputs by *40%* by extracting data from users' uploaded
|
[Optimised dashboard performance by *80%* by employing stale-while-revalidate
|
||||||
documents in S3 with Textract],
|
hooks in React and lazily loading backend models with Hibernate, improving user
|
||||||
[Planned and executed a roadmap to SOC 2 security compliance by *implementing
|
retention by *30%*],
|
||||||
role-based permissions with JWTs*, sending backend logs and alerts to
|
[Achieved SOC 2 security compliance by implementing *role-based permissions with
|
||||||
CloudWatch, and introducing new privacy controls for user data],
|
JWTs* in Spring Boot, logging alerts to CloudWatch, and introducing new privacy
|
||||||
[*Introduced CI/CD for AWS Lambdas* by codifying them in Terraform for
|
controls for user data],
|
||||||
reproducible and automatic deployments],
|
[Owned the creation of an end-to-end document vault feature for users to upload
|
||||||
[Set up a staging environment via *EC2 and Jenkins* with parallel builds to
|
and verify supplementary identity documents with Textract, reducing support
|
||||||
*reduce production errors by 90%*],
|
tickets by *25%*],
|
||||||
|
[Created CI/CD pipelines to automatically deploy AWS Lambdas with Terraform and
|
||||||
|
Jenkins],
|
||||||
),
|
),
|
||||||
)
|
)
|
||||||
|
|
||||||
@ -86,16 +76,16 @@
|
|||||||
location: "Remote",
|
location: "Remote",
|
||||||
tools: "React, Python, PostgreSQL, AWS",
|
tools: "React, Python, PostgreSQL, AWS",
|
||||||
tasks: (
|
tasks: (
|
||||||
[Owned and launched a HubSpot/Airtable to Google Sheets sync feature to *46000+
|
[Owned the creation of a new feature to sync *46000+* enterprise users'
|
||||||
enterprise users* by writing AWS Lambda functions in Python, abstracting
|
HubSpot/Airtable data to Google Sheets and emailed CSVs to databases with AWS
|
||||||
interfaces to easily support future sources],
|
Lambda and Python],
|
||||||
[Wrote and deployed a new *React* data query builder for our business
|
[Engineered a new React data query builder and FastAPI backend for our business
|
||||||
intelligence visualisation service to *33000+ users*, eliminating the need for
|
intelligence visualisation service to *33000+* users, eliminating the need for
|
||||||
clients to manually write API calls or Python code],
|
clients to manually write API calls],
|
||||||
[Designed a conflict resolution algorithm for merging databases, *reducing
|
[Designed a conflict resolution algorithm to merge databases, reducing
|
||||||
unrecoverable errors by 75%*],
|
unrecoverable errors by *75%*],
|
||||||
[Developed a *serverless microservice* that asynchronously ingests email
|
// [?? new tool dDeveloped a *serverless microservice* that asynchronously ingests
|
||||||
attachments directly into ETL pipelines],
|
// email attachments directly into ETL pipelines],
|
||||||
// [??? emphasise test driven development or remove line *Reduced request errors by 30%* by creating backend testing infrastructure with
|
// [??? emphasise test driven development or remove line *Reduced request errors by 30%* by creating backend testing infrastructure with
|
||||||
// *Pytest* to catch regressions and test new features across database engines with
|
// *Pytest* to catch regressions and test new features across database engines with
|
||||||
// *90% code coverage*],
|
// *90% code coverage*],
|
||||||
@ -103,23 +93,39 @@
|
|||||||
)
|
)
|
||||||
|
|
||||||
#work_entry(
|
#work_entry(
|
||||||
"Backend Lead",
|
"Backend Developer",
|
||||||
"Hack the North",
|
"Hack the North",
|
||||||
start_date: "Mar 2023",
|
start_date: "Mar 2023",
|
||||||
end_date: "present",
|
end_date: "present",
|
||||||
location: "Waterloo, ON",
|
location: "Waterloo, ON",
|
||||||
tools: "Express, Redis, GraphQL, Docker, Kubernetes",
|
tools: "Express, Redis, GraphQL, Docker, Kubernetes",
|
||||||
tasks: (
|
tasks: (
|
||||||
[Shipped a new *bus registration tool* for attendees by creating new database
|
[Built a new travel tool application leveraging GraphQL and Apollo Server to
|
||||||
tables and GraphQL resolvers],
|
track bus capacity for *500+* attendees through real-time registration processes],
|
||||||
[Created a real-time event statistics dashboard by implementing GraphQL
|
[Streamlined event check-in times by *80%* for *1900+* guests by generating
|
||||||
*subscriptions with Redis PubSub*],
|
Apple/Google Wallet passes over Express],
|
||||||
[Diagnosed and *recovered from load spike crashes* by debouncing queries, using
|
[Diagnosed and recovered from crashes due to load spikes up to 800k updates per
|
||||||
rate limiting as a stopgap],
|
hour at the world's largest student-run hackathon by autoscaling Redis and
|
||||||
[Reduced event check-in times by *80%* for *1900+ attendees* by generating
|
Postgres on Kubernetes],
|
||||||
digital wallet passes over Express],
|
[Slashed dev server reload time from *130 → 9 s* by introducing threading,
|
||||||
[*Led and mentored a team of 6* to ideate, scope, and deliver features for
|
profiling, and pre-transpiling generated code],
|
||||||
*Canada's largest hackathon*],
|
// [Upgraded the event statistics dashboard to update in real time with GraphQL
|
||||||
|
// *subscriptions with Redis PubSub*],
|
||||||
|
),
|
||||||
|
)
|
||||||
|
|
||||||
|
#work_entry(
|
||||||
|
"Technical Volunteer",
|
||||||
|
"Organization for Transformative Works",
|
||||||
|
start_date: "Feb 2024",
|
||||||
|
end_date: "present",
|
||||||
|
location: "Remote",
|
||||||
|
tools: "Python, Ruby, React, MySQL",
|
||||||
|
tasks: (
|
||||||
|
[Upgraded archive import pipelines in Python to make *4000+* user works
|
||||||
|
accessible to *7000000+* registered users],
|
||||||
|
[Developed a new React frontend for users to browse and search works with *Ruby
|
||||||
|
on Rails*],
|
||||||
),
|
),
|
||||||
)
|
)
|
||||||
]
|
]
|
||||||
@ -132,23 +138,27 @@
|
|||||||
"Mandown Comic Downloader",
|
"Mandown Comic Downloader",
|
||||||
"Python, Qt",
|
"Python, Qt",
|
||||||
repo_link: "https://github.com/potatoeggy/mandown",
|
repo_link: "https://github.com/potatoeggy/mandown",
|
||||||
|
tasks: (
|
||||||
|
[Created a CLI and GUI to scrape, download, and convert webcomics into
|
||||||
|
e-ink-optimised PDF/EPUBs],
|
||||||
|
[Designed a scalable ebook converter with a *plugin-based architecture* to
|
||||||
|
improve compatibility],
|
||||||
|
[Implemented *image caching and multithreading* to increase download and update
|
||||||
|
speeds by *300%*],
|
||||||
|
),
|
||||||
)
|
)
|
||||||
- Created a CLI and GUI to scrape, download, and convert webcomics into
|
|
||||||
e-ink-optimised ebooks (EPUB, PDF)
|
|
||||||
- Designed a *scalable data conversion system* via a canonical format, allowing
|
|
||||||
any format to convert to any other
|
|
||||||
- Implemented *image caching and multithreading* to increase download and update
|
|
||||||
speeds by *300%*
|
|
||||||
|
|
||||||
#project(
|
#project(
|
||||||
"Kobo Sync Server",
|
"Kobo Sync Server",
|
||||||
"Rust, Axum",
|
"Rust, Axum",
|
||||||
repo_link: "https://github.com/potatoeggy/kobink",
|
repo_link: "https://github.com/potatoeggy/kobink",
|
||||||
|
tasks: (
|
||||||
|
[Developed a *Rust* backend to sync custom files to Kobos by reimplementing the
|
||||||
|
Kobo Sync API],
|
||||||
|
[Deployed the service behind an nginx reverse proxy and Cloudflare DNS to sync
|
||||||
|
over the internet],
|
||||||
|
),
|
||||||
)
|
)
|
||||||
- *Reimplemented the Kobo Store Sync protocol* in an Axum backend to instead sync
|
|
||||||
personal ebook files
|
|
||||||
- Deployed the service behind an nginx reverse proxy and Cloudflare DNS to allow
|
|
||||||
syncing over the internet
|
|
||||||
|
|
||||||
// #project(
|
// #project(
|
||||||
// "Genshin Rewards Simulator",
|
// "Genshin Rewards Simulator",
|
||||||
@ -166,11 +176,13 @@
|
|||||||
"Vue, TypeScript, WebSockets",
|
"Vue, TypeScript, WebSockets",
|
||||||
repo_link: "https://github.com/potatoeggy/jeopardy",
|
repo_link: "https://github.com/potatoeggy/jeopardy",
|
||||||
demo_link: "https://jeopardy.eggworld.me/host",
|
demo_link: "https://jeopardy.eggworld.me/host",
|
||||||
|
tasks: (
|
||||||
|
[Developed a beautiful and polished game where participants buzz in on their
|
||||||
|
devices over *WebSockets*],
|
||||||
|
[Synchronised game state between clients and the host with shared types and the
|
||||||
|
*observer pattern*],
|
||||||
|
),
|
||||||
)
|
)
|
||||||
- Created a beautiful and polished remote game where participants buzz in on their
|
|
||||||
devices over WebSockets
|
|
||||||
- Synchronised game state between clients and the host with shared TypeScript
|
|
||||||
types and the observer pattern
|
|
||||||
|
|
||||||
// #project(
|
// #project(
|
||||||
// "RecipeReady",
|
// "RecipeReady",
|
||||||
@ -205,8 +217,8 @@
|
|||||||
|
|
||||||
#indent[
|
#indent[
|
||||||
#work_entry(
|
#work_entry(
|
||||||
|
"B.A.Sc. in Computer Engineering (Honours, Co-op)",
|
||||||
"University of Waterloo",
|
"University of Waterloo",
|
||||||
"B.A.Sc. in Computer Engineering",
|
|
||||||
start_date: "Sep 2022",
|
start_date: "Sep 2022",
|
||||||
end_date: "Apr 2027",
|
end_date: "Apr 2027",
|
||||||
location: "Waterloo, ON",
|
location: "Waterloo, ON",
|
||||||
|
77
lib.typ
77
lib.typ
@ -5,17 +5,22 @@
|
|||||||
|
|
||||||
#let TITLE_FONT = "Bitter"
|
#let TITLE_FONT = "Bitter"
|
||||||
#let HEADING_FONT = TITLE_FONT
|
#let HEADING_FONT = TITLE_FONT
|
||||||
#let BODY_FONT = "Open Sans"
|
#let BODY_FONT = "Calibri"
|
||||||
|
|
||||||
#let BODY_FONT_SIZE = 10pt
|
#let BODY_FONT_SIZE = 11pt
|
||||||
#let HEADING_FONT_SIZE = 0.95em
|
#let HEADING_FONT_SIZE = 0.8em
|
||||||
#let TITLE_FONT_SIZE = 2em
|
#let TITLE_FONT_SIZE = 1.8em
|
||||||
|
|
||||||
#let HEADING_LINE_GAP = 3mm
|
#let HEADING_LINE_GAP = 3mm
|
||||||
#let BODY_LINE_HEIGHT = 0.8em
|
#let BODY_LINE_HEIGHT = 0.65em
|
||||||
|
|
||||||
|
#let HEADING_DETAILS_LIST_SPACING = 0.8em
|
||||||
|
#let DETAILS_TITLE_FONT_SIZE = 1em
|
||||||
|
|
||||||
|
#let DETAILS_LIST_SPACING = 0.75em
|
||||||
|
|
||||||
// rest = not top
|
// rest = not top
|
||||||
#let PAGE_MARGINS = (rest: 0.5in, top: 0.4in, left: 0.4in, bottom: 0.4in)
|
#let PAGE_MARGINS = (right: 0.4in, left: 0.3in, rest: 0.4in)
|
||||||
|
|
||||||
// Format locations.
|
// Format locations.
|
||||||
#let format_location(location) = {
|
#let format_location(location) = {
|
||||||
@ -23,15 +28,15 @@
|
|||||||
}
|
}
|
||||||
|
|
||||||
#let indent(content) = {
|
#let indent(content) = {
|
||||||
block(inset: (left: 0.75em, right: 0.5em), content)
|
block(inset: (left: 0em, right: 0em), content)
|
||||||
}
|
}
|
||||||
|
|
||||||
// General entry that is split into a left and right half (for experience and education).
|
// General entry that is split into a left and right half (for experience and education).
|
||||||
#let cv_entry(left_content: none, right_content: none, details: none) = {
|
#let cv_entry(left_content: none, right_content: none, details: none) = {
|
||||||
stack(
|
stack(
|
||||||
dir: ttb,
|
dir: ttb,
|
||||||
spacing: 0.9em,
|
spacing: HEADING_DETAILS_LIST_SPACING,
|
||||||
grid(columns: (11fr, 3fr), column-gutter: 1cm, {
|
grid(columns: (13fr, 3fr), column-gutter: 0.5cm, {
|
||||||
set strong(delta: DEFAULT_STRONG)
|
set strong(delta: DEFAULT_STRONG)
|
||||||
set align(left)
|
set align(left)
|
||||||
left_content
|
left_content
|
||||||
@ -41,9 +46,8 @@
|
|||||||
right_content
|
right_content
|
||||||
}),
|
}),
|
||||||
{
|
{
|
||||||
set align(left)
|
set strong(delta: DEFAULT_STRONG)
|
||||||
set strong(delta: SMALL_STRONG)
|
list(tight: false, spacing: DETAILS_LIST_SPACING, marker: [•], ..details)
|
||||||
details
|
|
||||||
},
|
},
|
||||||
)
|
)
|
||||||
}
|
}
|
||||||
@ -58,52 +62,43 @@
|
|||||||
end_date: none,
|
end_date: none,
|
||||||
location: none,
|
location: none,
|
||||||
) = {
|
) = {
|
||||||
let task_list = if tasks != none {
|
|
||||||
let list = []
|
|
||||||
for task in tasks {
|
|
||||||
list += [- #task]
|
|
||||||
}
|
|
||||||
list
|
|
||||||
} else {
|
|
||||||
none
|
|
||||||
}
|
|
||||||
cv_entry(left_content: {
|
cv_entry(left_content: {
|
||||||
set par(leading: 0.75em)
|
set par(justify: true)
|
||||||
text(1.1em)[*#role*]
|
text(DETAILS_TITLE_FONT_SIZE)[*#company* | _#role _]
|
||||||
|
|
||||||
if tools != none {
|
if tools != none {
|
||||||
text[ | _ #tools _ ]
|
text(DETAILS_TITLE_FONT_SIZE)[ | *#tools* ]
|
||||||
}
|
}
|
||||||
|
|
||||||
"\n"
|
|
||||||
set strong(delta: SMALL_STRONG)
|
|
||||||
text[*#company*]
|
|
||||||
}, right_content: {
|
}, right_content: {
|
||||||
[*#start_date -- #end_date*]
|
[_#start_date -- #end_date _]
|
||||||
|
// if location != none {
|
||||||
if location != none {
|
// format_location(location)
|
||||||
"\n" + format_location(location)
|
// }
|
||||||
}
|
}, details: tasks)
|
||||||
}, details: task_list)
|
|
||||||
}
|
}
|
||||||
|
|
||||||
#let project(title, tools, repo_link: none, demo_link: none) = {
|
#let project(title, tools, repo_link: none, demo_link: none, tasks: none) = {
|
||||||
set strong(delta: DEFAULT_STRONG)
|
set strong(delta: DEFAULT_STRONG)
|
||||||
let text_link = if demo_link != none { demo_link } else { repo_link }
|
let text_link = if demo_link != none { demo_link } else { repo_link }
|
||||||
|
|
||||||
|
let content_title = {
|
||||||
if text_link != none {
|
if text_link != none {
|
||||||
link(text_link)[#text(1.1em)[*#title*]]
|
link(text_link)[#text[*#title*]]
|
||||||
} else {
|
} else {
|
||||||
text(1.1em)[*#title*]
|
text[*#title*]
|
||||||
}
|
}
|
||||||
|
|
||||||
if demo_link != none {
|
if demo_link != none {
|
||||||
link(demo_link)[#text(1.1em)[ #fa(link-icon) ]]
|
link(demo_link)[ #fa(link-icon) ]
|
||||||
}
|
}
|
||||||
|
|
||||||
if repo_link != none {
|
if repo_link != none {
|
||||||
link(repo_link)[#text(1.1em)[ #fa(github) ]]
|
link(repo_link)[ #fa(github) ]
|
||||||
}
|
}
|
||||||
[ | _ #tools _ ]
|
[ | _ #tools _ ]
|
||||||
|
}
|
||||||
|
|
||||||
|
cv_entry(left_content: content_title, right_content: none, details: tasks)
|
||||||
}
|
}
|
||||||
|
|
||||||
#let render_contact_data(data) = {
|
#let render_contact_data(data) = {
|
||||||
@ -118,9 +113,9 @@
|
|||||||
// Set name and contact data and format headings
|
// Set name and contact data and format headings
|
||||||
#let template(name, color, doc) = {
|
#let template(name, color, doc) = {
|
||||||
set page(margin: PAGE_MARGINS, paper: "us-letter")
|
set page(margin: PAGE_MARGINS, paper: "us-letter")
|
||||||
set list(indent: 1em, marker: [•])
|
set list(indent: 0.75em, marker: [•])
|
||||||
set text(font: (BODY_FONT), BODY_FONT_SIZE)
|
set text(font: (BODY_FONT), BODY_FONT_SIZE)
|
||||||
set par(justify: true, leading: BODY_LINE_HEIGHT)
|
set par(leading: BODY_LINE_HEIGHT)
|
||||||
align(center)[
|
align(center)[
|
||||||
#text(size: TITLE_FONT_SIZE, font: TITLE_FONT, fill: color)[*#name*]
|
#text(size: TITLE_FONT_SIZE, font: TITLE_FONT, fill: color)[*#name*]
|
||||||
#block(above: 0em, below: 1em)
|
#block(above: 0em, below: 1em)
|
||||||
|
Loading…
Reference in New Issue
Block a user