feat: finalise web
This commit is contained in:
parent
4d7215700a
commit
1fdfe05eea
117
Daniel Chen.typ
117
Daniel Chen.typ
@ -30,7 +30,7 @@
|
|||||||
= Skills
|
= Skills
|
||||||
|
|
||||||
#indent[
|
#indent[
|
||||||
*Languages:* Python, TypeScript, JavaScript, Rust, Java, HTML, CSS, SQL \
|
*Languages:* Python, TypeScript, JavaScript, Rust, C++, Java, HTML, CSS, SQL \
|
||||||
// MATLAB, Dart, Ruby
|
// MATLAB, Dart, Ruby
|
||||||
*Frameworks:* Node.js, Express, Axum, Spring Boot, React, Next.js, Vue, Flutter,
|
*Frameworks:* Node.js, Express, Axum, Spring Boot, React, Next.js, Vue, Flutter,
|
||||||
Prisma \
|
Prisma \
|
||||||
@ -42,6 +42,19 @@
|
|||||||
= Experience
|
= Experience
|
||||||
|
|
||||||
#indent[
|
#indent[
|
||||||
|
#work_entry(
|
||||||
|
"Technical Volunteer",
|
||||||
|
"Organization for Transformative Works",
|
||||||
|
start_date: "Feb 2024",
|
||||||
|
end_date: "present",
|
||||||
|
location: "Remote",
|
||||||
|
tools: "Ruby on Rails, React, MySQL",
|
||||||
|
tasks: (
|
||||||
|
[Upgraded ETL pipelines to import 4 000+ user works to be accessible to *7+
|
||||||
|
million registered users*],
|
||||||
|
),
|
||||||
|
)
|
||||||
|
|
||||||
#work_entry(
|
#work_entry(
|
||||||
"Backend Lead & Event Organiser",
|
"Backend Lead & Event Organiser",
|
||||||
"Hack the North",
|
"Hack the North",
|
||||||
@ -50,16 +63,16 @@
|
|||||||
location: "Waterloo, ON",
|
location: "Waterloo, ON",
|
||||||
tools: "Express, GraphQL, Docker, Kubernetes",
|
tools: "Express, GraphQL, Docker, Kubernetes",
|
||||||
tasks: (
|
tasks: (
|
||||||
[??? leadership],
|
[*Led and mentored a team of 7* to clean technical debt and deliver *3 more
|
||||||
|
features* than the previous year],
|
||||||
[Streamlined event check-in times by *80%* for *1900+ attendees* by generating
|
[Streamlined event check-in times by *80%* for *1900+ attendees* by generating
|
||||||
digital wallet passes over *Express*],
|
digital wallet passes over Express],
|
||||||
[Integrated backend *tracing* to monitor performance and trigger alerts with
|
[Integrated backend tracing to monitor performance and trigger alerts with
|
||||||
*OpenTelemetry* and *Grafana*],
|
OpenTelemetry and Grafana],
|
||||||
[Diagnosed system scaling issues during load spikes and *recovered from a system
|
[Diagnosed and *recovered from load spike crashes* by debouncing queries, using
|
||||||
crash* by asking the frontend team to debounce their queries, implementing rate
|
rate limiting as a stopgap],
|
||||||
limiting as a temporary stopgap measure],
|
|
||||||
[Shipped a new *bus registration tool* for attendees by creating new database
|
[Shipped a new *bus registration tool* for attendees by creating new database
|
||||||
tables and *GraphQL* resolvers],
|
tables and GraphQL resolvers],
|
||||||
),
|
),
|
||||||
)
|
)
|
||||||
|
|
||||||
@ -68,18 +81,19 @@
|
|||||||
"Immigr8",
|
"Immigr8",
|
||||||
start_date: "Jan 2024",
|
start_date: "Jan 2024",
|
||||||
end_date: "Apr 2024",
|
end_date: "Apr 2024",
|
||||||
location: "Clearwater, FL",
|
location: "Remote",
|
||||||
tools: "Next.js, Spring Boot, PostgreSQL, AWS, Terraform",
|
tools: "Next.js, Spring Boot, PostgreSQL, AWS, Terraform",
|
||||||
tasks: (
|
tasks: (
|
||||||
[??? deduplicated user form inputs by implementing aws textract w/s3 + lambda],
|
[Reduced manual user inputs by *40%* by extracting data from users' uploaded
|
||||||
|
documents in S3 with *Textract*],
|
||||||
// [Improved client responsiveness by *80* by employing stale-while-revalidate hooks for caching and ?? SWR performance, caching, error handling],
|
// [Improved client responsiveness by *80* by employing stale-while-revalidate hooks for caching and ?? SWR performance, caching, error handling],
|
||||||
[Set up CI/CD for AWS infrastructure by *codifying it in Terraform* for
|
[Introduced CI/CD for AWS Lambdas by *codifying them in Terraform* for
|
||||||
reproducible and automatic deployments],
|
reproducible and automatic deployments],
|
||||||
[Reduced CI/CD pipeline build times by *70%* by setting up a staging environment
|
[Set up a staging environment via EC2 and Jenkins with parallel builds to *reduce
|
||||||
with *EC2* and *Jenkins*, parallelising builds, and caching dependencies],
|
production errors by 90%*],
|
||||||
[Planned and executed a roadmap to SOC 2 security compliance by *implementing
|
[Planned and executed a roadmap to SOC 2 security compliance by *implementing
|
||||||
role-based permissions with JWTs*, sending backend logs and alerts to
|
role-based permissions with JWTs*, sending backend logs and alerts to
|
||||||
*CloudWatch*, and introducing new privacy controls for user data],
|
CloudWatch, and introducing new privacy controls for user data],
|
||||||
),
|
),
|
||||||
)
|
)
|
||||||
|
|
||||||
@ -88,38 +102,24 @@
|
|||||||
"Dropbase",
|
"Dropbase",
|
||||||
start_date: "May 2023",
|
start_date: "May 2023",
|
||||||
end_date: "Aug 2023",
|
end_date: "Aug 2023",
|
||||||
location: "Toronto, ON",
|
location: "Toronto",
|
||||||
tools: "React, FastAPI, PostgreSQL, AWS",
|
tools: "React, FastAPI, PostgreSQL, AWS",
|
||||||
tasks: (
|
tasks: (
|
||||||
[Designed a conflict resolution algorithm for merging databases, *reducing
|
[Designed a conflict resolution algorithm for merging databases, *reducing
|
||||||
unrecoverable errors by 75%*],
|
unrecoverable errors by 75%*],
|
||||||
[Owned and launched a HubSpot/Airtable to Google Sheets sync feature to *46000+
|
[Owned and launched a HubSpot/Airtable to Google Sheets sync feature to *46000+
|
||||||
enterprise users* by writing *AWS Lambda* functions in Python, abstracting
|
enterprise users* by writing AWS Lambda functions in Python, abstracting
|
||||||
interfaces to easily support future sources],
|
interfaces to easily support future sources],
|
||||||
[Wrote and deployed a new *React* data query builder for our business
|
[Wrote and deployed a new React data query builder for our business intelligence
|
||||||
intelligence visualisation service to *33000+ users*, eliminating the need for
|
visualisation service to *33000+ users*, eliminating the need for clients to
|
||||||
clients to manually write API calls or Python code],
|
manually write API calls or Python code],
|
||||||
[Developed a scalable microservice that asynchronously ingests email attachments
|
[Developed a serverless microservice that asynchronously ingests email
|
||||||
directly into ETL pipelines],
|
attachments directly into ETL pipelines],
|
||||||
// [??? emphasise test driven development or remove line *Reduced request errors by 30%* by creating backend testing infrastructure with
|
// [??? emphasise test driven development or remove line *Reduced request errors by 30%* by creating backend testing infrastructure with
|
||||||
// *Pytest* to catch regressions and test new features across database engines with
|
// *Pytest* to catch regressions and test new features across database engines with
|
||||||
// *90% code coverage*],
|
// *90% code coverage*],
|
||||||
),
|
),
|
||||||
)
|
)
|
||||||
|
|
||||||
#work_entry(
|
|
||||||
"Technical Volunteer",
|
|
||||||
"Organization for Transformative Works",
|
|
||||||
start_date: "Feb 2024",
|
|
||||||
end_date: "present",
|
|
||||||
location: "Remote",
|
|
||||||
tools: "Ruby on Rails, React, MySQL",
|
|
||||||
tasks: (
|
|
||||||
[??? Automated importing NUMBER of fics by scraping it into a database then
|
|
||||||
import to AO3, largest fanfic archive],
|
|
||||||
[?? Increased reliability probably (how?)],
|
|
||||||
),
|
|
||||||
)
|
|
||||||
]
|
]
|
||||||
|
|
||||||
= Projects
|
= Projects
|
||||||
@ -128,35 +128,38 @@
|
|||||||
// joycond-cemuhook?
|
// joycond-cemuhook?
|
||||||
|
|
||||||
#indent[
|
#indent[
|
||||||
// #project("Comic Downloader", "Python, Qt",
|
|
||||||
// repo_link: "https://github.com/potatoeggy/mandown")
|
|
||||||
// - Created a CLI and GUI to scrape, download, and convert webcomics into
|
|
||||||
// e-ink-optimised ebooks (EPUB, PDF)
|
|
||||||
// - Designed an *intermediate representation* to store metadata and images as a
|
|
||||||
// standardised conversion target
|
|
||||||
// - Implemented *image caching and multiprocessing* to increase download and update
|
|
||||||
// speeds by *300%*
|
|
||||||
|
|
||||||
#project(
|
#project(
|
||||||
"Genshin Simulator",
|
"Mandown Comic Downloader",
|
||||||
"Vue, TypeScript",
|
"Python, Qt",
|
||||||
repo_link: "https://github.com/potatoeggy/primoprod",
|
repo_link: "https://github.com/potatoeggy/mandown",
|
||||||
demo_link: "https://primoprod.vercel.app",
|
|
||||||
)
|
)
|
||||||
- Built a realistic and beautiful *loot box simulator* with fluid animations based
|
- Created a CLI and GUI to scrape, download, and convert webcomics into
|
||||||
off a popular game
|
e-ink-optimised ebooks (EPUB, PDF)
|
||||||
- Reduced action latency by *90%* by lazily loading art assets at opportune times,
|
- Designed a scalable data conversion system via a *canonical format*, allowing
|
||||||
such as during cutscenes
|
any format to convert to any other
|
||||||
|
- Implemented *image caching and multithreading* to increase download and update
|
||||||
|
speeds by *300%*
|
||||||
|
|
||||||
|
// #project(
|
||||||
|
// "Genshin Simulator",
|
||||||
|
// "Vue, TypeScript",
|
||||||
|
// repo_link: "https://github.com/potatoeggy/primoprod",
|
||||||
|
// demo_link: "https://primoprod.vercel.app",
|
||||||
|
// )
|
||||||
|
// - Built a realistic and beautiful *loot box simulator* with CSS animations based
|
||||||
|
// off a popular game
|
||||||
|
// - Reduced action latency by *90%* by lazily loading art assets at opportune times,
|
||||||
|
// such as during cutscenes
|
||||||
|
|
||||||
#project(
|
#project(
|
||||||
"Kobo Sync Server",
|
"Kobo Sync Server",
|
||||||
"Rust, Axum",
|
"Rust, Axum",
|
||||||
repo_link: "https://github.com/potatoeggy/kobink",
|
repo_link: "https://github.com/potatoeggy/kobink",
|
||||||
)
|
)
|
||||||
- Developed an Axum backend to *automatically convert and sync* ebooks to a Kobo
|
- Reimplemented the Kobo Store Sync protocol in an Axum backend to *instead sync
|
||||||
eReader over the network
|
personal ebook files*
|
||||||
- *Reverse-engineered* the Kobo eReader Sync API to support updating ebook
|
- Deployed the service behind an *nginx reverse proxy* and *Cloudflare DNS* to
|
||||||
metadata and covers
|
allow syncing over the internet
|
||||||
|
|
||||||
// #project(
|
// #project(
|
||||||
// "RecipeReady",
|
// "RecipeReady",
|
||||||
|
5
lib.typ
5
lib.typ
@ -12,10 +12,10 @@
|
|||||||
#let TITLE_FONT_SIZE = 2em
|
#let TITLE_FONT_SIZE = 2em
|
||||||
|
|
||||||
#let HEADING_LINE_GAP = 3mm
|
#let HEADING_LINE_GAP = 3mm
|
||||||
#let BODY_LINE_HEIGHT = 0.85em
|
#let BODY_LINE_HEIGHT = 0.8em
|
||||||
|
|
||||||
// rest = not top
|
// rest = not top
|
||||||
#let PAGE_MARGINS = (rest: 0.5in, top: 0.4in, left: 0.4in)
|
#let PAGE_MARGINS = (rest: 0.5in, top: 0.4in, left: 0.4in, bottom: 0.4in)
|
||||||
|
|
||||||
// Format locations.
|
// Format locations.
|
||||||
#let format_location(location) = {
|
#let format_location(location) = {
|
||||||
@ -78,7 +78,6 @@
|
|||||||
"\n"
|
"\n"
|
||||||
set strong(delta: SMALL_STRONG)
|
set strong(delta: SMALL_STRONG)
|
||||||
text[*#company*]
|
text[*#company*]
|
||||||
|
|
||||||
}, right_content: {
|
}, right_content: {
|
||||||
[*#start_date -- #end_date*]
|
[*#start_date -- #end_date*]
|
||||||
|
|
||||||
|
Loading…
Reference in New Issue
Block a user